|Thread title||Replies||Last modified|
|Bots starting position randomizing is not working||10||13:50, 31 July 2018|
|How to keep roborumble running forever||2||01:04, 4 April 2018|
|Melee client corrupted||3||16:24, 31 October 2017|
|cpuConstant||9||20:38, 17 October 2017|
|Awful||0||05:59, 8 September 2017|
|Fire power 2.95 bug||3||02:17, 4 December 2015|
|Head on gun in melee||12||03:48, 12 November 2015|
|Smart bots competition||6||14:24, 18 October 2014|
|Article||6||15:42, 18 September 2014|
|Strange regression in robocode 1.9.2 vs 1.8.2||1||03:25, 2 June 2014|
|Gertjan1996 needs help||0||20:54, 21 May 2014|
|Bad bots in roborumble||0||03:15, 25 February 2014|
|Rumble Client||8||04:03, 10 February 2014|
|What is wrong with meleerumble?||9||15:40, 28 November 2013|
|what to do with about printing to much?||14||09:17, 21 November 2013|
|EvBot||7||04:51, 16 November 2013|
Can you double check it for me. I just notice that 1on1 rounds seem to have the same bot starting location. I.e. in 35 rounds battle, every round start (at least bot positions) are identical.
I see it in robocode-220.127.116.11
I don't know about RoboRumble but in RoboCode I see the randomizing works.
All random here also. Stupid question (but just to be safe): No fixed position set in the properties ?
Bingo! It apparently was set somehow for one of the bots.
But I cannot even find how to setup it in the GUI. I spotted it in the config after your hint.
Looks like I was right, robocode opens some secret to people with more than 10 years of experience :)
I've looked into the newest robocode source code, but couldn't find where the robot start position settings are — except for the tests used by robocode itself. Could you still remember what config it was to fix start positions?
Edit: got it, it just hides behind some direct access instead of accessors.
the config is "robocode.battle.initialPositions".
Thanks for mentioning that feature anyway!
It is good that you find it. I already forgot what it was. Strange thing I did not remember to set it up, so some bug triggered it back then.
However, it doesn't work when set in robocode.properties.
And from source code I know that this config is valid only in battle files.
And with a bug of robocode, once you open intro.battle, then the first bot's initial position will be fixed as long as another battle file is not opened. Neither "New Battle" nor "Restart" works.
So did yours appear in robocode.properties?
Thanks for running roborumble for so so many years! I tried to do the same thing with my VPSs but roborumble client kept getting killed in a few days.
What's your settings to keep roborumble running for such a long time? Did you experience similar problems running roborumble?
Funny thing that I don't do anything special. I just run the clients on my own computers. They all currently have Debian stable (stretch) linux distribution with openjdk-8 java. The robocode is installed manually in my home folder.
I run clients in the 'screen' sessions, so if I get disconnected the session survives. From time to time, I restart it manually, mostly for selfish reason of a quicker bot list upadate. But the current run is almost 100 days straight without interruption. My machines have 4 to 6 cores, so I just do normal office load (browsing, youtube, short term calculations) on them as well.
One more thing. I just checked that I client consumes about 4GB of memory (including virtual one) after that 100 days.
I would suspect that VPS shares you CPU with other virtual OSes and kills CPU consuming process forcefully.
For many years that I run robocode clients, I never had any problem with sporadic crashes. The restart happens only if my home looses power and UPS runs dry.
I think one of your meleerumble clients is corrupted, since it is continuously re-running battles against old versions of Neuromancer and Firestarter. Normally a client should get new bots once every 2 hours, but it has been a few days since the Firestarter update and yours is still running the old versions. Could you investigate?
Weird. The client was running battles with new version of bots but also attempted to upload ratings for the old ones. I cleared caches (files) and temporary (temp) dirs. Strangely enough I see the following at start up
Iteration number 0 Could not load properties file: ./roborumble/files/codesizemelee.txt Downloading rating files ... Downloading participants list ... Downloading missing bots ... Downloaded cb.fire.Firestarter 2.0d into ./robots/cb.fire.Firestarter_2.0d.jar Removing old participants from server ... Removing entry ... cb.fire.Firestarter_2.0e from meleerumble OK. cb.fire.Firestarter 2.0e retired from meleerumble Removing entry ... eem.EvBotNG_v11.0 from meleerumble OK. eem.EvBotNG v11.0 retired from meleerumble
Note that eem.EvBotNG v11.0 is 3 version behind, cb.fire.Firestarter_2.0e is one version behind.
I cannot figure out where does it read these versions since I cleared caches. My only assumption it is coming from the literumble. Am I right?
Nevertheless, despite the reported removal at the end of the battle I see
Fighting battle 0 ... fire219.CatBot 1.0,stelo.Spread 0.3,gh.mini.Gruwel 0.9,ara.Shera 0.88,wee.Gem 1.8.4,davidalves.net.DuelistNanoMelee 1.01,pedersen.Moron 2.0,zyx.mega.YersiniaPestis 3.1,emp.Yngwie 1.11,zyx.nano.BacillusComma 1.0 RESULT = stelo.Spread 0.3 wins, gh.mini.Gruwel 0.9 is second. Fighting battle 1 ... pedersen.Ugluk 1.1.1,florent.XSeries.X2 0.12,gimp.GimpBot 0.1,yk.JahRoslav 1.1,agrach.Dalek 1.0,lrem.Spectre 0.4.4,jcs.Seth 1.8,maribo.FollowFire 1,DTF.Kludgy 1.2b,franzor.Lizt 1.3.1 RESULT = pedersen.Ugluk 1.1.1 wins, DTF.Kludgy 1.2b is second. Fighting battle 2 ... stelo.WangRobot 1.0,dsw.StaticD 1.0,kurios.DOSexe .9b,jk.melee.Neuromancer 6.9,amk.Punbot.Punbot 0.01,oog.melee.CapuletDroid 1.1,lessonz.robocode.Oz 0.5.0,cs.sheldor.Talon 1.1,eem.IWillFireNoBullet v2.8,aaa.ScaledBot 0.01d RESULT = jk.melee.Neuromancer 6.9 wins, aaa.ScaledBot 0.01d is second. ... ... Ignoring: cb.fire.Firestarter 2.0e,cb.fire.Firestarter 2.0d,SERVER Ignoring: amk.Punbot.Punbot 0.01,cb.fire.Firestarter 2.0e,SERVER
Note that Firestarter was not even fighting the battles. Where is it coming from?
It all might be related to the sometimes observed appearance of two versions of the same bot in the rumble ratings which lingers for several hours.
Interesting, maybe it isn't your client uploading these bots then. Maybe it is User:Rsalesc? I should really add some more logging to the rumble so this kind of thing is more obvious.
Hi all, I decided to work on my skipped turns and to watch my execution time. As the first step, I recalculated cpuConstant according to cpuManager code (I wish it would be available from within the robot) at the beginning of each round. Guess what. The number wildly fluctuate. By wildly, we are talkin more than FOUR times!!! On the short end of the spectrum cpuConstant is about 8mS and it can be as large as 35mS.
Any ideas what is going on? Everything is done with openjdk8. One difference from the original code: I measure time with
Yes with Turbo Boost, CPU speed is not constant ;) And I think every new CPU should have something like that.
I would expect the cpuConstant go lower as turbo kicks in. But my first run usually gives lower number (faster speed), and consequent executions typically but not always are longer/slower.
In either case it is a problem. Recall recently discussed ThreadDeath issue.
Do you relay highly on GC? I mean do you create objects in loops? If the speed gets slower, I can only guess that would be GC.
Well, I do have many HashMaps and several kdTrees but the fact that Java kicks in with GC at random times is certainly an issue. The init round should be relatively low CPU event on my bot part.
Strange part is that robocode does not report 37 mS long execution as a skipped turn while my "official" cpuConstant is about 6mS.
Kd-trees and HashMaps is fast and GC friendly IMO. Did u tried recalculate CPU constant?
You mean the robocode properties file says CPU constant = 6ms?
Robocode gives robots extra time at the beginning of a round. In addition, most of this time is probably taken away by one-time setup and initialization code, making your calculated CPU constant longer than it should be.
I was browsing the wiki and found a couple places where authors mention the fire power 2.95 bug. Can someone tell me what it is? I did my best, but I could not find a description or rationale behind this magic number.
It is actually an x.x5 power bug. If you look at the history of BasicSurfer/Code, you will see that the old method of comparing bullet powers to filter onHitByBullet to match to the correct wave multiplied both by 10, rounded, then checked for equality. This meant that due to rounding errors, some values of bullets with the pattern x.x5 would not be matched and the bot would not learn from the bullet hit. Since many bots used BasicSurfer as a base, this actually caused a noticable increase in rumble score.
Yeah... IIRC I was doing some extensive default bullet power testing and found that 1.95 drastically outperformed both 1.94 and 1.96 in a large test bed and this turned out to be why. I felt kind of dirty since I was both the person that introduced the bug and the one who figured out how to exploit it. :-/
Folks, I have an issue which drives me crazy. How come that HawkOnFire with only head on gun have a high bullet bonus in melee? My EvBotNG is in top 20 in melee, yet when I run a simulation with HawkOnFire included, the bullet bonus it gets is just a touch higher than for Hawk's. Yes, EvBotNG survives way more often, so the survival bonus compensates.
When I try to use only the head on gun myself, my performance became even worse than the Hawk's one. I looked at the code and Hawk has only the head on gun. So HOW does it pull this miracle?
I'll have to think about it more, but off the top of my head, are you using the same bullet power? HoF uses power=3.0 a lot, which is more likely to give high bullet damage bonus. Also, if your bot's more survival oriented, it might keep more distance from opponents, which would cost you bullet damage in general.
And big congrats getting to #14 in Melee! Some very tough competition that high in the rankings.
I think you are right, the average distance must be the parameter in question. How could I forget about it.
Also, thanks for warm words with my humble progress. One thing in my mind is anti Diamond gun :) It is amazing how your bot survives with relatively short motions around a given conner in melee. I am thinking about a gun which instead of the most common GF would aim at the most visited location. I think it would score quite well against Dimond. But from the other hand, it looks like your bot is quite unique in this regard.
Do you have any recollections if this was tried in the glory days of robocode?
I don't remember anything like that. There were schemes like that for movement, I think, with different areas being considered safe/unsafe.
One thing I found helped a lot in Neuromancer was the bullet power formula, and taking distance of the target I was aiming at into account. Close targets get lots of bullet power, further targets get less bullet power. If lots of targets are bunched together, they also get lots of bullet power.
Good luck. I've been tempted a few times over the last few weeks to pull Neuromancer out and implement some of the features that are missing. If I do that I'd also like to open-source the current version of Neuromancer, since it seems nobody else has seriously tried a surf-everybody strategy.
It's very interesting watching the different movement strategies of different bots - Diamond sticking in the corners and picking off anybody who approaches, Shadow finding empty spots and aggressively clearing out close bots, Neuromancer skipping through the middle of the chaos and somehow dodging bullets.
The bullet energy formula is probably the most important ingredient, but every time I touch this even a little, my score drops down like a rock.
I am actually surprised by how few bots aim at Diamond in melee. It dances in a corner with a relatively simple triangular path with short motions. I would think a pattern matching gun should follow it quite well. But yet Diamond tend to out survive quite a lot of bots. But my old pattern matching gun is so sloooow, I did not even put in EvBotNG. From other hand, Diamond seems to be quite unique with this corner hugging strategy, so having a special gun just for it would not bring a big score boost.
Also, I think my old EvBot is dodging everyones bullets, it did not do surfing in the canonical form of predicting the bullet hit position, but it did take in account every wave. With new bot,I am hitting the skipped turns a lot, since I do exact path calculation and precise intersection now. I have a feeling that I do it wrong and it should not take so much CPU. But iterative procedures sucks, I should pull out my linear algebra skills and do non iterative hit position calculation.
As for Neuromancer, this one tend to survive in impossible situations. I use to think that you knew some hacks and actually see the bullets :)
Seeing Neuromancer in the open-source would be great. I do not see what stops you, I would not expect literate copies from true developers. But it would be loverly to see the tricks you use.
But even more exiting, would be to see new wave of discussions which comes with new developments. The wiki seems to be very quite lately, with not much going on.
Strictly speaking, Diamond doesn't have any deliberate corner movement strategy, and he certainly doesn't have a hard-coded triangular pattern that he follows. It's just a well-tuned minimum risk movement with some randomness based on recent locations. So hopefully it looks more predictable than it is. :-) One thing he does is precisely predict a few ticks ahead to avoid hitting walls, so his wall smoothing is pretty graceful and that may help him stay in the corners (where it's safer) and still avoid recent locations. Looking at it now, his melee movement is actually a very small amount of code: MeleeMover.java.
As for what's been tried before - my attitude is generally that everything has been tried before ;), but that doesn't mean it's not worth pursuing. Those attempts could have been suboptimal or outright broken and the landscape of competition has changed over time. If I were climbing the Melee ranks, I would certainly consider specialty guns for the top bots. And I have certainly considered it for Shadow and DrussGT in 1v1. :-)
But no, I don't remember anyone trying that. oldwiki:AreaTargeting came to mind, but looks like something different. Maybe you could formulate something like a GF that's based on a corner / quadrant of the field, the way a GF assumes orbital movement and factors out orbit direction. Like polar coordinates from nearest corner, with 0/north being towards the center of the field.
The reason I didn't want Neuromancer to be open source was because I wanted there to me more than one implementation style of surfing for Melee. Shadow and SilverSurfer had two very different styles of surfing, but when rozu open-sourced Apollon, everybody took that style (true-surfing) and nobody touched goto style until I came along with DrussGT. Perhaps though, by open sourcing Apollon it kickstarted the wavesurfing movement for 1v1 which may have never happened otherwise. I'll release an open version of current Neuromancer in a few weeks, that will give me some motivation to get these features added =)
It is quite possible to get a fast Play It Forward - the gun in Neuromancer depends on it.
And I think you may be disappointed with how simple Neuromancer is =) I suspect those who read through it will go "Oh, obviously." and proceed to write their own equivalent from scratch, much like others have done with the BasicSurfer.
As we know, there is hard (bad badly controlled) maximum processing time per turn allowed per bot imposed by the robocode engine. This pushes all of us to do the best in the allocated time limit. But as result some of potentially better but slower strategies do not show up at the top of the rating.
I am looking for a way to set/control allowed calculation time per in the rumble clients, in a hope to find the best among slow and wise bots. So we can start new rumble codenamed "wise slopokes". Additionally, it may spark new wave of interest among the kings of the rating authors.
What do you all think about this idea?
I like the idea of more processing time in order to attempt more complex stats, but to be honest I've tried some pretty complex stuff (including Spectral Clustering) for KNN on recorded data and nothing has beat simple KNN with a square kernel and weighted on sample distance. Not to mention that a lot of these algorithms aren't just a constant 50x slower or whatever, they essentially become unsolvable at sizes above ~500 data points, scaling quadratically or worse.
A problem with unlimited processing time is that it becomes possible to simply keep a copy of other bots and simulate them to determine where they will shoot/move.
Does anybody have any specific techniques that they would use if there was more processing time available?
I personally was aiming to a better movement algorithm. I.e. one can do better than 1 or 2 wave surfing. That would be super critical in melee, where you can track/react/predict more than just the closest bot.
Idea of enemy simulator is also attractive. I personally do not like clustering algorithms, I would rather prefer emulate and distill a subset of physically possible and optimal tracks for a current situation.
A clustering algorithm probably would seize to work once you apply it in melee, since you have to account for a neighboring bot subset as an input.I would think this would drastically increase the problem dimensions.
Displacement vectors of Voidious would be the closest to real emulation but it is still just an approximation.
I also have a crazy idea of a fine control of the motion, i.e. something more complicated than a simple go to a point which does not change for a several ticks. I want to have a path with smaller/different rotation angles and potentially speed which does not accelerate to max or min state. This produces a lot of possibilities for the paths and takes a lot of time to properly evaluate the danger. Of course it is overkill since such paths are often only 1 pixel apart.
I think DrussGT has a trick where at every several ticks the bot rotates to left and right, so the bot moves in wave like path. This confuses a linear targeting quite a lot.
Though, my goal would be to fit in between coming bullets with a fine control of a motion and also meet a wave with bot moving.
I had some movement algorithm things in the works, and they would have been more straightforward to implement with more processing time sure, but half the fun so to speak was coming up with the right optimization strategies to make it very fast.
Honestly, overall I don't feel like higher per-tick processing time would be too helpful at this point. My reasoning is, high level Robocode tends to require a very large amount of automated testing to know if some tuning was was actually an improvement or not. If the processing time limit was increased, it would also increase how long that testing takes.
Basically, I think the potential gains from slower algorithms are outweighed by the gains of faster iterations during tuning/development.
If anything I'd be more interested in a "FastBots" league, except I think the engine's control over the processing limit is too unreliable/unstable for that to be reasonable.
Yes, your article have changed me. I am not sure should I thank your or not :), since robocoding ate large chunk of my time.
I'm sorry, I seem to be missing some context. To what article are you referring?
I see strange regression in performance of my bot when it compiled with different version of robocode.
Specifically if I take EvBot v4.6.4 and compile it with robocode v1.8.2 I am getting about 20% higher score against let's say sqTank.waveSurfing.LionWWSVMvoid 0.01, However, when I compile with robocode v1.9.2 I see drastically worse performance. In both cases I run the battles with robocode v1.9.2.
Does anyone see similar behavior?
When I compare the stats of my more recent bot vs older v4.6.4 for example [4.6.12 vs 4.6.4], I see drop in a score against some bots which my bot use to be good against. There were quite a lot of changes and I originally thought it was due to experimental code, but now I am quite convinced that the major APS drop is due the change of version of robocode, since from v 4.6.5 and up I used robocode 1.9.2.
My rumble client contaminated with a few bad bots: cb.mega.Remedy 0.2 uji.SiberianKhatru 1.0
They produce the following error message: Can't load 'uji.SiberianKhatru 1.0' because it is an invalid robot or team. May be we should comment them out from the participants wiki page? Alternatively, may be something wrong on my side?
Hey, I just want to start by thanking you for running a rumble client. However, for some reason the results from your client for DrussGT are very different from what I'm getting with my local tests. I'm just curious what your setup is and what might be causing this.
There is nothing special. Rumble clients runs with a stock config, where only user name is set and I exclude Lo_Ian.Gandalf_V4*, since it seems to halt on my side. This 6 CPU machine, but only 2 are heavily used for roborumble and meleerumble clients, though last week I left It unattended and firefox start to consume a lot of load, so may be it somehow skewed CPU constant of the client.
If you have some ideas what could it be, I will be happy to investigate.
Things often brought up in rumble client discrepancy cases are which OS is in use, and what exact version of Java is in use.
Also, other possible relevant factors are:
- If CPU frequency scaling is enabled, as it is by default on most newer machines (this can cause problems for robocode's CPU time limiting being fair)
- If the system in question has high memory usage such that it goes into swap space.
I am not sure about CPU scaling. This is what ever Debian does to relatively modern AMD CPU. As for the swap, I am 99.99% sure it is not used. This machine has 16GB out of which only about 6 is used. If I look at memory usage, I seen only 3MB of swap used which is way to small for JAVA. May be the problem with linux scheduler, which constantly move application from one CPU to another.
There is one strange thing though, I notice that my meleerumble client usually crash within couple days, but roborumble is not. But this is true for another computer of mine as well.
FWIW I don't think anybody's measured the effect of dynamic clock speed on skipped turns, only speculated that it should have an effect. For all I know, the timings coming from Java on modern CPUs could have so much variance that the dynamic clock speed doesn't increase the number of skipped turns that much.
Here I run a single client with blank config for a while so it calculates CPU speed alone, before dynamic clock kicks in. Then I copy config to all clients and run them in parallel.
The effect is they assume a worst case in CPU speed and skipped turns occurs less often than it should. Which I think is better than occurring more often than it should.
Thanks for the offer. Could you test DrussGT 3.2.1 vs. Garm, Shadow, Hydra and Gilgalad? Functionally, 3.2.1 should be almost equivalent to 2.8.16 (3.2 has 3-series bullet shielding disabled), but as this comparison shows they are getting very different scores.
There are also a few bots DrussGT now gets 100% against - this worries me, as it points to them crashing.
What should be the setting on a client to perform such test? Or should I use RoboRunner?
Also my observation showed that sometimes short test set shows drastically (1% APS) higher scores. I was fulled by it couple times thinking that my older bot is more superior to a new version. They end up to have the same score or even reversed as I keep waiting for stats accumulation.
Looks like meleerumble is not updated anymore. If one goes to http://literumble.appspot.com/Rankings?game=meleerumble she would see that the most recent update was done on 2013-11-07 while today is 2013-11-16.
That's because nobody is running a client for it. You're welcome to run it yourself, unfortunately I had to stop the client I was running when I moved out of my old lab.
I find it impossible to run a client for meleerumble, robocode just crushes on attempt to submit results. I have no such problem for roborumble, which continues to happily accept my CPU cycles :)
But someting is broken in meleerumble robocode. Here is the relevant part of the console log for ./meleerumble.sh v18.104.22.168 run
RESULT = mld.Wisdom 1.0 wins, bayen.UbaMicro 1.4 is second. Fighting battle 2 ... dq.Finity 0.2,jangs.ns51 1.0,simonton.micro.Sprout 1.1.3,baal.nano.N 1.42,ap.Frederick 1.1,amk.jointstrike.JointStrike 0.2,cs.sheldor.Talon 1.1,bvh.mini.Fenrir 0.39,arthord.NanoSatanMelee Beta,fullsail.TimbotNoPrediction 1.0 Can't load 'dq.Finity 0.2' because it is an invalid robot or team. RESULT = simonton.micro.Sprout 1.1.3 wins, bvh.mini.Fenrir 0.39 is second. Uploading results ... OK. Portia vs Improved added to queue in 5ms Exception in thread "Application Thread" java.lang.ArrayIndexOutOfBoundsException: 2 at net.sf.robocode.roborumble.netengine.ResultsUpload.uploadResults(ResultsUpload.java:176) at roborumble.RoboRumbleAtHome.main(RoboRumbleAtHome.java:137)
I just tried a client and it is working fine, so it seems like a problem with your install. Delete the /robocode/roborumble/files and /robocode/roborumble/temp directories and try again.
Thanks Skilgannon. Above recipe fixed the issue.
But I find it strange, I am quite sure that I was trying to run it from the fresh install.
Something is wrong with meleerumble. My client fails with
Uploading results ... Exception in thread "Application Thread" java.lang.ArrayIndexOutOfBoundsException: 1 at net.sf.robocode.roborumble.netengine.ResultsUpload.senddata(ResultsUpload.java:293) at net.sf.robocode.roborumble.netengine.ResultsUpload.uploadResults(ResultsUpload.java:184) at roborumble.RoboRumbleAtHome.main(RoboRumbleAtHome.java:137)
On top of it the last upload was 11 hours ago and it was from me, but I think it start to misbehave by this time already.
I think it's your client again, same fix procedure as last time. You aren't running two clients out of the same folder, are you? It can corrupt the results file, which causes the client to crash.
I am quite convinced that it is on my side, but I was suspecting that others had the same problem, due to lack of uploads.
Same procedure did not fix it, but today I started it again after 8 hours and it works now.
I do run roborumble and meleerumble in the same folder, strangely roborumble is stable as a rock, while meleerumble gives me hick ups. Do you suspect some race condition and advise to use separate folders?
I am trying to print quite a lot while I am debugging my bot. But I see
SYSTEM: This robot is printing too much between actions. Output stopped until next action.
I would like to see the whole output. Is there a way to force robocode to print everything, something analogous to debug mode where all graphics events are plotted.
One option would be to write to a file instead of printing debug output, particularly if you adjust the per-bot disk quota in the config file.
Do we have a code sniplet anywhere for writing into a file?
Kind of a weird link, but PEZ recently posted all his bots to a GitHub repo so that's fresh in my mind. See the robot.getDataFile() stuff: 
Thanks, I will look at it.
Is there a way to see if bot run with GUI or through a rumble client? So one can disable all of the debugs during rumbling matches to save on CPU and quotas.
One more questions, why rumble client needs a display under x11 environment? It does not use gui so it should run just fine.
I don't think Robocode exposes anything about the environment (app vs control API, graphics or not). You can tell if graphical debugging is enabled, though, because onPaint() will be called. You could use that as your debug switch. I mainly use graphical debugging and log severe stuff to file.
As for X11, I don't know, I would guess it's due to using some java.awt code internally or something. I have run RoboRumble over a terminal quite a bit, but maybe my terminal has the X stuff setup. (Mac OS X Terminal connecting to Ubuntu.)
Not sure if it's useful, but a secondary suggestion is to try debugging graphics. Some info that takes a lot of text is very simple and clear in graphical form.
Hey, a quick request with EvBot, could you only have one version in the rumble at a time? If you want to compare their scores you can do that in the details page.
Also, just a suggestion, you will get stable rumble results quicker if you make EvBot execute faster. Right now it runs quite slowly, and while that *is* allowed, it makes testing quite a bit more painful ;-)
I removed the second entry and apologize for such misbehavior.
Can you suggest a tool to test my bot offline? I use RoboRunner, but I see its result different from robocode graphical interface, and even then it is different for paint on/off cases. I am guessing it is doe to the fact that my bot is super slow, and has more skipped turns in no gui environment.
Don't worry about the second one, lots of people do that when they're starting out. It's just better to have one so the rankings don't get as full, and "top 100" actually means "top 100" etc...
Yeah, that sounds like you're dealing with skipped turns problems. RoboRunner should give you similar results to the main RoboRumble (at least, in pair-on-pair scores). If you minimise the graphical interface do the results look similar to what you're getting with RoboRunner? Also, make sure you aren't running too many simultaneous threads in RoboRunner, if you're running more threads than cores it can impact skipped turns pretty heavily.
Why don't you make a page for EvBot where you explain what your general strategy is, we might be able to help you with getting it to run faster and avoiding the skipped turn problems.
It's also worth noting you can always revert back to a previous version in the rumble and it retains all its battles. And you can compare to previous versions forever. So you don't lose much by having only 1 version at a time on the participants list.
Nice to hear you use RoboRunner! :-) +1 to what Skilgannon said about those issues. But it's also true that there's almost no replacement for entering your bot in the rumble, so don't stress too much about posting versions frequently. Unless you've got quite a few cores to throw at Robocode, you'll get more battles in the rumble, and you may get slightly different scores on different machines / JVMs too.
Guys, thanks for your suggestions. My windows manager does not has a minimize button so I am unable to see the difference with minimized robocode. But I remove my slow code and everything seems to be way more predictable. The slow code is a pif gun with which I step in all beginners pitfalls myself.
Voidious, I have a suggestion for RoboRunner installer. Do not copy robots folder, just link to it if you think it is a must. At first I was shocked how long the initialization was performed and then I saw 500MB directory for each thread.
The whole separate "robots" directory thing is not for no reason: In order to avoid conflicts between the running instances, I'm pretty sure one needs a separate "robot.database" file, and separate ".data" directory inside it. Unless I'm missing something, without modifying the robocode engine a little there's probably no way to do that with (sym)links, except by creating a link for each and every jar file, which would be a touch silly in some ways.
To improve that without symlink-spam I see one of two possible approaches:
- Modify the robocode engine to separate the "robot.database" and ".data" directories from
- Modify the robocode engine so it'll read jar files nested in a folder in the "robots" directory (thus allowing a simple symlink to a central robots directory)
For this purpose, personally I'd lean toward the first option simply because symlink APIs/commands are slightly less cross-platform (though I'd personally also not mind #2 supported for other reasons)
Doh! Sorry about that Beaming. The script is a little quick and dirty compared to what it could be. I haven't touched RoboRunner in a while, but I'll take a look when I get a chance.
@Rednaxela - It's true you need separate robots directories for each install, but I can indeed fix the issue Beaming describes with a more sophisticated shell script. The RoboRunner setup script just uses "cp -r" to clone X copies of whatever Robocode install you point it at, which means it makes X copies of the robots dir. But when using RoboRunner, it copies bots as needed from the "bots" directory into the Robocode installs, so it would be fine to leave them empty during setup.