Jump to content
Search In
  • More options...
Find results that contain...
Find results in...
david_a

Visual Doom Bot Competition

Recommended Posts

The 2016 IEEE Computational Intelligence and Games Conference will include a very interesting AI competition - Visual Doom! Participants will create a Doom bot (hooking into a custom version of ZDoom) that works purely off the same input that human players have. The competition will involve good ol' deathmatch with the winner having the highest frag count. More details about entering the competition (for you brave souls out there) can be found at the Visual Doom competition page.

Share this post


Link to post

Yipe. This looks cool as hell. All this is happening while I'm sitting here contemplating life and wasting time on the internet.

Share this post


Link to post
AD_79 said:

visual doom page linking to wikia smh

cool though

I noticed that. One of us should probably correct them :)

I don't know how plugged-in to the community the people behind this are, but since they based their custom port on ZDoom and linked to a wiki (even if it's the wrong one) they seem like they know a little bit about the game.

Share this post


Link to post
david_a said:

Participants will create a Doom bot (hooking into a custom version of ZDoom) that works purely off the same input that human players have.

Has anyone managed to run ViZDoom on their computer (I haven't), so that they would confirm if the automap is available or not, respectively if IDDT cheat works? I bet the bots would use it pretty much solely instead of normal 3D vision.

Share this post


Link to post
david_a said:

I noticed that. One of us should probably correct them :)


I've left a note on their github.

Share this post


Link to post

I wish this was for vanilla Doom or PrBoom+, or maybe even Zandronum, but not ZDoom.

Share this post


Link to post
scifista42 said:

Has anyone managed to run ViZDoom on their computer (I haven't), so that they would confirm if the automap is available or not, respectively if IDDT cheat works? I bet the bots would use it pretty much solely instead of normal 3D vision.

Well apparently the competition involves deathmatch, so cheats like IDDT will presumably be disabled.

Share this post


Link to post

This is cool.

scifista42 said:

Has anyone managed to run ViZDoom on their computer (I haven't), so that they would confirm if the automap is available or not, respectively if IDDT cheat works? I bet the bots would use it pretty much solely instead of normal 3D vision.

That'd defeat the purpose of the exercise, plus it'd mean not being able to tell anything about where height changes are. Relying on the automap view might possibly make basic navigation easier in the short term, but it'd limit the bot's ability to improve.

Share this post


Link to post
esselfortium said:

That'd defeat the purpose of the exercise,

And that's exactly why I'm asking whether the engine makes it possible or not, because if yes, the participants will surely exploit it. If not entirely, at least to some extent.

Share this post


Link to post

I don't see what is cheating here. If you can parse the automap easier than the player view, just do it. You'll still lose some information, but you cam tell the bot to press tab periodically and record the player view.

Can you also rely on sounds with this? Or is ViZDoom totally silent?

Share this post


Link to post

Uhh, if they are going to use the automap then it's not interesting at all... I want to see them analyze a mess of pixels in a 320x200 resolution and try to get some useful info from it. I hope they will clarify on this.

Share this post


Link to post

This reminds me of that Starcraft 2 bot that someone created. It actually analysed the screen for information and nothing else, if I understood it correctly.

Anyway, I want to see more info as this progresses. It seems so hard to me to just rely on the player's viewport. Using the automap too would help with navigating. It should be allowed IMO.

Share this post


Link to post
Linguica said:

Apparently they are allowed to use a depth buffer as well, which seems to defeat the purpose?

Yeah a big part of the challenge would be to figure out how close a nearby wall or object is. Using the depth buffer from the rendered scene is a massive cop-out. Makes it much easier to figure out what is a wall (same depth going down), what is floor or ceiling (same depth going across) and what is an object (same depth going down and across).

Share this post


Link to post
Jon said:

I guess that implies GL-only too.

I guess so, though Quake's software renderer uses a depth buffer.

Share this post


Link to post

I was able to approximate a depth buffer in Wolf4SDL by calculating the pixel color change density per area. I took advantage of the pixellated graphics. However, this is not effective at low resolutions, and can be fooled by textureless areas.

Share this post


Link to post

Does prize money constitute commercial usage?

This would probably have been more interesting if it were for example Chocolate Doom playing with just the screen buffer and sound channels compared to say an OpenGL port with a depth buffer.

Share this post


Link to post

Why the hate for the depth buffer? Isn't that basically what other real-world systems use? Like autonomous cars using Lidar.

Share this post


Link to post
boris said:

Why the hate for the depth buffer?

Participants will create a Doom bot that works purely off the same input that human players have.

Can AI effectively play Doom using only raw visual input?

Share this post


Link to post

But the human brain has an built-in depth buffer, too, so to speak. Giving this information to the program is a more realistic scenario IMO.

Share this post


Link to post
boris said:

But the human brain has an built-in depth buffer, too, so to speak. Giving this information to the program is a more realistic scenario IMO.

Oh come on, monitors display a 2D image. If the point is to research a computer program that can reconstruct a 3D understanding from a 2D image, then giving access to a depth buffer is totally missing the point. If the point is to navigate an environment using a depth buffer as pseudo-lidar, then why put any emphasis on a comparison to human play?

Share this post


Link to post
Linguica said:

If the point is to navigate an environment using a depth buffer as pseudo-lidar, then why put any emphasis on a comparison to human play?

To be fair, the bots are going to suck even with access to the depth buffer, so that comparison to human play would be interesting and worth researching/showcasing even then. But it's still contradicting the point that they are proposing the challenge with - "only raw visual input".

Share this post


Link to post

Yes, when put like that, the bots should have access only to the displayed frame and the sounds being played. And they should use the player available controls in any way they seem fit. Like opening the automap every few milliseconds for example.

But it's still very interesting. I want to see what progress will be made in these short few months. Maybe printz can get some ideas or code for his AutoDoom project. (although it is different)

Share this post


Link to post
×