-
Notifications
You must be signed in to change notification settings - Fork 1
Include a licence #10
Comments
lol, i wasn't even aware that people were using this repo. cool to hear though! what are your plans exactly? i'm def open to the idea of turning this repo into something that is compatible with |
I'll add MIT for now, gotta say: I've not looked at this code for a while. |
also ... sushigo [the card game] is great, 'aint it? |
you might also be interested in this: https://www.youtube.com/watch?v=KUvB1hqF2OA&t=166s |
@koaning thanks for the quick response! I did also stumble upon your blog post regarding sushi-go earlier. My interest is in applying reinforcement learning to card/board games and sushi-go was a natural one that came to mind, I was pleased to discover that you've made an engine for the game already. Regarding api, it might be different for the gym api as gym is designed for single agent environments (open issue on multi-agent environments here: openai/gym#934). For games like this (turned-based, multi-agent, partially observable) it might require a different approach. Ideally (if time permits) I would like to design (or find a good existing implementation) of a general gym-like api for turn based multi-agent games which could work with plug-and-play environments like sushi-go. |
you might be interested to hear about ‘evol’ its an evolutionary algotirhms
package with support for “contestpopulations”. these are populations where
individuals ‘duke it out’ instead of being evaluated against a single
metric. this might be a fun avenue.
alternatively, as far as a single individual is concerned, you could
imagine that it still feels like the agent talks to a single environment.
you merely need to initialize the environment with agents in it.
another benefit of a gym is in the generality. it might be easier for
others to try out algorithms and there may also be agents that you can plug
and play.
On Wed 16. Jan 2019 at 15:46, Janis Klaise ***@***.***> wrote:
@koaning <https://github.com/koaning> thanks for the quick response! I
did also stumble upon your blog post regarding sushi-go earlier.
My interest is in applying reinforcement learning to card/board games and
sushi-go was a natural one that came to mind, I was pleased to discover
that you've made an engine for the game already.
Regarding api, it might be different for the gym api as gym is designed
for single agent environments (open issue on multi-agent environments here:
openai/gym#934 <openai/gym#934>). For games
like this (turned-based, multi-agent, partially observable) it might
require a different approach. Ideally (if time permits) I would like to
design (or find a good existing implementation) of a general gym-like api
for turn based multi-agent games which could work with plug-and-play
environments like sushi-go.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#10 (comment)>, or mute
the thread
<https://github.com/notifications/unsubscribe-auth/AA-Pj_CyKjsGGsdNqjsQwOX284dbp9AYks5vDztSgaJpZM4aC5iR>
.
--
*- Vincent*
|
@koaning sure, you can always make a multi-agent turn-based game into a single-agent environment by treating the other agents as static and thus part of the environment. It gets tricky if you want to train all agents at once, non-stationary environment and book-keeping become concerns. Thanks for the heads-up on evol, I was just wondering what is the motivation behind your company releasing this? Looks like a direct competitor to DEAP, although I think when I tried to use DEAP ~3 years ago I couldn't make a simple example work... |
its more a hobby project where collegues got interested. my company pays me
to do open source sometimes and this is one of iur contributions.
feedback is welcome. tomorrow we have support of parallel processing in the
evaluation phase.
On Thu 17. Jan 2019 at 10:31, Janis Klaise ***@***.***> wrote:
@koaning <https://github.com/koaning> sure, you can always make a
multi-agent turn-based game into a single-agent environment by treating the
other agents as static and thus part of the environment. It gets tricky if
you want to train all agents at once, non-stationary environment and
book-keeping become concerns.
Thanks for the heads-up on evol, I was just wondering what is the
motivation behind your company releasing this? Looks like a direct
competitor to DEAP, although I think when I tried to use DEAP ~3 years ago
I couldn't make a simple example work...
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#10 (comment)>, or mute
the thread
<https://github.com/notifications/unsubscribe-auth/AA-Pj5K_bt0G9slDtjtWrUmU3kazX1Dyks5vEELbgaJpZM4aC5iR>
.
--
*- Vincent*
|
Would it be possible to add a licence to the project (I would recommend either MIT or Apache-2)? I'm thinking of extending the current API to ultimately be able to train multi-agent RL algorithms, but it's tricky to reuse parts of the code without a licence in place. Thanks!
The text was updated successfully, but these errors were encountered: