Releases: beyretb/AnimalAI-Olympics
Releases · beyretb/AnimalAI-Olympics
v2.0.0 release
- v2.0.0:
- fix small bugs
- adds tutorial notebooks
- Bump ml-agents from 0.7 to 0.15 which:
- allows multiple parallel environments for training
- adds Soft actor critic (SAC) trainer
- has a new kind of actions/observations loop (on demand decisions)
- removes brains and some protobufs
- adds side-channels to replace some protobufs
- refactoring of the codebase
- GoodGoalMulti are now yellow with the same texture (light emitting) as GoodGoal and BadGoal
- The whole project including the Unity source is now available on our ml-agents fork
v1.1.1 release
Hotfix for curriculum which was loading in the wrong order
v1.1.0 release
Add curriculum learning to animalai-train
to use yaml configurations
v 1.0.5
- v1.0.5
- Adds customisable resolution during evaluation
- Update
animalai-train
to tf 1.14 to fixgin
broken dependency - Release source code for the environment (no support to be provided on this for now)
- Fixes some legacy dependencies and typos in both libraries
v 1.0.3 - submission ready
- v1.0.3
- Adds inference mode to Gym environment
- Adds seed to Gym Environment
- Submission example folder containing a trained agent
- Provide submission details for the competition
- Documentation for training on AWS
Plus previous additions not tagged as release:
-
v1.0.2
- Adds custom resolution for docker training as well
- Fix version checker
-
v1.0.0
- Adds custom resolution to both Unity and Gym environments
- Adds inference mode to the environment to visualise trained agents
- Prizes announced
- More details about the competition
v0.6.0 release
- Adds score in playmode (current and previous scores)
- Playmode now incorporates lights off directly (in
examples
try:python visualizeArena.py configs/lightsOff.yaml
) - To simplify the environment several unnecessary objects have been removed see here
- Several object properties have been changed also here
- Frames per action reduced from 5 to 3 (i.e.: for each action you send we repeat it for a certain number of frames
to ensure smooth physics) - Add versions compatibility check between the environment and API
- Remove
step_number
argument fromanimalai.environment.step
v0.5 release
- v0.5 Package
animalai
, gym compatible, dopamine example, bug fixes- Separate environment API and training API in Python
- Release both as
animalai
andanimalai-train
PyPI packages (forpip
installs) - Agent speed in play-mode constant across various platforms
- Provide Gym environment
- Add
trainBaselines,py
to train usingdopamine
and the Gym wrapper - Create the
agent.py
interface for agents submission - Add the
HotZone
object (equivalent to the red zone but without death)
v0.4 release
0.4 - Lights off moved to Unity, colors configurations, proportional goals, bugs fixes
- The light is now directly switched on/off within Unity, configuration files stay the same
- Blackouts now work with infinite episodes (t=0)
- The rand_colors configurations have been removed and the user can now pass RGB values, see here
- Rewards for goals are now proportional to their size (except for the DeathZone), see here
- The agent is now a ball rather than a cube
- Increased safety for spawning the agent to avoid infinite loops
- Bugs fixes
v0.3 release
We added the possibility to switch the lights off at given intervals
- visualizeLightsOff.py displays an example of lights off, from the agent's point of view
- Beams objects have been removed
- A Cylinder
object has been added (similar behaviour to the Woodlog
)
- The immovable Cylinder
tunnel has been renamed CylinderTunnel