Generated by GPT-5-mini| Gym-MiniGrid | |
|---|---|
| Name | Gym-MiniGrid |
| Developer | OpenAI Gym community |
| Initial release | 2018 |
| Programming language | Python |
| License | MIT License |
| Repository | GitHub |
Gym-MiniGrid
Gym-MiniGrid is a lightweight set of reinforcement learning environments designed for rapid prototyping and research. It provides procedurally generated gridworlds suitable for agents implemented with frameworks like TensorFlow, PyTorch, JAX, and Keras. The project interfaces with ecosystems such as OpenAI Gym, Stable Baselines3, and RLlib to enable comparisons across algorithms like Deep Q-Network, Proximal Policy Optimization, and Soft Actor-Critic.
Gym-MiniGrid offers a collection of compact, partially observable tasks that emphasize sample efficiency, sparse rewards, and compositional generalization for agents. Researchers from labs affiliated with institutions such as Berkeley Artificial Intelligence Research Laboratory, DeepMind, OpenAI, MIT Computer Science and Artificial Intelligence Laboratory, and Google Research have used similar gridworld platforms to study exploration, hierarchical learning, and curriculum design. The environment suite complements benchmarks like Atari 2600, VizDoom, Procgen Benchmark, and DeepMind Lab by focusing on simpler state representations and rapid iteration.
The suite includes tasks such as goal-reaching, key-and-door puzzles, multi-room navigation, and object manipulation, comparable in purpose to scenarios in Sokoban, Montezuma's Revenge, and Taxi (video game). Each environment exposes observations (e.g., partially observable egocentric views) analogous to representations used in ImageNet pretraining pipelines and in studies from Stanford University and Carnegie Mellon University. Tasks are parameterized to support variations used in curriculum learning studies led by teams at University of California, Berkeley and University College London. Procedural generation settings permit evaluation protocols similar to those in competitions hosted by NeurIPS and benchmarks endorsed by ICML and IJCAI.
Installation is typically performed via Python (programming language) package managers and integrates with toolchains from Anaconda (software distribution), pip, and Conda Forge. Users combine Gym-MiniGrid with agents implemented using libraries maintained by organizations like Facebook AI Research (for PyTorch), Google Brain (for TensorFlow and JAX), and the Hugging Face ecosystem. Example usage patterns mirror tutorials from Coursera and educational material from edX courses taught by faculty at Stanford University and MIT. Continuous integration and testing workflows are often configured with services provided by GitHub Actions and Travis CI.
The codebase is implemented in Python (programming language) with modular classes reflecting designs inspired by OpenAI Gym and data structures similar to those in NumPy. The architecture separates level generation, tile semantics, and agent interfaces, enabling integrations with simulators and wrappers used in projects from Unity Technologies and DeepMind. Observations, actions, and reward signals follow conventions established in competitions hosted by NeurIPS and influenced by best practices from research groups at Oxford University and ETH Zurich. Developers often instrument experiments with visualization tools from Matplotlib and logging frameworks like those from Weights & Biases.
Benchmarks for Gym-MiniGrid emphasize sample efficiency and generalization across procedurally generated maps, comparable to evaluations published in papers from NeurIPS, ICML, and ICLR. Performance baselines frequently reference algorithms developed at DeepMind, OpenAI, and Facebook AI Research, and are reported alongside comparisons to environments like MiniWorld and Gridworld (programming) tasks used in studies from Cornell University and Princeton University. Metrics include success rate, episode length, and cumulative reward, following evaluation protocols advocated by researchers at Carnegie Mellon University and in workshops at AAAI.
The project has an active contributor base on GitHub with pull requests, issues, and discussions involving researchers from Microsoft Research, Google Research, DeepMind, and independent contributors who publish experiments in venues such as NeurIPS and ICML. Educational use is notable in university courses at Stanford University, MIT, and University of Toronto, where instructors employ the environments for assignments and reproducible research. Community tooling and extensions are hosted in forks and companion repositories maintained by organizations like Hugging Face and community groups participating in challenges at conferences such as NeurIPS and ICLR.
Category:Reinforcement learning environments