Generated by GPT-5-mini| MLIR | |
|---|---|
| Name | MLIR |
| Developer | Google (company) |
| Released | 2018 |
| Programming language | C++ |
| Platform | LLVM |
| License | Apache License 2.0 |
MLIR
MLIR is a modular compiler infrastructure project originated at Google (company) and developed in collaboration with projects across the LLVM ecosystem, TensorFlow developers, and members of academic labs at institutions such as Stanford University, MIT, and University of California, Berkeley. It provides a reusable intermediate representation and framework intended to bridge domain-specific systems from high-level frameworks like PyTorch (machine learning) and JAX (software) down to low-level runtimes and hardware backends developed by vendors such as NVIDIA, Intel Corporation, and ARM Limited. MLIR emphasizes extensibility, multi-level optimization, and integration with compiler toolchains including Clang, GCC-adjacent workflows, and specialized accelerators from companies like Google (company) and Apple Inc..
MLIR was introduced in 2018 as part of ongoing efforts by engineers at Google (company) collaborating with contributors from TensorFlow teams and maintainers of LLVM to address rising complexity in compilation for machine learning models and heterogeneous hardware. Early design discussions referenced lessons from projects such as GCC, LLVM, and domain-specific compiler efforts at Microsoft, Facebook, Inc. (now Meta Platforms, Inc.), and academic prototypes developed at Carnegie Mellon University and ETH Zurich. The project gained adoption through integrations with prominent frameworks including TensorFlow and PyTorch (machine learning), and through presentations at venues like LLVM Developers’ Meeting, NeurIPS, and International Conference on Machine Learning. Major milestones included upstreaming of core libraries into the LLVM project tree and cross-industry collaboration with hardware vendors such as NVIDIA, Intel Corporation, AMD, and firms in the MLPerf ecosystem.
MLIR's architecture centers on a hierarchical intermediate representation that supports multiple abstraction layers, enabling transformations between representations inspired by precedents such as Static Single Assignment (SSA) form implementations in LLVM and graph-based models from TensorFlow. The design uses a dialect mechanism allowing teams at Google (company), Apple Inc., Huawei, and research groups at University of Toronto and Tsinghua University to introduce domain-specific operations and semantics without altering the core infrastructure. Core components include an extensible type system, region- and block-based control flow borrowed from LLVM builders, and a pass manager modeled after compilation flows used by GCC and clang. The infrastructure interoperates with code generation backends contributed by organizations such as NVIDIA and Intel Corporation, enabling lowering to device-specific ISAs and runtimes like CUDA, ROCm, and Vulkan.
MLIR enables creation of multiple dialects to represent distinct computational paradigms, with notable dialects developed by entities including TensorFlow, XLA (Accelerated Linear Algebra), and research labs at MIT. Examples include tensor-oriented dialects used by PyTorch (machine learning) integrations, linear algebra dialects informed by work at Princeton University and EPFL, and hardware description dialects driven by collaborations with ARM Limited and Cadence Design Systems. Each dialect defines operations and types; operations are used in transformations and analyses carried out by contributors at Google (company), Facebook, Inc./Meta Platforms, Inc., Amazon Web Services, and academic teams at University of California, Berkeley. The dialect system has enabled domain-specific lowering pipelines used by compilers targeting accelerators from NVIDIA and custom ASIC efforts from startups backed by firms like Sequoia Capital and Andreessen Horowitz.
Tooling around MLIR includes a suite of utilities, pass managers, and integration points with existing projects such as LLVM, Clang, and Bazel. Companies including Google (company), IBM, Microsoft, and Intel Corporation have contributed passes and tooling for optimization, verification, and code generation. The project supports debugging and visualization utilities used at events like LLVM Developers’ Meeting and demonstrated in workshops hosted by SIGPLAN and ACM conferences. Development workflows frequently leverage build systems and continuous integration practices common at Google (company), GitHub, Inc., and enterprise contributors like Red Hat and SUSE to maintain cross-platform compatibility and reproducibility.
MLIR is applied across domains by teams at Google (company), research groups at Stanford University and UC Berkeley, and industry partners such as NVIDIA and Intel Corporation. Common applications include compilation of TensorFlow graphs and PyTorch (machine learning) models, optimization pipelines for domain-specific languages developed at MIT and Carnegie Mellon University, and backend generation for accelerators in collaborations with ARM Limited and AMD. MLIR has been used in production systems for model deployment on mobile platforms like those produced by Apple Inc. and Qualcomm, in datacenter inference pipelines run by Amazon Web Services and Google Cloud Platform, and in research prototypes presented at NeurIPS and ICML.
Adoption of MLIR spans contributors from Google (company), TensorFlow, PyTorch (machine learning), LLVM maintainers, hardware vendors such as NVIDIA, Intel Corporation, and ARM Limited, and academic collaborators at MIT, Stanford University, and University of California, Berkeley. Community contributions are coordinated via platforms such as GitHub, Inc. and discussed at venues including LLVM Developers’ Meeting, NeurIPS, and ACM SIGPLAN conferences. Commercial ecosystems involving firms like Red Hat, IBM, Microsoft, and startup accelerators funded by Sequoia Capital have produced tooling and dialects that expand support for heterogeneous platforms and novel ML workloads. Continued growth is driven by cross-industry standards efforts and collaborations between major research labs and corporations.
Category:Compilers