Generated by DeepSeek V3.2| PyTorch | |
|---|---|
| Name | PyTorch |
| Developer | Meta AI |
| Released | October 2016 |
| Programming language | Python, C++, CUDA |
| Operating system | Linux, macOS, Microsoft Windows |
| Genre | Machine learning, deep learning |
| License | BSD |
PyTorch is an open-source machine learning framework primarily developed by Meta AI and now part of the Linux Foundation's umbrella. It is widely recognized for its flexibility and intuitive design, which leverages the dynamic computation graph paradigm known as automatic differentiation. This approach has made it a favorite for both academic research, such as at Stanford University and MIT, and industrial applications at companies like Tesla and Microsoft.
Built as a successor to the earlier Torch library, it provides a seamless path from research prototyping to production deployment, often integrated with platforms like ONNX Runtime. The core tensor computation is accelerated using libraries such as Intel MKL and NVIDIA's CUDA for GPU support, while its design emphasizes Pythonic principles. Its architecture allows for deep integration with the broader Python ecosystem, including NumPy and SciPy, facilitating rapid experimentation in fields like computer vision and natural language processing.
The project originated within Meta AI's research lab, with key contributions from researchers like Soumith Chintala. It was publicly released in October 2016, quickly gaining traction as an alternative to static graph frameworks like Google's TensorFlow. A significant milestone was the introduction of TorchScript in 2018, enabling graph-based execution for performance optimization. In 2022, governance was transferred to the newly formed PyTorch Foundation under the Linux Foundation, joining other projects like Kubernetes to ensure neutral stewardship.
A defining characteristic is its imperative, define-by-run style, which utilizes an autograd system for building dynamic neural networks. The torch.nn module provides building blocks for complex architectures, while distributed training is enabled through backends like NCCL and Gloo. The framework also includes LibTorch for C++ deployment, supports quantization via FBGEMM, and offers profiling tools integrated with TensorBoard. For specialized hardware, it can leverage AMD's ROCm and Google's TPU support through XLA.
It is extensively used in cutting-edge research published at conferences like NeurIPS and ICLR, particularly for developing transformer models and generative adversarial networks. Companies such as Uber and Airbnb employ it for recommendation systems, while OpenAI has utilized it for projects like GPT-2. In academia, institutions like Carnegie Mellon University use it for robotics research, and it is the backbone of many tutorials on platforms like Coursera and Fast.ai.
A rich set of libraries extends its capabilities, including TorchVision for computer vision, TorchText for NLP, and TorchAudio for audio processing. The Hugging Face platform hosts thousands of pre-trained models built with it, and tools like PyTorch Lightning and Ignite simplify high-level workflow abstraction. For mobile deployment, PyTorch Mobile supports iOS and Android, and the ecosystem interfaces with Apache Spark for large-scale data processing.
Unlike TensorFlow's historical static graph approach, it offers dynamic computation, which is often compared to Chainer's early methodology. While JAX provides similar autodiff capabilities, it is more tightly integrated with NumPy and functional programming paradigms. Performance-wise, it competes with Microsoft's Cognitive Toolkit and Apache MXNet, but its Python-first design and strong community support, similar to that of scikit-learn, have made it dominant in research circles.
Category:Free software programmed in Python Category:Free software programmed in C++ Category:Machine learning Category:Meta Platforms Category:Linux Foundation