Generated by GPT-5-mini| PyTorch (software) | |
|---|---|
| Name | PyTorch |
| Developer | Meta Platforms, Inc.; originally by Facebook's AI Research Lab |
| Released | October 2016 |
| Programming language | C++, CUDA, Python |
| Platform | Linux, macOS, Windows |
| License | BSD-style |
PyTorch (software) PyTorch is an open-source machine learning framework for tensor computation and deep learning, developed initially by researchers at Facebook AI Research and maintained by Meta Platforms, Inc. and a broad community. It provides dynamic computation graphs, automatic differentiation, and GPU acceleration, and is widely used alongside projects from Google Research, Microsoft Research, OpenAI, and academic institutions such as Stanford University and University of Toronto. PyTorch integrates with ecosystems like NumPy, CUDA, and tools from NVIDIA and Intel Corporation for production deployment and research collaboration.
PyTorch emerged in the context of contributions from researchers associated with Facebook AI Research, NYU, and University of Oxford following earlier work on tensor libraries such as Torch (machine learning), Theano, and Caffe. Early releases in 2016 coincided with advances announced at conferences like NeurIPS and ICLR, with subsequent development influenced by software from Google Brain and industry shifts exemplified by projects at Microsoft Research and OpenAI. Major milestones included integration of a stable C++ frontend, announcement of the TorchScript mechanism for model serialization, and collaboration with hardware partners such as NVIDIA and Intel Corporation to optimize kernels for CUDA and oneAPI. The project’s governance evolved through contributions from corporate labs, academic groups including Massachusetts Institute of Technology and University of California, Berkeley, and community participants active on platforms like GitHub and events such as PyCon and NeurIPS workshops.
PyTorch offers features comparable to frameworks produced by Google Research and Microsoft Research while emphasizing a flexible API inspired by Torch (machine learning). Core features include dynamic computation graphs for research workflows discussed at ICLR and NeurIPS, automatic differentiation via an autograd engine influenced by designs from Theano and Autograd (software), and tensor operations interoperable with NumPy and SciPy. PyTorch supports distributed training using paradigms common to systems from TensorFlow and Horovod and provides model serialization compatible with deployment tools developed by ONNX partners such as Microsoft Corporation and Amazon Web Services. Additional features include mixed-precision utilities aligned with work from NVIDIA and quantization techniques studied at institutions like University of Edinburgh.
The architecture comprises a Python frontend that orchestrates computation, a C++ backend for performance similar to designs in XLA and MLIR, and GPU kernels leveraging CUDA and libraries from NVIDIA. The autograd engine implements reverse-mode differentiation analogous to approaches from Autograd (software) and Theano, while just-in-time (JIT) compilation via TorchScript enables static graph generation for production systems akin to TensorFlow Serving. The modular core allows integration with compiler infrastructures such as MLIR from Google and runtime systems like ONNX Runtime and TVM, reflecting cross-industry patterns in projects at Intel Corporation and ARM Holdings. The build and extension mechanisms interact with package ecosystems managed on Conda (package manager) and pip.
PyTorch’s ecosystem includes libraries developed by corporate labs and academic groups: TorchVision for computer vision tasks with datasets referenced by groups at Stanford University and University of Illinois Urbana-Champaign; TorchAudio with influences from research at Johns Hopkins University; and TorchText linking to NLP benchmarks promulgated by ACL community members. Higher-level libraries include PyTorch Lightning promoted by contributors from NYU and Montreal Institute for Learning Algorithms, and research frameworks like Detectron2 from Facebook AI Research and FastAI associated with University of San Francisco instructors. Model zoos, datasets, and tools interoperate with Hugging Face initiatives originated by teams in Paris and New York City, while deployment stacks integrate with cloud providers such as Amazon Web Services, Google Cloud Platform, and Microsoft Azure.
PyTorch is used across industry and academia in domains connected to research groups at MIT, Caltech, ETH Zurich, and corporate labs like DeepMind and OpenAI. Use cases span computer vision pipelines in enterprises such as Amazon and Uber Technologies, Inc., natural language processing systems developed by teams at Hugging Face and Google Research, reinforcement learning research at DeepMind and OpenAI, and scientific computing projects at Lawrence Berkeley National Laboratory. PyTorch supports workflows in healthcare research associated with Harvard Medical School and Johns Hopkins University, autonomous systems collaborations with Waymo and Tesla, Inc., and recommendation systems deployed by Netflix and Spotify.
Benchmarking literature compares PyTorch performance against frameworks from Google Research and Microsoft Research using hardware provided by NVIDIA (A100, V100) and Intel Corporation accelerators, with optimization studies published in venues like ICML and NeurIPS. Performance varies by workload—convolutional networks, transformer models popularized by researchers at Google Research and Google Brain, and graph neural networks explored at Stanford University—and benefits from integrations with CUDA libraries, kernel fusions researched at Facebook AI Research, and compiler optimizations from TVM and MLIR. End-to-end throughput and latency comparisons often reference deployment platforms from Amazon Web Services and Microsoft Azure and are reported in benchmarks conducted by industrial labs including NVIDIA and Meta Platforms, Inc..
Category:Machine learning software