Generated by GPT-5-mini| Glow (machine learning) | |
|---|---|
| Name | Glow |
| Developer | Dmitry Kaliakin; Kingma, Diederik and Dhariwal, Prafulla (original authors linked works) |
| Released | 2018 |
| Programming language | Python (programming language), C++ |
| Operating system | Linux, macOS, Windows |
| License | MIT License |
| Platform | TensorFlow, PyTorch |
Glow (machine learning) is a deep generative model introduced in 2018 that implements invertible flow-based transformations for density estimation and generative sampling. It combines ideas from normalizing flows, invertible residual networks, and bijective transformations to enable exact likelihood computation and efficient sampling for high-dimensional data such as images, audio, and tabular records. The model has been influential in research communities associated with Google Research, OpenAI, DeepMind, Facebook AI Research, and academic groups at Massachusetts Institute of Technology, Stanford University, and University of Toronto.
Glow originated from research on normalizing flows alongside works from Diederik P. Kingma, Durk P. Kingma, and colleagues at OpenAI and Google Brain; it extended prior methods such as RealNVP, NICE (machine learning), and invertible architectures like RevNets. The design emphasizes reversible bijections, enabling exact evaluation of the data likelihood under the model via the change-of-variables formula, an approach related to research at Berkeley AI Research, Princeton University, and Carnegie Mellon University. Glow has been compared and contrasted with autoregressive models developed at Google DeepMind and variational models associated with University of Oxford and New York University.
Glow's core architecture uses a sequence of invertible layers built from three principal components: actnorm scaling layers inspired by normalization strategies from Batch normalization developments at University of Toronto and Google Brain; invertible 1x1 convolutions that generalize permutation operations explored in works at Facebook AI Research; and affine coupling layers conceptually derived from studies at OpenAI and UC Berkeley. The model arranges these into multi-scale architectures reminiscent of hierarchical compression schemes in projects at ETH Zurich, EPFL, and Max Planck Institute for Intelligent Systems, enabling tractable Jacobian determinants and memory-efficient backpropagation techniques used in NVIDIA-backed implementations. Implementation toolchains often leverage infrastructure developed at PyTorch Foundation, TensorFlow teams, and compute clusters at Amazon Web Services or Google Cloud Platform.
Training Glow typically optimizes the exact log-likelihood using stochastic gradient descent variants such as Adam (optimization algorithm), RMSprop, or LAMB (optimizer). Techniques from large-scale training studies at OpenAI, DeepMind, and Microsoft Research—including gradient clipping strategies, learning rate schedules from AdaDelta research, and mixed-precision training popularized by NVIDIA—are commonly applied. Regularization and stability practices borrow from work at Stanford University and Harvard University on initialization schemes and Jacobian conditioning, while distributed training frameworks developed by Horovod and Ray (distributed computing) are used for scaling on clusters managed by Kubernetes or Slurm Workload Manager.
Glow has been applied to image synthesis tasks evaluated in benchmarks such as ImageNet, CIFAR-10, and CelebA for high-fidelity generation, and to audio modeling problems connected to datasets curated by Mozilla and research teams at DeepMind. It has been used for anomaly detection in industrial settings linked to deployments by Siemens and GE Digital, and for data compression prototypes influenced by work at Apple Inc. and Google Research. Researchers at MIT Media Lab and UC Berkeley have explored Glow for style transfer, latent space manipulation inspired by projects at Adobe Research, and interpretability studies aligned with efforts at Allen Institute for AI.
Evaluations contrast Glow against autoregressive baselines such as models from PixelCNN research at Google DeepMind and variational approaches like VAE studies at NYU. Metrics include bits per dimension on image benchmarks maintained by Stanford DAWNBench and sample quality assessments analogous to Inception Score and Fréchet Inception Distance, with empirical comparisons published by teams at UC San Diego and University College London. Computational trade-offs are analyzed relative to GAN research popularized by Ian Goodfellow and colleagues at OpenAI, and in latency-sensitive contexts evaluated by engineering groups at Intel and AMD.
Subsequent work extended Glow with conditional architectures for supervised synthesis explored at Facebook AI Research and Microsoft Research, discrete flow adaptations studied at ETH Zurich, and continuous-time formulations related to neural ODE research from Columbia University and MIT. Hybrid models combine flow components with autoregressive decoders developed at Google Research and latent-variable schemes advanced by University of Cambridge. Practical toolboxes and forks have been maintained by open-source communities on platforms such as GitHub and incorporated into model hubs supported by Hugging Face and industrial research groups at OpenAI and NVIDIA.
Category:Machine learning models