LLMpediaThe first transparent, open encyclopedia generated by LLMs

COCO

Generated by GPT-5-mini
Note: This article was automatically generated by a large language model (LLM) from purely parametric knowledge (no retrieval). It may contain inaccuracies or hallucinations. This encyclopedia is part of a research project currently under review.
Article Genealogy
Expansion Funnel Raw 76 → Dedup 23 → NER 16 → Enqueued 9
1. Extracted76
2. After dedup23 (None)
3. After NER16 (None)
Rejected: 7 (not NE: 7)
4. Enqueued9 (None)
Similarity rejected: 6
COCO
NameCOCO
CaptionCommon Objects in Context dataset
Released2014
DevelopersMicrosoft Research
DomainsComputer vision, Image recognition, Object detection, Segmentation, Captioning

COCO COCO is a large-scale image dataset for object detection, segmentation, and captioning used widely in computer vision research. It provides richly annotated images containing everyday scenes from sources such as Flickr and benchmarked tasks facilitating comparison across models from institutions like Microsoft Research, Facebook AI Research, and Google Research. COCO has influenced challenges organized by groups including CVPR, ICCV, and NeurIPS and underpins evaluation suites for systems developed by teams at OpenAI, DeepMind, and various universities.

Overview

COCO contains hundreds of thousands of images annotated with object instance segmentation, keypoints, stuff labels, and image captions, enabling tasks spanning object detection, semantic segmentation, instance segmentation, image captioning, and pose estimation. The dataset is commonly used alongside other corpora such as ImageNet, PASCAL VOC, Cityscapes, ADE20K, and Visual Genome in benchmarks and challenge leaderboards curated at conferences like ECCV and CVPR. Tools and frameworks that consume COCO annotations include Detectron2, Mask R-CNN, Faster R-CNN, YOLO, SSD, MMDetection, and libraries from TensorFlow and PyTorch ecosystems.

History and Development

COCO was introduced by researchers at Microsoft Research and collaborators to address limitations of prior datasets such as PASCAL VOC and ImageNet by providing context-rich scenes and instance-level segmentations. Early presentations and papers were circulated at venues including ECCV 2014 and subsequently expanded through community contributions and challenge tracks at COCO Challenges hosted with CVPR and ECCV. Successive releases added annotations for keypoints influenced by work from groups at University of Oxford, Stanford University, MIT, and Carnegie Mellon University on human pose and landmark detection. The dataset design drew on annotation platforms and crowd-sourcing paradigms pioneered by Amazon Mechanical Turk and methodological ideas from projects like LabelMe and Open Images.

Dataset Composition and Annotation

COCO’s image corpus includes diverse everyday scenes featuring objects across dozens of categories such as person, bicycle, car, airplane, bus, train, truck, boat, traffic light, fire hydrant, stop sign, parking meter, bench, bird, cat, dog, horse, sheep, cow, elephant, bear, zebra, giraffe, backpack, umbrella, handbag, tie, suitcase, frisbee, skis, snowboard, sports ball, kite, baseball bat, baseball glove, skateboard, surfboard, tennis racket, bottle, wine glass, cup, fork, knife, spoon, bowl, banana, apple, sandwich, orange, broccoli, carrot, hot dog, pizza, donut, cake, chair, couch, potted plant, bed, dining table, toilet, TV, laptop, mouse, remote, keyboard, cell phone, microwave, oven, toaster, sink, refrigerator, book, clock, vase, scissors, teddy bear, hair drier, toothbrush—categories frequently referenced in benchmarks alongside datasets from Google, Stanford, Berkeley AI Research, and Tencent research groups. Annotation pipelines used crowdworkers to collect polygonal instance masks, bounding boxes, segmentation maps, keypoint coordinates for human poses, and multiple natural language captions per image, drawing on practices from projects at Carnegie Mellon University and annotation standards influenced by research at University of Washington.

Evaluation Metrics and Benchmarks

COCO popularized metrics such as mean Average Precision (mAP) at varying Intersection over Union thresholds (IoU), average recall, and AP across object sizes (small, medium, large), shaping evaluation protocols used by teams at Facebook AI Research, Google Research, Microsoft Research Asia, NVIDIA, and academic groups at University of Oxford and ETH Zürich. Leaderboards track performance of architectures like Mask R-CNN, RetinaNet, EfficientDet, Swin Transformer, and end-to-end systems from OpenAI and DeepMind on tasks including object detection, panoptic segmentation, and keypoint detection. COCO-style metrics inform challenge rules at NeurIPS competitions and benchmark suites in industrial research labs such as Intel Labs and IBM Research.

Applications and Impact

COCO has driven progress in autonomous driving research at companies like Waymo, Tesla, and Uber ATG by improving scene understanding models originally tested on datasets such as KITTI and Cityscapes. It supports advancements in assistive technologies developed by groups at Apple, Amazon, and Google for image captioning and accessibility features. COCO annotations have been used in multi-modal model training by teams at OpenAI and Meta AI to bridge vision and language tasks seen in systems like CLIP, DALL·E, BLIP, and VisualBERT. Research on robotics perception at MIT CSAIL, Stanford AI Lab, and Carnegie Mellon University Robotics Institute has exploited COCO for object recognition and manipulation pipelines, while biomedical imaging groups have referenced COCO-style annotation schemes in adapting instance segmentation methods to medical datasets curated by institutions like Harvard Medical School and Johns Hopkins University.

Limitations and Criticisms

Critics from academia and industry, including commentators at MIT Technology Review and researchers at AI Now Institute, have noted dataset biases in COCO’s object category distribution, geographic skew, and dataset artifacts affecting fairness and generalization across populations encountered by systems deployed by Google, Facebook, and governments. Concerns raised parallel critiques of ImageNet and Open Images about representational harms, annotation errors, and the ecological validity of training on curated images versus in-the-wild video corpora like YouTube-8M. Methodological debates at NeurIPS and ICLR have argued for richer annotation types and longitudinal evaluation protocols advocated by groups at Stanford HAI and Berkeley AI Research, while reproducibility efforts from Papers with Code and open-source initiatives like Hugging Face encourage standardized splits and metadata for COCO-derived benchmarks.

Category:Computer vision datasets