LLMpediaThe first transparent, open encyclopedia generated by LLMs

LCS

Generated by GPT-5-mini
Note: This article was automatically generated by a large language model (LLM) from purely parametric knowledge (no retrieval). It may contain inaccuracies or hallucinations. This encyclopedia is part of a research project currently under review.
Article Genealogy
Parent: Overwatch League Hop 5
Expansion Funnel Raw 65 → Dedup 0 → NER 0 → Enqueued 0
1. Extracted65
2. After dedup0 (None)
3. After NER0 ()
4. Enqueued0 ()
LCS
NameLCS
FieldComputer science
Introduced1960s
RelatedLongest increasing subsequence, Edit distance, Sequence alignment

LCS

The LCS concept identifies a maximal common subsequence shared by two or more sequences and is central to sequence comparison, alignment, and diffing. It underpins classical results and algorithms in theoretical computer science, bioinformatics, data compression, and software engineering, with deep connections to combinatorics and dynamic programming. Researchers and practitioners from institutions such as Massachusetts Institute of Technology, Stanford University, University of Waterloo, University of California, Berkeley, and California Institute of Technology have contributed foundational work alongside figures at Bell Labs and companies like Google and Microsoft.

Definition and scope

In formal terms, given two finite sequences over an alphabet used in contexts like DNA sequencing and text processing, the object is a sequence that appears in both original sequences with indices increasing but not necessarily contiguous. Fundamental problems consider two-sequence and multiple-sequence variants studied in settings such as Knuth-inspired algorithmics, Donald Knuth-adjacent literature, and curricula at institutions like Princeton University and Carnegie Mellon University. The scope spans exact LCS, constrained LCS, and approximate variants used in frameworks developed at National Institutes of Health labs and industrial research groups at IBM Research and AT&T Labs.

Variants and formulations

Common variants include the pairwise LCS, the multiple-sequence LCS (mLCS) problem prominent in bioinformatics pipelines at European Bioinformatics Institute, the shortest common supersequence problem explored in Oxford University combinatorics groups, and constrained LCS that introduces pattern-enforcement or weightings as in work from École Polytechnique Fédérale de Lausanne. Other formulations relate to edit-distance, longest increasing subsequence studied by researchers affiliated with Columbia University, and sequence alignment models used by teams at Broad Institute and Cold Spring Harbor Laboratory. Parameterized versions focus on bounds such as alphabet size or subsequence length, a topic pursued at ETH Zurich and Max Planck Institute for Informatics.

Algorithms and computational complexity

Classic dynamic programming yields an O(mn) time and O(mn) space algorithm for two sequences, as detailed in textbooks from Addison-Wesley and courses at University of Illinois Urbana-Champaign. Improvements exploit bit-parallelism (Myers' algorithm), sparse dynamic programming, and Hunt–Szymanski techniques developed in venues like ACM SIGMOD and IEEE FOCS. Multiple-sequence problems are NP-hard, with hardness results linked to reductions from problems studied at Stanford and complexity classifications appearing in journals from SIAM and Elsevier. Parameterized complexity and fixed-parameter tractable (FPT) approaches have been advanced by groups at University of Edinburgh and Université de Paris, while approximation algorithms and heuristics are common in software from Adobe Systems and research at University College London.

Applications

Applications are broad: genome comparison and read assembly in projects at National Human Genome Research Institute and Wellcome Sanger Institute, diff and merge tools in version control systems like Git and Subversion, plagiarism detection used by services developed by Turnitin partners, and data deduplication in cloud services offered by Amazon Web Services and Microsoft Azure. Other areas include natural language processing pipelines at Google Research and Facebook AI Research, music sequence analysis in labs at Berklee College of Music, and error correction in communication systems researched at Bell Labs and Nokia Bell Labs.

Examples and case studies

Canonical textbook examples compare strings such as "AGGTAB" and "GXTXAYB" yielding a subsequence "GTAB", discussed in materials from Pearson Education and taught at University of Toronto. Case studies include large-scale comparative genomics projects at National Center for Biotechnology Information and pairwise alignment components of the BLAST pipeline, where LCS-inspired heuristics aid preprocessing. Industrial case studies demonstrate LCS usage in differencing engines in Microsoft Word and merge conflict resolution in GitHub repositories, while academic benchmarks include datasets curated by UCI Machine Learning Repository and sequence corpora from Project Gutenberg.

Implementation and optimization techniques

Practical implementations use memory-efficient rolling arrays, bitset acceleration as in Myers' bit-vector algorithm implemented in languages popular at Oracle Corporation and Google, and suffix-array or suffix-tree-based optimizations derived from work by researchers at University of Helsinki and University of Waterloo. Parallel and GPU-accelerated variants are implemented using toolkits from NVIDIA and libraries from Intel to scale pairwise comparisons in cloud environments at Amazon and Microsoft Research. Engineering trade-offs between exact algorithms and heuristics are documented in open-source projects hosted on GitHub and in performance studies published in conferences like USENIX and ACM SIGCOMM.

Category:Computer science topics