11-667: Large Language Models Methods and Applications - Carnegie Mellon University
Product
Capabilities11 decomposed
llm fundamentals curriculum delivery and structured learning progression
Medium confidenceDelivers a comprehensive, sequenced curriculum covering large language model theory, architecture, and applications through structured course modules. The system organizes learning materials into progressive difficulty levels (beginner to advanced) with integrated lectures, assignments, and practical exercises that build foundational understanding of transformer architectures, attention mechanisms, training methodologies, and deployment patterns. This is implemented as a university-level course structure with curated content pathways rather than ad-hoc documentation.
Combines rigorous academic curriculum design with practical LLM applications, structured as a full-semester course at a top-tier institution rather than scattered tutorials or documentation. Integrates theoretical foundations (attention mechanisms, training algorithms) with contemporary applications (prompt engineering, RAG, agents) in a coherent learning progression.
Provides deeper theoretical grounding than most online tutorials or documentation, with university-level rigor and peer-reviewed content, while remaining more accessible than academic papers alone
transformer architecture deep-dive with mathematical foundations
Medium confidenceTeaches the complete transformer architecture including self-attention mechanisms, multi-head attention, positional encoding, feed-forward networks, and layer normalization through mathematical derivations and conceptual explanations. The curriculum covers how attention computes query-key-value projections, why positional encoding is necessary, and how transformer stacks compose these components into a complete model. This goes beyond high-level descriptions to explain the 'why' behind architectural choices and mathematical properties.
Provides rigorous mathematical treatment of transformer components with derivations of attention formulas, complexity analysis, and proofs of why certain design choices work, rather than treating transformers as black boxes. Integrates theory with implementation details showing how mathematics translates to code.
Deeper mathematical rigor than most online tutorials, with formal derivations comparable to research papers but presented pedagogically for learners rather than assuming expert background
llm application architecture patterns and system design
Medium confidenceTeaches architectural patterns for building production LLM applications, covering system design considerations, integration with existing systems, scalability patterns, and operational concerns. The curriculum covers different application architectures (simple prompting, RAG, agents, multi-model systems), how to structure applications for reliability and maintainability, and how to integrate LLMs with databases, APIs, and other services. This includes both high-level architectural patterns and practical implementation considerations.
Covers complete application architecture from high-level patterns through operational concerns, with explicit focus on production considerations and integration with existing systems. Treats LLM applications as complete systems rather than just adding an LLM to existing code.
More comprehensive than most LLM application guides, covering architectural patterns and system design while remaining more practical than academic software architecture research
llm training and fine-tuning methodology instruction
Medium confidenceTeaches practical and theoretical aspects of training large language models from scratch and fine-tuning pre-trained models, covering data preparation, tokenization strategies, loss functions, optimization algorithms, distributed training, and evaluation metrics. The curriculum explains how to structure training pipelines, handle different data formats, implement various fine-tuning approaches (full fine-tuning, LoRA, prompt tuning), and measure model performance. This includes both the mathematical foundations and practical implementation considerations for training at different scales.
Integrates theoretical understanding of training objectives with practical pipeline implementation, covering both classical training approaches and modern parameter-efficient methods (LoRA, adapters). Addresses infrastructure and scaling challenges specific to large models rather than treating training as a generic ML problem.
More comprehensive than framework-specific tutorials while remaining more practical than academic papers, with explicit guidance on computational trade-offs and modern techniques like parameter-efficient fine-tuning
prompt engineering and in-context learning techniques
Medium confidenceTeaches systematic approaches to prompt design, few-shot learning, chain-of-thought prompting, and in-context learning strategies that improve LLM performance without model retraining. The curriculum covers how to structure prompts for different tasks, leverage examples effectively, use intermediate reasoning steps, and combine multiple prompting techniques. This includes both empirical best practices and theoretical understanding of why certain prompting strategies work better than others for different model sizes and capabilities.
Combines empirical prompt engineering techniques with theoretical understanding of in-context learning, explaining both what works and why it works. Covers systematic approaches to prompt optimization rather than treating it as an art, including evaluation frameworks for measuring prompt effectiveness.
More systematic and theoretically grounded than most prompt engineering guides, while remaining practical and immediately applicable without requiring model retraining or fine-tuning
retrieval-augmented generation (rag) system design and implementation
Medium confidenceTeaches how to build RAG systems that augment LLM generation with retrieved context from external knowledge sources, covering document indexing, retrieval mechanisms, ranking strategies, and integration with generation models. The curriculum explains how to structure knowledge bases, implement semantic search, handle retrieval failures, and optimize the retrieval-generation pipeline. This includes both the architectural patterns for RAG systems and practical considerations for production deployment with large document collections.
Provides end-to-end RAG system design covering both retrieval and generation components, with explicit focus on production considerations like handling retrieval failures, ranking optimization, and latency management. Treats RAG as a complete system architecture rather than just adding a retrieval step to an LLM.
More comprehensive than framework-specific RAG tutorials, covering architectural patterns and trade-offs while remaining more practical than academic information retrieval papers
llm-based agent design and planning strategies
Medium confidenceTeaches how to design autonomous agents that use LLMs for reasoning and decision-making, including planning algorithms, tool use and function calling, memory management, and multi-step task decomposition. The curriculum covers different agent architectures (ReAct, chain-of-thought, hierarchical planning), how to structure tool definitions for function calling, and strategies for handling agent failures and loops. This includes both the theoretical foundations of planning and practical implementation patterns for building reliable agents.
Covers complete agent design including planning strategies, tool integration, and failure handling, rather than treating agents as simple LLM + tools combinations. Addresses practical challenges like loop detection, error recovery, and cost management specific to LLM-based agents.
More comprehensive than framework-specific agent tutorials, with explicit coverage of planning algorithms and reliability patterns while remaining more practical than academic planning research
llm evaluation, benchmarking, and metrics instruction
Medium confidenceTeaches how to evaluate LLM performance across different dimensions including accuracy, fluency, factuality, safety, and efficiency, covering both automatic metrics and human evaluation methodologies. The curriculum explains how to select appropriate benchmarks, design evaluation protocols, interpret results, and understand the limitations of different metrics. This includes coverage of standard benchmarks (GLUE, SuperGLUE, MMLU, etc.), task-specific metrics, and emerging evaluation challenges for large models.
Provides comprehensive evaluation methodology covering both automatic metrics and human evaluation, with explicit discussion of metric limitations and when different evaluation approaches are appropriate. Addresses evaluation challenges specific to large generative models rather than treating evaluation as a standard ML problem.
More thorough than most model evaluation guides, covering both standard benchmarks and emerging evaluation challenges while remaining more practical than academic evaluation research
llm deployment, optimization, and inference efficiency
Medium confidenceTeaches how to deploy LLMs in production environments with focus on inference optimization, latency reduction, and cost efficiency, covering quantization, distillation, batching strategies, caching, and hardware selection. The curriculum explains how to profile model performance, identify bottlenecks, implement optimization techniques, and measure trade-offs between quality and efficiency. This includes both software optimization techniques and hardware considerations for different deployment scenarios (cloud, edge, on-premise).
Covers complete deployment pipeline from profiling and optimization through production monitoring, with explicit focus on inference-specific challenges and trade-offs. Addresses both software optimization techniques and hardware selection rather than treating deployment as a generic ML problem.
More comprehensive than framework-specific deployment guides, covering multiple optimization techniques and hardware options while remaining more practical than academic optimization research
safety, alignment, and responsible llm development practices
Medium confidenceTeaches how to identify and mitigate risks in LLM systems including bias, toxicity, hallucination, and misuse, covering safety evaluation methodologies, alignment techniques, and responsible deployment practices. The curriculum covers red-teaming approaches, bias detection and mitigation, factuality verification, and ethical considerations in LLM development. This includes both technical safety measures and broader considerations for responsible AI deployment.
Integrates technical safety measures with broader ethical and responsible AI considerations, covering both detection and mitigation of safety risks. Addresses LLM-specific safety challenges rather than treating safety as a generic ML concern.
More comprehensive than most safety guides, covering technical evaluation methods alongside ethical frameworks while remaining more practical than academic AI ethics research
multimodal llm capabilities and vision-language model understanding
Medium confidenceTeaches how multimodal LLMs process and generate content combining text, images, and other modalities, covering vision encoders, cross-modal alignment, and applications like image captioning and visual question answering. The curriculum explains how vision-language models integrate visual and textual information, the architectures used for multimodal fusion, and how to leverage multimodal capabilities in applications. This includes both understanding existing multimodal models and considerations for building or fine-tuning multimodal systems.
Covers multimodal LLM architectures and applications with explicit focus on how vision and language components interact, rather than treating vision and language as separate problems. Addresses challenges specific to multimodal systems like cross-modal alignment and fusion.
More comprehensive than most vision-language model guides, covering both architecture understanding and application development while remaining more practical than academic multimodal learning research
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with 11-667: Large Language Models Methods and Applications - Carnegie Mellon University, ranked by overlap. Discovered automatically through the match graph.
llm-course
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
COS 597G (Fall 2022): Understanding Large Language Models - Princeton University

CS11-711 Advanced Natural Language Processing
in Large Language Models.
LLM Bootcamp - The Full Stack

AI-Systems (LLM Edition) 294-162
in AI System.
awesome-generative-ai-guide
A one stop repository for generative AI research updates, interview resources, notebooks and much more!
Best For
- ✓Computer science students and researchers building LLM expertise
- ✓ML engineers transitioning from traditional deep learning to large language models
- ✓Technical founders and architects designing LLM-based products
- ✓Teams evaluating LLM frameworks and needing theoretical grounding for decisions
- ✓Researchers developing novel transformer variants or architectures
- ✓ML engineers implementing transformers from scratch or optimizing existing implementations
- ✓Technical leaders making decisions about model architecture choices
- ✓PhD students and advanced practitioners in NLP/ML
Known Limitations
- ⚠Requires significant time investment (full semester course) — not suitable for quick reference or rapid prototyping
- ⚠Curriculum is fixed and may lag behind rapidly evolving LLM landscape (new architectures, training techniques)
- ⚠Primarily theoretical with limited hands-on coding exercises relative to lecture content
- ⚠No interactive sandboxes or live model experimentation environments embedded in course materials
- ⚠Heavy mathematical content may be challenging without strong linear algebra background
- ⚠Focuses on standard transformer architecture — limited coverage of recent variants (mixture-of-experts, sparse attention, etc.)
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
About

Categories
Alternatives to 11-667: Large Language Models Methods and Applications - Carnegie Mellon University
Are you the builder of 11-667: Large Language Models Methods and Applications - Carnegie Mellon University?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →