Language models sit at the center of public attention, but “language” in industry increasingly means text plus whatever else you can tokenize: images, audio, video frames, sensor traces, and tool outputs fed back into context. This hub describes how those pieces hang together without pretending that a single architecture solves every modality. The foundational explainer on this site remains Large Language Models: Probability, Not Personhood—read it first if you have not yet internalized autoregression and finite context windows.

One objective, many surfaces

Whether the interface is chat, code completion, or a vision-language assistant, the recurring pattern is the same: a model assigns mass to next pieces conditioned on prior context. Multimodal systems typically fuse representations early or late; the engineering trade-off is between tight integration (harder to debug) and modular pipelines (easier to swap components, harder to keep coherent). Neither choice removes the need for Optimization & learning literacy: you are still walking down a loss landscape shaped by data and architecture.

Retrieval, tools, and the boundary of the model

When people say “grounding,” they often mean one of three different things: (1) retrieving documents and conditioning on them, (2) calling APIs or calculators, or (3) training-time objectives that encourage faithfulness. Each has different failure modes—stale corpora, injection attacks through retrieved text, or confident hallucinations when tools return empty results. Our Deployment reality hub discusses how to operationalize logging and rollback when these paths break in production.

Benchmarks as narrow probes

Leaderboards reward narrow skills: multiple-choice knowledge, short-form reasoning, coding puzzles. Real work blends tasks, tolerates ambiguity, and punishes subtle factual drift. A high score is a hint, not a certificate—especially when vendors tune on test-adjacent material. Cross-check claims with the checklist in Alignment: Goals, Feedback, and Failure Modes, which stresses separating demo fluency from verified reliability.

Safety sits downstream of capability

Capable language interfaces raise misuse and accident rates simultaneously. Policy discussions belong in Safety & alignment; the through-line is simple: the same model that helps a clinician summarize notes can assist someone harmful if guardrails, access control, and monitoring are absent. Technical mitigations and governance are both required—see the alignment article for a non-alarmist framing.

Speech, vision, and other modalities: adapters and fusion

Audio and image encoders map raw signals into embedding spaces that can align with text—either through joint pre-training on paired data or through adapters bolted onto frozen backbones. Video adds time: memory and compute grow quickly with resolution and frame rate. Teams often start with APIs and off-the-shelf encoders, then optimize latency and cost once product–market fit is clearer. The same deployment discipline applies: latency SLOs, fallbacks when a modality fails, and user-visible error messages that do not blame “the AI” abstractly.

Cost, latency, and the product surface

Long contexts and multimodal stacks are expensive. Caching, prompt compression, selective retrieval, and smaller specialist models for sub-tasks are engineering responses—not afterthoughts. When pricing features, tie them to measurable outcomes: tickets deflected, drafts shortened, code review time reduced. Otherwise you risk shipping a fascinating demo that finance cannot sustain. For optimization fundamentals that underpin cost trade-offs, revisit How Neural Networks Actually Learn.

Suggested reading order

  1. How Neural Networks Actually Learn — if you need the optimization vocabulary.
  2. Large Language Models: Probability, Not Personhood — core LLM mechanics.
  3. Alignment: Goals, Feedback, and Failure Modes — objectives, feedback, and misuse.