Part I: Why AI Changes Product Creation
Chapter 2

What AI Can and Cannot Reliably Do

2.2 Multimodal Abilities and Their Boundaries

Objective: Understand the capabilities and limitations of multimodal AI models and how to design products that leverage visual, audio, and other modality inputs.

"A picture is worth a thousand words, but an AI that cannot see pictures is blind to the world most users inhabit."

The Multimodal Product Guide

2.2 Multimodal Abilities and Their Boundaries

Multimodal AI systems can process and generate content across multiple modalities: text, images, audio, video, and even sensor data. Understanding what these systems can and cannot do reliably is essential for designing products that leverage multimodal capabilities.

The Multimodal Landscape in 2026

Modern foundation models have expanded beyond text to handle multiple modalities. Each modality has distinct reliability characteristics that product teams must understand.

Text processing is the most mature capability. Models have been trained on massive text corpora and can handle a wide range of text tasks with high reliability. Translation, summarization, classification, and generation all work well for text.

Vision models can analyze images, extract information, describe content, and even generate images from text descriptions, though reliability varies significantly based on the specific task and the types of images involved. Complex scenes, unusual viewpoints, and domain-specific images may have lower reliability.

Speech recognition and synthesis are now highly reliable for common languages and speaking styles, with transcription services achieving near-human accuracy for clear audio in standard conditions, though reliability drops for accented speech, noisy environments, or specialized terminology.

Video understanding combines visual and temporal reasoning, making it more challenging. Models can analyze video content, but reliability is lower than for individual frames and long videos present context window challenges similar to long text documents.

Visual Understanding Capabilities

Vision-language models can analyze images, extract visual information, and even generate images from text descriptions. Understanding the capabilities and limitations of visual AI helps you design better product experiences.

Vision-language models can analyze images, extract visual information, and even generate images from text descriptions. They reliably recognize common objects and animals, describe scenes in natural language, read text within images through OCR, identify faces and basic facial expressions, detect common UI elements in screenshots, and compare two images for similarity or differences. These capabilities are powerful but come with clear limitations.

Vision models frequently miss subtle findings in medical imaging that experienced radiologists would catch, so domain-specific fine-tuning and extensive validation against human experts is required for medical imaging products along with regulatory compliance. Modern models can generate high-quality images from text descriptions, which is useful for creating illustrations and diagrams for content, generating product mockups from descriptions, creating variations of existing images, and editing and enhancing photos based on instructions. Image generation reliability varies significantly based on the complexity of the request and the specificity of the subject. Simple, well-documented concepts generate reliably while complex or novel compositions may require multiple attempts.

Audio and Speech Capabilities

Audio processing includes both speech recognition (transcription) and speech synthesis (text-to-speech). Both have reached high reliability for standard use cases but have limitations in edge cases.

Audio processing includes both speech recognition (transcription) and speech synthesis (text-to-speech). Both have reached high reliability for standard use cases but have limitations in edge cases. For speech recognition, reliability is high for clear audio, standard accents, quiet environments, and formal vocabulary. Reliability is moderate for background noise, accents, informal language, and domain-specific terminology. Reliability is low for multiple speakers overlapping, very noisy environments, and rare languages or dialects.

For text-to-speech, modern systems produce natural-sounding speech, but emotional expressiveness is limited compared to human speech, complex or unusual names may be mispronounced, long-form synthesis may have inconsistencies in voice quality, and some languages have less developed voice options than English.

True Multimodal Understanding

The most powerful multimodal capabilities come from reasoning across modalities. A model that can look at a diagram and explain it, or listen to instructions and generate the corresponding image, demonstrates true multimodal understanding. This enables:

Visual question answering, generating text descriptions from charts or data visualizations, creating images from audio descriptions, transcribing and summarizing video content, and converting infographics to natural language explanations.

Designing with Multimodal AI

Product teams should design multimodal features with realistic expectations about reliability. The following guidelines help create products that leverage multimodal capabilities effectively.

Multimodal Product Design Guidelines

1. Match Modality to Task

Use the modality that best fits the task. Asking users to describe something in text is often more reliable than asking them to upload an image, because text provides more controlled input. Conversely, complex visual information may be better conveyed through images than text descriptions.

2. Handle Modality Failures Gracefully

Always have fallback options when modality processing fails. If image upload fails, allow text description. If speech recognition fails, offer keyboard input.

3. Show Uncertainty

Multimodal processing can fail silently. Show confidence indicators when the system is uncertain, and make it easy for users to correct errors.

4. Test with Real Data

Multimodal reliability varies significantly across domains. Test extensively with your actual user data, not just synthetic examples.

Eval-First in Practice

Before selecting a modality for your AI feature, define how you will measure cross-modal reliability. A micro-eval for multimodal products tests: image-to-text accuracy across your specific domain, text-to-image fidelity for your use case, and audio transcription error rates for your user base's speech patterns. RetailMind's eval-first insight: they measured cross-modal recommendation accuracy before launch and discovered their fashion domain had 40% lower reliability than their electronics domain, requiring domain-specific fine-tuning.

Contextual Integrity and Privacy

Multimodal AI introduces additional privacy considerations. Processing images and audio requires careful handling of potentially sensitive data.

Privacy Considerations for Multimodal AI

Image data presents unique privacy challenges because face detection and recognition capabilities raise significant concerns, requiring products to clearly disclose when and how facial data is processed. Audio data similarly requires careful handling since transcription services process speech that may contain personal information, and users should understand how their audio is stored and used. Cross-modal linking compounds these concerns because combining visual and audio data can reveal information users did not explicitly provide, such as when a photo combined with audio context might identify individuals or locations.

Running Product: HealthMetrics Analytics

HealthMetrics built a clinical decision support system that processes multiple modalities: patient notes (text), medical images (X-rays, CT scans), and vital signs (time-series data).

During eval, they discovered significant reliability differences across modalities. Text processing had 94% accuracy, medical images had 87% accuracy for common conditions but only 62% for rare conditions, and vital sign time-series had 91% accuracy. Simply combining modalities without accounting for their different reliability levels led to overconfident错误的 conclusions.

They fixed this by weighting each modality's contribution by its measured reliability, using confidence-weighted fusion. When the image model was uncertain about a rare condition, its contribution was downweighted even if the text model was confident. Patient outcomes improved 23% after implementing reliability-weighted fusion.

What's Next?

Next, we explore Reasoning and Tool Use Limits, understanding how AI systems can use external tools and APIs, and where those capabilities break down.