MLflow 3.11 introduces multimodal tracing, which automatically extracts binary content (images, audio, PDFs) from AI agent traces instead of storing raw base64 strings in the trace database. Binary data is offloaded to artifact storage (S3, Azure Blob, GCS, etc.) while spans retain only lightweight reference URIs. The feature works out of the box with autologging for OpenAI, Anthropic, Gemini, Bedrock, and LangChain, covering 8 multimodal data patterns. A manual Attachment API handles custom file types. The MLflow UI renders images as thumbnails, audio with inline playback controls, and PDFs in an embedded viewer, enabling proper debugging of vision and audio model behavior.
Table of contents
Why Text-Only Traces Fall Short How It Works Auto-Extraction: Zero Code Changes Manual Attachments for Custom Content Rich Rendering in the Trace UI Getting Started Sort: