MLflow 3.11 introduces multimodal tracing, which automatically extracts binary content (images, audio, PDFs) from AI agent traces instead of storing raw base64 strings in the trace database. Binary data is offloaded to artifact storage (S3, Azure Blob, GCS, etc.) while spans retain only lightweight reference URIs. The feature works out of the box with autologging for OpenAI, Anthropic, Gemini, Bedrock, and LangChain, covering 8 multimodal data patterns. A manual Attachment API handles custom file types. The MLflow UI renders images as thumbnails, audio with inline playback controls, and PDFs in an embedded viewer, enabling proper debugging of vision and audio model behavior.

4m read timeFrom mlflow.org
Post cover image
Table of contents
Why Text-Only Traces Fall Short ​How It Works ​Auto-Extraction: Zero Code Changes ​Manual Attachments for Custom Content ​Rich Rendering in the Trace UI ​Getting Started ​

Sort: