
Building Multimodal Data Pipelines
Instructor: Gilberto Hernandez
- Intermediate
- 1 Hour 1 Minute
- 7 Video Lessons
- 3 Code Examples
- Instructor: Gilberto Hernandez
What you'll learn
Extract structured, queryable information from unstructured images, audio, and video using OCR, Automatic Speech Recognition (ASR), and Vision Language Models (VLMs).
Build a VLM-backed pipeline that reasons across video frames to generate timestamped scene descriptions and track events over time.
Implement a multimodal RAG application on a real-world dataset, taking raw images, audio, and video into a fully queryable interface with grounded, cited answers.
About this course
Images, audio, and video make up a growing share of the data companies generate today, but most pipelines are still built for structured data alone. This course teaches you to build AI-powered pipelines that process multimodal data and turn it into LLM-ready text.
You’ll start with the foundations: using ASR to extract transcripts from audio and turning images into LLM-ready text descriptions. From there, you’ll see how Vision Language Models generate descriptions from video segments, capturing not just what’s visible in a single frame, but what unfolds across a scene over time. You’ll then apply these skills to implement a multimodal RAG pipeline that searches across slides, audio, and video from meetings to answer questions about their content. By combining all three modalities, you give LLMs the rich context they need to deliver detailed answers across complex, real-world content.
In detail, you’ll:
- Survey the multimodal data landscape, the unique challenges each data type presents, and the techniques that transform unstructured content into searchable text.
- Apply OCR and ASR to convert images and audio into structured text, then embed them into a unified vector space for cross-modal semantic search.
- Prompt Vision Language Models effectively, and choose the right frame sampling and embedding strategy for video.
- Run a Vision Language Model on meeting videos to generate timestamped segment descriptions, then embed them alongside audio and slides for unified semantic, and time-based search.
- Build a multimodal RAG system that retrieves across audio, slides, and video to generate grounded, cited answers from meeting recordings.
Every technique you’ll learn serves the same goal data engineers have always had: take messy, unstructured data and turn it into something you can query, analyze, and build on.
Who should join?
Data engineers and ML practitioners who want to extend their pipelines beyond structured data to handle images, audio, and video. Familiarity with Python, SQL queries and basic data engineering concepts is recommended.
Course Outline
7 Lessons・3 Code ExamplesIntroduction
Video・2 mins
Multimodal Data Overview
Video・7 mins
Automatic Transcription, OCR, and Semantic Search
Video with code examples・16 mins
Processing Video with a VLM
Video・7 mins
Building a VLM‐Backed Pipeline
Video with code examples・8 mins
Multimodal RAG System
Video with code examples・9 mins
Conclusion
Video・1 min
Quiz
Reading・10 mins
Instructor
Building Multimodal Data Pipelines
- Intermediate
- 1 Hour 1 Minute
- 7 Video Lessons
- 3 Code Examples
- Instructor: Gilberto Hernandez
Additional learning features, such as quizzes and projects, are included with DeepLearning.AI Pro. Explore it today
Want to learn more about Generative AI?
Keep learning with updates on curated AI news, courses, and events, as well as Andrew’s thoughts from DeepLearning.AI!

