Automatic Speech Recognition: Build vs. Buy
Learn how to build a scalable ASR pipeline using Ray and Whisper, with batching, GPU optimization, and real-world tips from production deployments
Learn how to build a scalable ASR pipeline using Ray and Whisper, with batching, GPU optimization, and real-world tips from production deployments
Multimodal Monday Week 10: Xiaomi's 7B model outperforms GPT-4o, Ming-Omni unifies all modalities with 2.8B params, and specialized efficiency beats raw scale. The AI landscape is shifting fast.
A weekly pulse on everything multimodal—models, data, tools & community.
A weekly pulse on everything multimodal—models, data, tools & community.
Why the future of AI isn’t about bigger models — it’s about better data.
A weekly pulse on everything multimodal—models, data, tools & community.
Late interaction models enable precise retrieval from multimodal data like PDFs and images by comparing query tokens with token or patch embeddings—ideal for RAG, search, and document understanding.
A weekly pulse on everything multimodal—models, data, tools & community.
A weekly pulse on everything multimodal—models, data, tools & community. 🎯 Quick Take (TL;DR) * OpenAI GPT-Image 1 arrives in
Segmentation turns raw video into searchable chunks—objects, actions, scenes—boosting precision in multimodal search. It bridges unstructured content with structured queries like “man falling in warehouse,” enabling faster, more accurate retrieval across large datasets.
Visual CoT, video gen, and color benchmarks highlight this week's multimodal AI leaps—plus tools, papers, and real-world use cases.
CraftFlow uses Mixpeek to extract talk patterns, auto-score calls, surface objections, and summarize key sales moments with AI-powered embeddings.
Apple’s new scaling law research redefines how multimodal models are built, while Moonshot and OpenGVLab drop powerful open-source VLMs with reasoning and tool-use.
Even with data-driven targeting, most ads still miss. Contextual AI changes that—boosting relevance, clicks, and ROI without cookies.
As Google phases out third-party cookies, advertisers face declining performance from behavioral targeting. Learn how Contextual AI offers a privacy-safe, high-precision alternative.
📢 Quick Take (TL;DR) * Major multimodal model releases: Meta unveiled Llama 4 Scout & Maverick – open Mixture-of-Experts models with native
By applying the classic group_by pattern to structured video data at index time, you can turn raw frames into searchable, analyzable DataFrames aligned with how your users explore footage.
Researchers introducing new methods to replace embeddings with discrete IDs for faster cross-modal search
Intelligent video chunking using scene detection and vector embeddings. This tutorial covers how to break down videos into semantic scenes, generate embeddings, and enable powerful semantic search capabilities.
AI video tagging used to mean manual review and basic object detection. With multimodal models and dynamic taxonomies, you can now automatically detect brand moments, inappropriate content, actions, moods and trending content at scale.
This guide will walk developers through building a modern Media Asset Management (MAM) system with semantic search capabilities using Mixpeek's infrastructure.
AI-powered image discovery app using Mixpeek's multimodal SDK and MongoDB's $vectorSearch. Features deep learning, vector embeddings, and KNN search for advanced visual content management.
Build a scalable MCP pipeline on S3 using AWS Lambda, Temporal, Ray, and Qdrant to process and index unstructured data like video, audio, and PDFs for real-time AI search and retrieval.
This article demonstrates how to build a reverse video search system using Mixpeek for video processing and embedding, and Weaviate as a vector database, enabling both video and text queries to find relevant video segments through semantic similarity.
Deep dive into multimodal AI, data processing, and best practices from our engineering team.