Multimodal Monday AI #37: Less-Is-More, Analogical Vision

Week of Dec 9-14, 2025: Apple proves one attention layer beats dozens in diffusion models, MokA shows low-rank adaptation outperforms full fine-tuning, Adobe's relsim captures analogical relationships between images, and X-VLA controls different robot types with one transformer.
Multimodal Monday AI #37: Less-Is-More, Analogical Vision

📢 Quick Hits (TL;DR)

AI understands analogies between images. Adobe’s new relsim metric helps models grasp deep visual relationships. A peach’s cross-section mirrors Earth’s layers. Your model can finally see these connections.

Apple proves one attention layer can beat dozens. Their new paper shows a single attention layer converts pretrained vision features into SOTA image generators. MokA shows you can adapt multimodal models with fewer parameters and better results.

AI gets a body (and a phone). X-VLA is a soft-prompted transformer that controls totally different robots like they’re one species. And AutoGLM, open-source phone agent framework, enables autonomous task completion on Android devices through natural language commands.

🧠 Research Highlights

Relational Visual Similarity

Adobe Research and UW-Madison developed relsim, a metric that captures analogical relationships between images rather than surface-level features. The model understands that a peach’s layers relate to Earth’s structure the same way a key relates to a lock.
Links: Paper

One Attention Layer is Enough

Apple demonstrates that a single attention layer transforms pretrained vision features into state-of-the-art image generators. This approach simplifies diffusion models while maintaining top-tier quality.
Links: Paper

Comparison between standard VAE (Rombach et al., 2022b), VA-VAE (Yao et al., 2025), RAE (Zheng et al., 2025) and their proposed FAE. The number shows the channel dimension of the generative modeling space.

Unison: A Fully Automatic, Task-Universal, and Low-Cost Multimodal Framework

Unison automates multimodal tasks across text, images, and video without task-specific retraining. The framework uses efficient fusion techniques that work across different modalities with minimal computational overhead.
Links: Paper

Overview of the Unison. The left shows the stage-one understanding model, the right shows the stage-two generation model. These two components are connected via a projector.

MokA: Multimodal Low-Rank Adaptation for MLLMs

MokA reveals that current multimodal fine-tuning wastes parameters and proposes a low-rank method that improves visual-language integration. The approach beats standard fine-tuning on visual grounding benchmarks while using fewer parameters.
Links: Paper

Evaluation results of LoRA, its variants, and MokA on visual-text benchmarks. #A and #B are the number of low-rank matrices. Here N=2 refers to the number of modalities.

Computational Emotion Analysis with Multimodal LLMs

Researchers evaluate how well multimodal LLMs analyze emotions from audio-visual data. The paper identifies gaps in current methods and shows where generative AI can improve affective computing.
Links: Paper

Relationship between video-based 3-shot emotion intensity scores of mLLMs and human coders’ average ratings in the RAVDESS dataset

WonderZoom: Generates multi-scale 3D worlds from text descriptions. Links: Paper

X-VLA: Soft-prompted transformer that controls different robot types using a unified interface. Links: Docs

MoCapAnything: Captures 3D motion for arbitrary skeletons from single-camera videos. Links: Paper

RouteRAG: Uses reinforcement learning to efficiently retrieve from both text documents and knowledge graphs. Links: Paper | GitHub

🛠️ Tools, Models and Techniques

AutoGLM

z.ai released AutoGLM, an open-source framework that completes tasks on Android phones through natural language commands. AutoGLM-Phone-9B is available on Hugging Face and ModelScope.

Why it matters: You can now build agents that interact with real-world smartphone apps.
Links: Website

Apriel-1.6-15B-Thinker

ServiceNow’s 15B multimodal model scores 57 on the Artificial Analysis Intelligence Index, matching the performance of 200B-scale models. The model handles reasoning tasks at a fraction of the size.

Why it matters: You get frontier-level multimodal reasoning in a model small enough to self-host.
Links: Model | Blog | Demo

GLM-4.6V

Z.ai released GLM-4.6V with tool-calling and 128K context window for vision-language tasks. The model handles multilingual development and API integration.

Why it matters: You get long-context multimodal processing with native tool support.
Links: Blog | GitHub | Demo

GPT-5.2

OpenAI released GPT-5.2, their latest frontier model. The model advances capabilities across reasoning, generation, and multimodal understanding.

Why it matters: The new baseline for frontier model performance just shifted.
Links: Twitter | Blog

DMVAE

Tencent and PKU released DMVAE, a VAE that matches latent distributions to any reference. The model achieves state-of-the-art image synthesis with fewer training epochs.

Why it matters: You can control the distribution of your generated images more precisely.
Links: Paper | Model

Qwen-Image-i2L: First open-source tool that converts a single image into a custom LoRA. Links: ModelScope | Code

Dolphin-v2: 3B parameter model that parses any document type. Links: Hugging Face

RealGen: Improves text-to-image photorealism using detector-guided rewards. Links: Website | Paper | GitHub | Models

Qwen 360 Diffusion: State-of-the-art 360° text-to-image generation. Links: Hugging Face | Viewer

Any4D: Unified transformer for feed-forward, dense, metric-scale 4D reconstruction. Links: Website | Paper | Demo

Shots: Generates 9 cinematic camera angles from a single image with perfect consistency. Links: Twitter

The Great Simplification

We’re watching complexity collapse in real time. Apple needs one attention layer where others use dozens. Unison handles any modality without retraining. MokA beats full fine-tuning with a fraction of the parameters.

This isn’t about making things smaller. It’s about removing the cruft that never belonged there.

What This Means for You

Your multimodal systems can run faster and cheaper. A single-layer model fits on hardware that was out of reach last year. You can index a billion images on a single GPU.

Your systems can become more adaptable. Unison adds new modalities without retraining. You can build applications that were too complex before.

Your models become more transparent. Low-rank adaptation means fewer parameters to debug. You can explain why your system retrieved a specific image.

This is multimodal AI for teams with real constraints. Not just for companies with unlimited compute budgets.

🧩 Community + Shoutouts

Nano Banana Pro Solution Shoutout to rob from ComfyUI for sharing a solution that addresses the biggest downside to Nano Banana Pro (the cost and slow generation speed). 1 prompt = 9 distinct images at 1K resolution (~3 cents per image). Links: Twitter

Multimodal Benchmarks GitHub Repo Shoutout to Mixpeek for releasing a Multimodal Benchmarks GitHub repo. A great resource for anyone who is working on multimodal AI. Links: GitHub

67 New Z Image LoRAs Shoutout to Reddit user malcolmrey for releasing 67 new Z Image LoRAs. A great resource for anyone who is using Z Image. Links: Reddit


That’s a wrap for Multimodal Monday #37! Apple proves one attention layer generates SOTA images. Adobe’s relsim captures analogies between images. MokA beats full fine-tuning with low-rank adaptation. RouteRAG navigates knowledge graphs with reinforcement learning. X-VLA controls different robots with one transformer. AutoGLM completes tasks on Android phones. Apriel-1.6-15B matches 200B-scale models at 15B parameters.

Ready to build multimodal solutions that actually work? Let's talk

About the author
Philip Bankier

Philip Bankier

Philip’s expertise in RAG, data engineering, and AI agents makes him a natural fit to write about scaling multimodal data warehousing systems.

Mixpeek Engineering Blog

Deep dive into multimodal AI, data processing, and best practices from our engineering team.

Great! You’ve successfully signed up.

Welcome back! You've successfully signed in.

You've successfully subscribed to Mixpeek Engineering Blog.

Success! Check your email for magic link to sign-in.

Success! Your billing info has been updated.

Your billing was not updated.