- The Rundown AI
- Posts
- Google's ultra-efficient Gemma 3
Google's ultra-efficient Gemma 3
PLUS: Sakana's peer-reviewed AI-authored paper
Good morning, AI enthusiasts. The era of massive compute requirements for cutting-edge AI may be coming to an end, with Google’s latest open-source release outperforming giants at just a fraction of the size.
With Gemma 3’s high-level performance, multimodal capabilities, and on-device operation on just a single GPU, the AI efficiency barrier is quickly getting destroyed.
In today’s AI rundown:
Google’s Gemma 3 for single-GPU deployment
Gemini Flash with new image capabilities
Create your AI-powered Telegram assistant
Sakana’s peer-reviewed AI-authored paper
4 new AI tools & 4 job opportunities
LATEST DEVELOPMENTS

Image source: Google
The Rundown: Google just unveiled Gemma 3, a new family of lightweight AI models built from the same technology as Gemini 2.0 — delivering performance that rivals much larger models while running efficiently on just a single GPU or TPU.
The details:
The model family comes in four sizes (1B, 4B, 12B, and 27B parameters) optimized for different hardware configurations from phones to laptops.
The 27B model outperforms larger competitors like Llama-405B, DeepSeek-V3, and o3-mini in human preference evaluations on the LMArena leaderboard.
Other new capabilities include a 128K token context window, support for 140 languages, and multimodal abilities to analyze images, text, and short videos.
Google also released ShieldGemma 2, a 4B parameter image safety checker that can filter explicit content — with easy integration into visual applications.
Why it matters: Gemma 3’s performance is mind-blowing, beating out top-level systems that dwarf it in both size and compute. Running on just a single GPU, these models hit a once incomprehensible sweet spot of being open-source, powerful, fast, multimodal, and small enough to be deployed across devices — a truly massive feat.
TOGETHER WITH DAGSTER
The Rundown: Dagster consolidates your AI capabilities into one powerful orchestrator that developers love — helping reduce costs, eliminate complexity, and ensure reliable pipelines from prototype to production.
With Dagster, you can:
Consolidate all AI capabilities under one intuitive interface
Save 40%+ on infrastructure costs by optimizing AI workloads
Ship AI features 3x faster with standardized development practices

Image source: Google
The Rundown: Google released new experimental image-generation capabilities for its Gemini 2.0 Flash model, letting users upload, create, and edit images directly from the language model without requiring a separate image-generation system.
The details:
A 2.0-flash-exp model is available via API and in the Google AI Studio with support for both image and text outputs and editing via text conversation.
Gemini uses reasoning and a multimodal foundation to maintain character consistency and understand real-world concepts throughout a conversation.
For instance, you can prompt it to generate a story with pictures and then guide it to the perfect version through natural dialogue.
Google says Flash 2.0 also excels at text rendering compared to competitors, allowing for ads, social posts, and other text-heavy design generations.
Why it matters: This upgrade is a major step in shifting how AI generates visual content — moving away from dedicated image models toward language models that natively understand both text and visuals. Just as natural language prompting has taken over other domains, image editing appears to be next on the list.
AI TRAINING

The Rundown: In this tutorial, you will learn how to build a personal AI assistant on Telegram that can answer questions, remember conversations, and eventually connect to other services using n8n's automation platform.
Step-by-step:
Create a Telegram bot - search for "BotFather" in Telegram, type /newbot, and save the API token you receive.
Sign up for n8n (free 14-day trial) and create a new workflow with a Telegram trigger using your bot token.
Add an AI Agent node after the trigger, connect it to your preferred AI model (like those from OpenAI), and use the message text as the prompt.
Add a Telegram "Send Message" node to return the AI's responses to your chat using the chat ID from the trigger.
Enable Window Buffer Memory in the AI Agent settings so your bot remembers previous conversations.
Pro tip: You can also expand the assistant’s capabilities by connecting it to calendars, emails, notes apps, and other services. We did an extensive workshop on how to create your own AI Agent to automate tasks with n8n here.
PRESENTED BY JOTFORM
The Rundown: Jotform AI Agents let organizations provide 24/7, conversational customer service across multiple platforms — no coding required.
With Jotform AI Agents, you can:
Get started easily with over 7,000 ready-to-use AI agent templates
Automate workflows and trigger custom actions in real time
Handle voice, text, and chat inquiries seamlessly
Customize your agent’s look and feel to align with your brand identity
SAKANA AI

Image source: Sakana AI
The Rundown: Japanese AI startup Sakana announced that its AI system successfully generated a scientific paper that passed peer review, with the company calling it the first fully AI-authored paper to clear the scientific bar.
The details:
AI Scientist-v2 generated three papers, creating the hypotheses, experimental code, data analyses, visualizations, and text without human modification.
One submission was accepted at the ICLR 2025 workshop with an average reviewer score of 6.33, ranking higher than many human-written papers.
Sakana also pointed out some caveats, including the AI making citation errors and workshop acceptance rates being higher than typical conference tracks.
The company concluded that the paper did not meet its internal bar for ICLR conference papers but displayed “early signs of progress.”
Why it matters: While this milestone comes with significant asterisks, it also represents a major early marker of AI's advancing role in academic research processes. Between models like Sakana’s and Google’s AI co-scientist, a seismic shift is getting closer and closer for the scientific world.
QUICK HITS
⚙️ Responses API and Agents SDK - OpenAI’s DIY tools for custom agents
⚡️ Reka Flash 3 - Open, 21B parameter reasoning AI for on-device deployment
👨🏻⚖️ Harvey - AI for law firms, service providers, and Fortune 500 companies
🗣️ Wispr Flow for Windows - Use voice to write 3x faster in every application
New legal filings revealed that Google owns 14% of Anthropic, with its investments totaling over $3B in the rival AI startup.
Alibaba researchers open-sourced R1-Omni, a new multimodal reasoning model that can ‘read’ emotions using visual and audio context.
Google DeepMind introduced Gemini Robotics and Gemini Robotics-ER, two Gemini 2.0-based models to help robots accomplish real-world tasks without training.
Perplexity launched a new Model Context Protocol (MCP) server for its Sonar model, allowing Claude to access real-time web search capabilities.
Snap released its first AI Video Lenses, powered by its own in-house model, offering premium subscribers three AR animations with new options planned to launch weekly.
Moonvalley released Marey, an AI video model that claims to be trained exclusively on licensed content for use in filmmaking — capable of creating 30-second-long HD clips.
Captions unveiled Mirage, a foundation model designed specifically for generating UGC-style content for ad campaigns.
COMMUNITY
Join our next workshop this Friday at 4 PM EST to learn about how to use Manus AI to boost productivity and automate everyday tasks easily, with Dr. Alvaro Cintas, The Rundown’s AI professor.
RSVP here. Not a member? Join The Rundown University on a 14-day free trial.
We’ll always keep this newsletter 100% free. To support our work, consider sharing The Rundown with your friends, and we’ll send you more free goodies.
That's it for today!Before you go we’d love to know what you thought of today's newsletter to help us improve The Rundown experience for you. |
See you soon,
Rowan, Joey, Zach, Alvaro, and Jason—The Rundown’s editorial team
Reply