• Everyday AI
  • Posts
  • Ep 751: Hands on with Google’s Gemma 4: How to Use The Open Source Model Locally and Why It Matters

Ep 751: Hands on with Google’s Gemma 4: How to Use The Open Source Model Locally and Why It Matters

Breaking: Meta's new Muse Spark AI model released, Anthropic's Mythos Raises Security Alarms, New: Claude Managed Agents and our hands-on with Gemma 4.

 

Outsmart The Future

Today in Everyday AI
8 minute read

🎙 Daily Podcast Episode: Running powerful AI locally for free used to sound impossible, but Google’s new Gemma 4 model is making it a reality. Give today’s show a watch/read/listen to learn more.

🕵️‍♂️ Fresh Finds: Alibaba launched a new AI data center, Gemma 4 can now run locally with OpenClaw, and CapCut released Dreamina Seedance 2.0, and more. Read on for Fresh Finds.

🗞 Byte Sized Daily AI News: Breaking: Meta's new Muse Spark AI model released, Anthropic's Mythos Raises Security Alarms, New: Claude Managed Agents and more. Read on for Byte Sized News.

💪 Leverage AI: Your company is likely paying millions for AI that can now run locally for free. Keep reading for that!

↩️ Don’t miss out: Miss our last newsletter? We covered: Anthropic shocks with Mythos drop, Broadcom, Anthropic and Google struck a major AI chip deal, OpenAI is pushing for an investigation into Elon Musk and more. Check it here!

Ep 751: Hands on with Google’s Gemma 4: How to Use The Open Source Model Locally and Why It Matters


Get this: there's a FREE open source model that runs on consumer hardware you can use TODAY that's as powerful as frontier AI models from 14 months ago. 😱

Are we living in the future? Thanks to Google's new Gemma 4, we apparently are.

Want to have a frontier level AI agent running for you around the clock?

Or trying to lighten your API bill?

Or, maybe you just want a more private and offline model?

Also on the pod today:

• Gemma 4: local AI revolution 💻
• 31B parameters vs giants 🤖 
• Apache 2.0 license: use anywhere 📝 

It’ll be worth your 41 minutes:

Listen on our site:

Click to listen

Subscribe and listen on your favorite podcast platform

Listen on:

Here’s our favorite AI finds from across the web:

New AI Tool Spotlight – Velo takes raw recordings and makes them awesome with AI, Flint Makes the perfect landing in Claude and your favorite GTM tools, Marble lets you create, edit and share high-fidelity, persistent 3D worlds.

Alibaba Launches Data Center — Alibaba just opened a data center powered by 10,000 of its own chips as China steps up its AI game.

Gemma 4 x OpenClaw — You can run Google’s new Gemma 4 model locally in minutes using Ollama and OpenClaw. Find out how to get a powerful AI agent up and running on your own device fast.

Dreamina Seedance 2.0 Released — CapCut just dropped Dreamina Seedance 2.0 in the US, with a free trial and huge Pro discount for new users.

Spotify New Podcast Tool — Spotify’s AI-powered Prompted Playlists can now make custom podcast playlists, not just music.

HappyHorse-1.0 Tops AI Charts — HappyHorse-1.0 just shot to the top spot for AI video, outpacing Seedance 2 with impressive multi-shot and prompt-following skills.

OpenAI Retiring Models — OpenAI is retiring several older Codex models on April 14 for ChatGPT sign-ins, but you can still access them via API.

Prism Updates — Prism just launched Paper Review, an AI-powered tool that checks scientific papers for rigor and accuracy, not just grammar

WorldLabs Model Updates — World Labs just dropped major Marble model upgrades, letting you turn photos or floor plans into huge, VR-ready 3D worlds. Curious how it works?

Grok new Tools — X just rolled out a revamped photo editor with blur and text tools, plus Grok-powered image editing.

Meta Drops new Model — Meta just dropped its new Muse Spark AI model, shaking up the Llama family.

1. Meta Unveils Muse Spark AI Model, Takes on OpenAI and Anthropic

Meta just launched Muse Spark, its latest AI model, which the company says finally closes much of the gap with top competitors’ technology.

The nine-month project, led by Alexandr Wang, will immediately power Meta’s AI app and website, with plans to roll it out across Facebook, Instagram, and WhatsApp. Unlike rivals, Muse Spark offers specialized modes like “shopping mode” and combines AI with user interests to personalize results, though it only outputs text.

2. Anthropic's Mythos AI Raises Security Alarms🚨

Anthropic has unveiled the wild capabilities of its new Claude Mythos Preview model, which shocked testers with cunning tactics like exploiting internet restrictions, manipulating grading AIs, and hiding its own tracks.

The company is only letting select tech and cybersecurity partners try it out, citing the need for stronger security measures as AI models grow more powerful. This limited-access approach could signal a new trend in how AI giants roll out advanced systems, especially as OpenAI gears up for a similar release.

3. OpenAI Unveils Child Safety Blueprint to Combat AI-Driven Exploitation 📜

OpenAI has just released a new policy blueprint aimed at tackling the growing challenge of AI-powered child sexual exploitation, drawing on insights from law enforcement, child safety advocates, and tech leaders.

The blueprint calls for updated laws, tighter industry coordination with authorities, and safety-by-design features in AI systems to stay ahead of evolving threats. With backing from the National Center for Missing & Exploited Children and key state attorneys general, the plan signals a major push for stronger, shared standards across the tech industry.

4. SpaceX’s AI Colossus 2 Kicks Off Massive Multi-Model Training 😮‍💨

Elon Musk just revealed on X that SpaceX’s AI Colossus 2 is now training seven models at once, including the much-anticipated 10T model, which will spend about two months in its pre-training phase.

This surge in simultaneous training signals a new level of speed and ambition in AI development, as teams race to push boundaries in both model size and capability. Industry watchers are buzzing over how the largest model will perform and what this leap means for the future of synthetic and real-world data.

5. Intel Partners with Musk’s Terafab AI Chip Project 💻

Intel just announced it’s teaming up with Elon Musk’s Terafab AI chip initiative alongside SpaceX and Tesla, shaking up the semiconductor scene as Musk pushes for next-gen robotics and data center tech. T

he move sent Intel’s stock climbing and marks a pivotal moment in the company’s turnaround, signaling renewed investor confidence. Musk’s vision includes two massive chip factories in Austin, Texas, aimed at powering autonomous vehicles, humanoid robots, and space-based data centers.

6. Arena Drops Full AI Leaderboard History as Open Dataset 📊

In a major move for transparency, Arena has released the full history of its AI leaderboards as a public dataset, covering almost three years of frontier model rankings across multiple domains.

This release, now available on Hugging Face, lets researchers and the curious alike track how AI models have evolved and compare open source versus proprietary dominance in different arenas. The data is sliced by task and time, making it easy to spot trends and shifts in the fast-changing AI landscape.

7. Anthropic Rolls Out Claude Managed Agents for Rapid AI Deployment 🦾

Anthropic has just unveiled Claude Managed Agents, a new suite of cloud APIs designed to help developers launch AI-powered agents much faster, slashing development times from months to days.

The platform takes care of all the messy backend work like security, sandboxing, and scaling, so teams can focus on building features and user experience instead of infrastructure headaches. The public beta is now live on the Claude Platform, signaling a major push to make enterprise-grade AI agents accessible to more companies, more quickly.

Your company is prolly paying 5-7 figures PER MONTH for AI that a $2,200 laptop can now run for free.

Completely free. Forever. 

(And sorry…. your competitors who figure this out before you do are gonna have a real fun time at the next board meeting.)

Google DeepMind dropped Gemma 4 and the entire cost equation for enterprise AI just broke open. 

But here’s the blunt truth: remember those banger models from like 14 months ago? GPT-4o and laude 3.5 Sonnet?

Yeah, Gemma 4 is free, local, offline and benchmarks better than those frontier faves from 2025.  

So we showed you the ins and outs and the whys on today’s edition of Working Wednesday on Everyday AI. 

Time to capitalize shorties.

1. Frontier-Ish AI Performance Is Now Free 🔥

Fourteen months ago, GPT-4o was reportedly built on two trillion parameters. The undisputed best general-use AI in the world.

Google's new Gemma 4 31B model just scored 1,452 on Arena. That's the blind taste-test leaderboard where real users compare model outputs head to head with no brand labels. Fifteen months ago? That score would've TOPPED the entire global rankings.

Wild.

We're talking 31 billion parameters punching twenty times above its weight class. A mid-range MacBook Pro at roughly $2,200 runs the 26B Mixture of Experts version. The 31B dense model needs about 32GB of RAM (48-64 is better). That's it. That's the whole setup. And a few mouse clicks.

Google released Gemma 4 under Apache 2.0 licensing. Zero restrictions. Full commercial freedom. Build products, sell them, deploy in your enterprise stack. No vendor lock-in, no licensing fees, no usage caps after download. Build something and sell it. Legally. Forever.

Try This

Go to ollama.com right now and download Ollama. It's free, installs in two minutes, and gives any local model a ChatGPT-style chat interface. No terminal skills required.

Once it's installed, search "gemma4," grab the CLI command for your hardware, paste it into Ollama, and let it download. The 31B model is about nine gigabytes and takes roughly five minutes. Most mid-range MacBook Pros can run the 26B version out of the box.

Open it up. Ask it something hard. It's free forever, no account, no subscription, no card required.

2. Run AI Agents 24/7 For Zero Dollars ⚡

Agentic AI around the clock used to mean a serious API bill. Not anymore.

Anthropic recently told users they can no longer run OpenClaw through a standard Claude subscription. You gotta pay via API now. When agents run continuously, that cost stacks fast.

Gemma 4 was built SPECIFICALLY for this. Native function calling and structured JSON output baked directly in. Not just chat. Actual autonomous agents that execute tasks, call tools, and chain multi-step logic without a human anywhere near the loop.

Point your local agents at Gemma 4 running on your own machine. Zero dollars. Zero API keys. Zero usage caps. For fifty to eighty percent of routine agentic use cases, this model is gonna be good enough to fully replace what you're currently paying for.

The math on this is cooked in your favor. You'd be wild to ignore it.

Try This

Already running OpenClaw or a local agent on an Ollama-compatible endpoint? Swap your model config to point at Gemma 4 today. Takes about two minutes.

Update your agent config to localhost:11434 and set the model name to gemma4. Run a real task you'd normally pay for. Compare the output quality against whatever you're currently running.

For routine agentic work, it's prolly good enough to stop paying for that API right now. Start with one workflow, benchmark it yourself, then scale from there.

3. Private, Offline AI Is Now Actually Good 🚀

Healthcare. Legal. Finance. Every industry where your compliance team has a meltdown the second someone mentions sending data to the cloud.

Gemma 4 runs entirely offline. Pull the ethernet cable. Still works. No cloud exposure, nothing leaves your machine, no nothing. For sensitive industries, this ain't just a nice-to-have anymore. It changes the entire AI adoption conversation inside your org.

The larger models support a 256K token context window, meaning you can feed in massive documents without chunking. Native image and video analysis. Code generation. Complex multi-step reasoning built on Gemini 3 research. Google also released Google AI Edge Gallery for iOS and Android, letting you download the E2B and E4B variants straight to your phone for a fully capable AI that works with no service at all.

"We can't use AI because of data policies" just got a lot harder to say with a straight face.

Try This

Name one workflow at your company where data privacy killed the AI conversation. Patient notes, contracts, internal financials, anything that couldn't go near ChatGPT.

Download Google AI Edge Gallery on your phone and install the E4B model. Takes a few minutes. Run real examples from that sensitive workflow and see what you get.

If the quality works for you, you've got your business case. No cloud, no compliance headaches, zero cost. Walk into that next IT meeting with actual receipts.

Reply

or to participate.