Google Gemma 3

Google’s Gemma 3 is a family of lightweight, open-source LLMs that perform remarkably well for their low resource requirements.

/images/agents/icons/gemma3.png
8.9
Shipyard Score
Excellent

What is it?

Gemma 3 is Google’s family of lightweight, open-source LLMs built alongside Gemini 2.0. Gemma 3 models come in four levels of parameters (1B, 4B, 12B, and 27B). They’re designed to be efficient and high-performing, and as non-reasoning models, they excel at pattern-based tasks. Gemma 3 models have multimodal capabilities and a 128K context window, and they can even be deployed locally or integrated into custom applications.

You can try Gemma 3 in your browser at Google AI Studio.

Google Gemma 3 model interface showing code generation capabilities Gemma 3 generating Python code based on a simple prompt for a scientific notation calculator

What can I use it for?

We found that Gemma 3 models were really strong when it came to implementing simple, common programming patterns. Gemma is also a good candidate for AI-based apps.

  • Code completion: Get suggestions based on the context of your program
  • Debugging: Recognize common errors + respond accordingly
  • Refactoring: Improve code structure and apply best practices while maintaining functionality
  • App dev: Use in your apps (esp. the smaller 1B and 4B models) for on-device AI
  • Agentic dev: Gemma 3 can be used with agents like Continue or Cursor for agentic dev tasks
  • Semantic search: Generate embeddings for document search and discovery
  • Multimodal processing: Gemma 3 can process both text and images using the built-in SigLIP image encoder
  • Self-hosting: Run it on your own infra for data privacy

What we like about it

Gemma 3 is an easy-to-onboard LLM for most general, simple tasks. The 27B model is especially impressive, and ranks 9th globally on Chatbot Arena on only a single GPU. We liked having an LLM we could easily run locally.

As a non-reasoning model, Gemma 3 was great for simple programming tasks; we used it as a secondary model for less complex code gen and ops.

We also appreciated how accessible Gemma 3 was when it came to fine-tuning for more advanced use cases. This is a huge plus both for using it for programming tasks, and customizing it as your LLM for AI-based apps.

Gemma 3 LLM output on Google AI Studio Gemma 3 explaining its code output

Try Shipyard today

Get isolated, full-stack ephemeral environments on every PR.

Shipyard's Agentic Corner

Get better agentic workflows by instantly deploying to ephemeral environments.

Shipyard's environments are agent-friendly, so you can have a more continuous and productive agentic dev loop.

Stay connected

Latest Agents

Shipyard Newsletter
Stay in the (inner) loop

Hear about the latest and greatest in cloud native, container orchestration, DevOps, and more when you sign up for our monthly newsletter.