News: 1755261130

  ARM Give a man a fire and he's warm for a day, but set fire to him and he's warm for the rest of his life (Terry Pratchett, Jingo)

Little LLM on the RAM: Google's Gemma 270M hits the scene

(2025/08/15)


Google has unveiled a pint-sized new addition to its "open" large language model lineup: Gemma 3 270M.

Weighing in at 270 million parameters and requiring around 550MB of memory, it's designed to make waves in on-device deployment and rapid model iteration — despite the usual caveats around hallucinations, shaky output, and probable copyright entanglements baked into its training data.

Google launched the original Gemma family in [1]February 2024 , and at the time offered two flavours: a two-billion-parameter version designed for on-CPU execution and a more capable seven-billion-parameter version targeting systems with GPU- or TPU-based accelerators.

[2]

While positioned as "open" models, in contrast to the company's proprietary Gemini family, they, like most competing "open" models, did not include the source nor training data - only pre-trained models and weights, something which remains true for the latest entry in the family (or, as Google would have it, the " [3]Gemmaverse ."

[4]

[5]

The new, smaller model — optimized for on-device use and capable of running in as little as 550MB of RAM - is ideal for "high-volume, well-defined" tasks, says Google, or when "you need to make every millisecond and micro-cent count."

It is pitched as being ideal for rapid development thanks to the speed with which it can be fine-tuned, and in turn says that can lead to the easy creation of "a fleet of specialized task models."

[6]

Based on unverified internal benchmarking, Google claims that Gemma 3 270M outperforms similarly-sized models including SmollLM2-360M-Instruct and Qwen 2.5 0.5B Instruct at the IFEval instruction-following benchmark, though naturally delivers much poorer performance than the four-times-the-size Gemma 3 1B at a score of 51.2 to 80.2.

The model isn't, Google is keen to point out, designed for raw performance. Instead, the company is making much of its energy efficiency: when quantized down to INT4 precision - with quantization-aware trained (QAT) checkpoints already provided, and the promise of a minimum of performance impact over INT8 precision - Google's again-unverified internal testing showed a battery drain of 0.75 percentage points for 25 conversations of unspecified length when running on a Pixel 9 Pro smartphone.

While the model itself is smaller than its siblings, its training dataset is not. It includes a similar spread of material as its larger siblings - including web documents, source code, mathematical text, and images - yet the 270M-parameter model was trained with a claimed six trillion tokens, three times as many as the 1B-parameter version and half again as many as the 4B-parameter model.

[7]Are you willing to pay $100k a year per developer on AI?

[8]Claude Code's copious coddling confounds cross customers

[9]OpenAI's GPT-5 looks less like AI evolution and more like cost cutting

[10]Infosec hounds spot prompt injection vuln in Google Gemini apps

Only the biggest 12- and 27-billion parameter models beat it, at 12 trillion tokens and 14 trillion tokens respectively. Like all the other Gemma 3 models, the dataset has a "knowledge cut-off date" of August 2024, meaning anything newer than that will have to be fed to the model during fine-tuning or as part of a prompt.

As with the earlier, larger Gemma models, the new compact model is made available for free - but with a set of usage restrictions, the breach of which gives Google "the right to restrict (remotely or otherwise) usage of any of the Gemma Services that Google reasonable believes are in violation."

[11]

These restrictions are outlined in the [12]prohibited use policy , and include a ban on generating content "that infringes, misappropriates, or otherwise violates any individual's or entity's rights," the performance of "dangerous, illegal, or malicious activities," unlicensed practise of medicine and accounting, the generation or distribution of spam, and, more controversially, "attempts to override or circumvent safety filters" and the generation of "sexually explicit content," with a carve-out in the latter clause for "content created for scientific, educational, documentary, or artistic purposes."

Those interested in getting hands-on with the latest model in the "Gemmaverse" can find it on [13]Hugging Face , [14]Ollama , [Kaggle](https://www.kaggle.com/models/google/gemma-3), [15]LM Studio , and [16]Docker .

Google has also released a guide to [17]fine-tuning the model .

Get our [18]Tech Resources



[1] https://www.theregister.com/2024/02/22/google_gemma_llms/

[2] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=2&c=2aJ9ZmkQhL9a1kkOpVVYZ6gAAAAc&t=ct%3Dns%26unitnum%3D2%26raptor%3Dcondor%26pos%3Dtop%26test%3D0

[3] https://deepmind.google/models/gemma/gemmaverse/

[4] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=4&c=44aJ9ZmkQhL9a1kkOpVVYZ6gAAAAc&t=ct%3Dns%26unitnum%3D4%26raptor%3Dfalcon%26pos%3Dmid%26test%3D0

[5] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=3&c=33aJ9ZmkQhL9a1kkOpVVYZ6gAAAAc&t=ct%3Dns%26unitnum%3D3%26raptor%3Deagle%26pos%3Dmid%26test%3D0

[6] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=4&c=44aJ9ZmkQhL9a1kkOpVVYZ6gAAAAc&t=ct%3Dns%26unitnum%3D4%26raptor%3Dfalcon%26pos%3Dmid%26test%3D0

[7] https://www.theregister.com/2025/08/15/are_you_willing_to_pay/

[8] https://www.theregister.com/2025/08/13/claude_codes_copious_coddling_confounds/

[9] https://www.theregister.com/2025/08/13/gpt_5_cost_cutting/

[10] https://www.theregister.com/2025/08/08/infosec_hounds_spot_prompt_injection/

[11] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=3&c=33aJ9ZmkQhL9a1kkOpVVYZ6gAAAAc&t=ct%3Dns%26unitnum%3D3%26raptor%3Deagle%26pos%3Dmid%26test%3D0

[12] https://ai.google.dev/gemma/prohibited_use_policy

[13] https://huggingface.co/collections/google/gemma-3-release-67c6c6f89c4f76621268bb6d

[14] https://ollama.com/library/gemma3

[15] https://lmstudio.ai/models/google/gemma-3-270m

[16] https://hub.docker.com/r/ai/gemma3

[17] https://ai.google.dev/gemma/docs/core/huggingface_text_full_finetune

[18] https://whitepapers.theregister.com/



Wait what

cyberdemon

They want to police/censor usage of a *local* LLM, by "disabling it remotely"?

So is that an open admission that anyone who uses their software is opening a backdoor to Google and sending them all their data anyway?

GIGO

Eye Know

Garbage in, garbage out.

Re: GIGO

Anonymous Coward

I was expecting at least one of those Gs to stand for 'Google'.

Re: GIGO

elsergiovolador

Google is a sound someone makes trying to say gargle with someone's balls inside the pie hole.

Re: GIGO

Snowy

Google in, Google observe. Happy now?

A MODERN FABLE

Aesop's fables and other traditional children's stories involve allegory
far too subtle for the youth of today. Children need an updated message
with contemporary circumstance and plot line, and short enough to suit
today's minute attention span.

The Troubled Aardvark

Once upon a time, there was an aardvark whose only pleasure in life was
driving from his suburban bungalow to his job at a large brokerage house
in his brand new 4x4. He hated his manipulative boss, his conniving and
unethical co-workers, his greedy wife, and his snivelling, spoiled
children. One day, the aardvark reflected on the meaning of his life and
his career and on the unchecked, catastrophic decline of his nation, its
pathetic excuse for leadership, and the complete ineffectiveness of any
personal effort he could make to change the status quo. Overcome by a
wave of utter depression and self-doubt, he decided to take the only
course of action that would bring him greater comfort and happiness: he
drove to the mall and bought imported consumer electronics goods.

MORAL OF THE STORY: Invest in foreign consumer electronics manufacturers.
-- Tom Annau