I replaced ChatGPT with 5 'tiny' AI tools — they are faster, greener and most can run offline

A close up photo of someone's hands while typing on a laptop
(Image credit: Shutterstock)

It’s no secret that AI comes with an environmental cost. According to the International Energy Agency, data centers consumed around 415 terawatt-hours of electricity in 2025 — roughly 1.5% of global electricity demand — and that figure is projected to climb to about 945 TWh by 2030, largely driven by AI growth. In the U.S., the agency says data centers accounted for around half of all electricity demand growth in 2025. That doesn’t mean AI is inherently bad for the planet, but it does mean every prompt has a footprint — which is exactly why smaller, more efficient AI models deserve more attention this Earth Day...and every day.

When most people think about AI, they picture the biggest names like ChatGPT, Gemini or Claude. Unfortunately, those big names come with big electricity and water useage, too. While AI can help solve real-world problems, it’s fair to ask whether every task needs a system designed to solve quantum physics just to rewrite an email.

That's where small language models, often called SLMs can be useful. And yes, they are exactly what they sound like, AI models built to be smaller and more efficient than the most popular chatbots. And after spending time with them, I’m starting to think they’re a smarter fit for everyday life.

What is a small language model?

man texting

(Image credit: Future)

Think of today’s biggest AI systems as enormous SUVs: powerful, impressive and capable of hauling just about anything. Small language models, on the other hand, are more like hybrids or compact cars.

Article continues below

They use fewer parameters (the internal settings that help AI learn patterns), need less computing power and can often run faster and cheaper. Some are even small enough to run directly on phones or laptops instead of relying entirely on cloud servers.

And they are honest enough for tackling many everyday AI tasks like rewriting an email, summarizing notes, organizing a grocery or to-do list, translating a short message, brainstorming captions,answering simple questions and sorting data or forms.

Why this matters for our Earth

data center cooling

(Image credit: Shutterstock)

Large AI models can be resource-intensive. Training them takes enormous compute power, and serving millions of prompts every day requires energy-hungry infrastructure.

Smaller models aren’t impact-free, but they can reduce that load in several ways:

  • They use less compute. Less complexity often means less processing power for common tasks.
  • They can run locally. When AI runs on a device instead of a server, it can reduce constant back-and-forth cloud processing. In fact, your computer or phone might already have a dedicated "brain" called an NPU (Neural Processing Unit), built specifically for these models, making them run even more efficiently.
  • They’re faster for simple jobs. You don’t need a supercomputer to make a to-do list.
  • They may extend device usefulness. As on-device AI improves, people may rely more on current hardware instead of always chasing the next upgrade.

I still use large AI systems when I need deep reasoning, advanced writing help or complex research.

And, of course I test and review these models for my job, but when I'm off the clock and doing quick tasks, I’m becoming more intentional.

Five small language models worth knowing about

Google Gemma

(Image credit: Google)

Here are some small language models already shaping the AI landscape:

  • Microsoft Phi-3 Mini. I feel as though Microsoft’s Phi family is one of the clearest examples of compact AI done right. I like this model for strong reasoning in a lightweight package.
  • Google Gemma. You can't go wrong with small models when they are inspired by the research behind Gemini. Gemma is a very capable AI. I really like using this for small tasks.
  • Meta Llama 3.2 1B and 3B. I wasn't a huge user of this model at first, simply because I was mostly using Gemma, but the on-device focus of Meta's small model is great for speed without constant cloud dependence.
  • Alibaba Cloud Qwen 2.5 1.5B. I have been pleasantly surprised by Alibaba’s Qwen family of compact models. They pack a huge punch with strong efficiency, multilingual ability and affordability.
  • IBM Granite. I have tested this the least, but I'm a fan so far. This model focuses on practical enterprise tasks like summarization, automation and internal workflows

Bottom line

The AI race has focused on bigger, faster and more powerful. But Earth Day is a good reminder that smarter doesn’t always mean larger. Small language models won’t replace every chatbot but for many everyday tasks, they are worth checking out. Not to mention, a great tool for speed, privacy and potentially a lighter footprint.

Have you tried a small language model? Give it a try and let me know what you think in the comments.


Google News

Follow Tom's Guide on Google News and add us as a preferred source to get our up-to-date news, analysis, and reviews in your feeds.


More from Tom's Guide

Amanda Caswell
AI Editor

Amanda Caswell is one of today’s leading voices in AI and technology. A celebrated contributor to various news outlets, her sharp insights and relatable storytelling have earned her a loyal readership. Amanda’s work has been recognized with prestigious honors, including outstanding contribution to media.

Known for her ability to bring clarity to even the most complex topics, Amanda seamlessly blends innovation and creativity, inspiring readers to embrace the power of AI and emerging technologies. As a certified prompt engineer, she continues to push the boundaries of how humans and AI can work together.

Beyond her journalism career, Amanda is a long-distance runner and mom of three. She lives in New Jersey.

You must confirm your public display name before commenting

Please logout and then login again, you will then be prompted to enter your display name.