Small Language Models: The Future AI Tools for Schools

Summary

While Large Language Models (LLMs) like GPT-5 have captured the world's attention, a quiet but powerful revolution is underway in AI research. In a notable research report from June 2025, researchers from NVIDIA, among others, argue that the future of "agentic AI"—specialized AI assistants—belongs to Small Language Models (SLMs). These smaller models are not just scaled-down versions of their larger siblings; they represent a paradigm shift towards efficiency, specialization, and safety. For schools, such a development means an opportunity to access AI tools that are significantly cheaper, faster, and can be run locally, which is crucial for protecting student privacy.


What is a Small Language Model? A Technical Explanation

A Small Language Model is a type of neural network, just like an LLM, but constructed with significantly fewer parameters. Parameters can be likened to the connections in a network that collectively store the model's knowledge and ability to spot patterns. While an LLM might have hundreds of billions of parameters, an SLM typically ranges from hundreds of millions up to low tens of billions of parameters and can run on consumer/edge hardware with low latency.

The core arguments for SLMs are:

  • Powerful enough: Modern research shows well-constructed SLMs can perform at the same level as, or even outperform, significantly larger models on specific, bounded tasks. Microsoft's Phi-3 (3.8B) is designed to run directly on a phone and shows strong results.
  • Better suited for AI agents: Most tasks an AI assistant performs are repetitive and non-conversational. When an SLM is fine-tuned and evaluated against clear format requirements, it tends to be more faithful to the format and predictable than a general LLM, though hallucinations can still occur and must be managed.
  • Economically superior: An SLM is significantly cheaper (often by multiples) to run than 70B-class models. Furthermore, the cost to fine-tune and specialize an SLM is a fraction of the cost for an LLM.

The idea is not for SLMs to replace LLMs entirely. Future systems are expected to be heterogeneous, or "hybrid systems," where SLMs form the base and handle the majority of all tasks locally. When a truly complex task arises, the system can selectively call upon a large, cloud-based LLM.


Practical Applications in the Classroom

The technical potential of Small Language Models can be translated into concrete and safe tools for the classroom:

  • The Local Writing Assistant: A tool on every student's computer providing help with spelling and grammar. Running the model locally minimizes data sharing and can facilitate GDPR compliance (data minimization, purpose limitation), but still requires internal routines, DPIA, and contractual support.
  • Subject-Specific AI Tutor: A school can fine-tune an SLM on its own course material for, say, biology or history. Students can then get answers grounded in approved material, but it is important to remember that even a specialized SLM can hallucinate outside its training data and quality control is necessary.
  • Safe Planning Partner for Teachers: An SLM can be installed locally on the teacher's computer to help create lesson plans and assessment rubrics based on curriculum requirements.

Obstacles on the Road and Why Change Takes Time

If SLMs are so beneficial, why do LLMs still dominate the market? According to research, there are several hurdles:

  1. Industrial Inertia: Massive capital investments have been made in centralized infrastructure for LLMs.
  2. Misleading Performance Metrics: Many standardized tests are designed to measure broad general knowledge, favoring generalist models and not always capturing an SLM's high performance on a specific task.
  3. Low Awareness: Broad discussion on AI has so far focused on the largest models, which is a market observation rather than a research fact.

Next Steps for Teachers and Schools

To start leveraging the potential of Small Language Models, schools and teachers can take the following steps:

  • Start Small: Initiate a pilot project with an open-source SLM (e.g., a model from the Llama, Mistral, or Phi family) for a well-defined task.
  • Focus on Digital Sovereignty: Use the argument of increased data security to justify investments in local AI infrastructure to school leadership.
  • Set Requirements at Implementation: When evaluating new digital tools, ask providers if their AI features can run locally and if they are based on SLM architecture.

Conclusion: A Smarter, Not Just Bigger, Future

The shift from a one-sided focus on LLMs to a more nuanced ecosystem where Small Language Models play a central role is a step towards more sustainable and responsible AI. For schools, this approach offers a concrete path forward to implement AI in a way that is economically defensible, pedagogically relevant, and safer for students.


Read More