NEW EVENT! Cutting-edge Trends for Every Industry at PACK EXPO Southeast
Discover packaging & processing solutions for all industries at the all-new PACK EXPO Southeast in Atlanta, GA, March 10-12, 2025

Small Language Models Enable Generative AI in Packaging and Processing Equipment

OEMs can introduce AI into their equipment by embedding small language models. They can run directly on a standalone/embedded PC, don’t require an Internet connection, eliminate latency and costs associated with AI. (Updated December 2024.)

Working The Phones

Editor's note: Advances in small language models are proceeding rapidly. This article was updated in December 2024 to account for the newest developments. (It was originally published in August 2024.)

In my previous column, I wrote about how declines in workforce quantity and quality are taking their toll on CPG customers. I sketched out how OEMs might leverage generative AI to create a new generation of machines that talk to operators and technicians, in any language. Essentially, packaging and processing OEMs can build their own version of ChatGPT right into their equipment, allowing operators and technicians to carry on a dialog with the machine. In this column I’ll walk through actual small language models that your engineers can download and begin experimenting with today.

Wait a minute, you say...there’s no way that your CPG customers will stand for your machine maintaining an open connection to the cloud in order for generative AI to work. You’re right about that, which is why there’s so much buzz right now in the AI community about a new generation of self-contained large language models that are designed to run on a local machine. No Internet connection required. Yes, you read that right.

In fact, there are several language models (which some have taken to calling small language models or SLMs) that are available for download that are open source or freely available via a commercial-friendly license that costs you exactly nothing. Why free? There’s an arm’s race right now among the big tech companies with AI model development. Users of these models, like you, benefit from their innovation. For more context, see this recent post from Mark Zuckerberg on why he believes the future of AI is open source, and what’s in it for his company.

Small language models can run on any PC based control or PC-based HMI, or even a dedicated PC embedded in your machine for the purpose of running a generative AI interface. A recurring theme around small language models is that they are a fraction of the size of current large language models, require a fraction of the computing power, and can yield performance coming close to that of large language models, depending on the application. They also eliminate any latency associated with round-trip communications to the cloud. All of the above makes this game-changing technology.

The size of language models is measured by the number of parameters. (If you’re curious, see a detailed explanation of what is meant by parameters when it comes to LLMs.) For example, ChatGPT 3.5 uses 175 billion parameters.  

In this column I’m going to focus on several SLMs that your engineers can begin experimenting with that are all 8 billion parameters or smaller. In fact, the more interesting models are 2 billion parameters and smaller. Counterintuitively, when it comes to embedding AI into your equipment, smaller is better: The smaller number of parameters, the less computing power is needed. I’m guessing if you create your own SLM specific to your machine that hoovers up every single word on every page of every manual ever written on that equipment, including hours of interviews with your design engineers, 2-billion-parameter models will be plenty powerful enough.

One final technical note. Many of these models are designed to run faster on a PC with a GPU (a Graphics Processing Unit), typically from Nvidia. These are the chips every AI company is desperate to lay their hands on to power their data centers. But GPUs (typically from Nvidia) can also be found in everyday PC workstations. Even some HMIs may contain GPUs to aid in visualization and graphics. You’ll want to check the hardware that you are using in or on your equipment, as a GPU will make these models run faster. That said, speed differences may be immaterial depending on what you design and how much information you incorporate into your custom models.

For details on the following models, have your engineers download this spreadsheet that we compiled that contains details on all the language models including:

  • Name of the model
  • Company who provides the model
  • Whether it’s open source
  • Links to detailed information on the model
  • Links to AI chats with Perplexity for details on the model

All the models are free, and all run on a PC or PC-based HMI. The spreadsheet also covers the hardware requirements of each model, where I was able to discern it. Once your engineers download this spreadsheet, they can begin tinkering.

The six families

INTRODUCING! The Latest Trends for All Industries at PACK EXPO Southeast
The exciting new PACK EXPO Southeast 2025 unites all vertical markets in one dynamic hub, generating more innovative answers to your production challenges. Don’t miss this extraordinary opportunity for your business!
Read More
INTRODUCING! The Latest Trends for All Industries at PACK EXPO Southeast
New e-book on Flexible Packaging
In this e-book, you’ll learn key considerations for vertical and horizontal f/f/s and how to choose between premade bags and an f/f/s system. Plus, discover the pitfalls to avoid on bagging machinery projects.
download
New e-book on Flexible Packaging