Small Language Models (SLMs)

Microsoft released in April 2024 Phi-3 mini, the first among three small language models (SLMs) the company plans to release.

SLMs are compact versions of LLMs. Phi-3 mini is an SLM that has 3.8 billion parameters, and is trained on a much smaller dataset, as compared to GPT-4.

It supports a context window of up to 1.28 lac tokens.

The company will add more models including Phi-3 small and Phi-3 medium to the Phi-3 family.

These models are cost-effective to operate and perform better on smaller devices such as laptops and smartphones.

SLMs are fine-tuned and are customized for specific tasks. They undergo targeted training (less computing power and energy consumption).

After receiving the prompt, the time taken by a model to make predictions is called inference latency. SLMs process quickly as they are smaller in size. They are more responsive and are suitable for real-time applications.

According to Microsoft, Phi-3 mini has outperformed models of the same size and next size across a variety of benchmarks (language, reasoning, coding, math).

It is ideal for analytical tasks as Phi-3 mini has strong reasoning and logic capabilities.

print

Leave a Reply

Your email address will not be published. Required fields are marked *