top of page

Small Language Models (SLMs)

Language Models

Classification

AI Model Design and Deployment

Overview

Small Language Models (SLMs) are neural network-based language models characterized by a significantly reduced parameter count compared to large language models (LLMs). Examples include Phi-2 and Mistral 7B, which are designed to offer efficient performance with lower computational and energy requirements. SLMs can be deployed on edge devices or in settings with limited resources, making them attractive for privacy-sensitive, cost-constrained, or latency-critical applications. However, their smaller size can limit the complexity of tasks they can perform and may reduce their ability to generalize across diverse topics. SLMs may also be more susceptible to bias and adversarial attacks due to reduced training data and model capacity. Despite these limitations, SLMs are increasingly used where transparency, speed, and data locality are prioritized, and ongoing research aims to close the performance gap with larger models.

Governance Context

Governance of SLMs involves ensuring responsible development and deployment, especially since their accessibility increases the risk of misuse. The EU AI Act, for example, imposes transparency obligations such as requiring clear documentation of intended use, data provenance, and risk assessment, regardless of model size. Organizations must implement at least two concrete controls: (1) maintain detailed records of training data sources and model limitations, and (2) conduct and document ongoing risk assessments throughout the model lifecycle. NIST's AI Risk Management Framework (AI RMF) also applies, mandating organizations to assess risks, implement monitoring, and maintain robust access controls. SLMs may be exempt from certain high-risk provisions but still require adherence to data protection (GDPR) and cybersecurity standards. Developers must also ensure compliance with sector-specific regulations (e.g., HIPAA for healthcare SLMs) and address model update practices to mitigate emergent vulnerabilities.

Ethical & Societal Implications

SLMs can enhance privacy and accessibility, but their limited capacity may lead to incomplete or biased outputs, especially in high-stakes domains. Their ease of deployment increases the risk of misuse, such as generating harmful content or facilitating social engineering attacks. SLMs may also perpetuate biases if trained on insufficient or unrepresentative datasets. Conversely, their transparency and resource efficiency can democratize AI access, but without robust governance, they risk undermining trust and safety in AI-driven systems. Additionally, the widespread use of SLMs could exacerbate digital divides if not accompanied by equitable access and oversight.

Key Takeaways

SLMs offer efficiency and privacy benefits but have reduced generalization capacity.; Governance frameworks like the EU AI Act and NIST AI RMF apply to SLMs.; SLMs are widely used in edge computing and regulated sectors with unique risks.; Transparency, documentation, and risk assessment are critical for SLM deployment.; Failure to update or validate SLMs can introduce security and ethical risks.; Concrete controls such as data provenance tracking and ongoing risk assessment are required.; SLMs' accessibility increases both opportunities and risks, demanding robust governance.

bottom of page