株式会社オブライト
AI2026-05-17

SLM (Small Language Model)

Also known as: Small Language Model / 小規模言語モデル

A language model in the hundreds-of-millions to low-billions parameter range. Key advantages are local execution, lower cost, and low latency — ideal for edge devices and domain-specific tasks.


Overview

SLM is the umbrella term for language models smaller than frontier LLMs — examples include Gemma, Qwen, Phi, and Mistral. They run on consumer GPUs or Apple Silicon Macs without cloud API fees. While less capable than frontier models in raw generality, fine-tuning on domain data often brings them to production quality for specific tasks.

SMB fit

On-premises or local execution eliminates data-leakage risk and keeps monthly costs near zero. Paired with Ollama or OpenClaw, an SMB can build an internal RAG system or customer-support bot without any cloud dependency.

Related Columns

AI
Small Language Models Are the Star of 2026: Why SMBs Should Adopt SLMs Now and How to Get Started
Gartner has named Domain-Specific Language Models a top strategic technology trend for 2026. Small Language Models (SLMs) are transforming AI adoption for SMBs with lower costs, higher accuracy for specific tasks, and zero data leakage risk. This guide covers benefits, leading models, practical use cases, and step-by-step adoption.
AI
Qwen3.5-9B Complete Guide: Run on Ollama with Just 5GB — Features, Benchmarks & Use Cases
Comprehensive guide to Qwen3.5-9B: Ollama setup instructions, hybrid Gated DeltaNet + Sparse MoE architecture, 262K context window, GPQA 81.7 and IFBench 76.5 (beating GPT-5.2's 75.4), comparison with GPT-4o-mini and Claude Haiku, and practical business use cases. Runs on just 5GB RAM.
AI
Gemma 4 Complete Guide — Features, System Requirements & Ollama Setup [2026]
Complete guide to Google Gemma 4 (released April 2, 2026): 4 model variants (E2B/E4B/26B MoE/31B Dense), Apache 2.0 license, system requirements, multimodal capabilities, AIME 89% benchmark, 140+ languages, and step-by-step Ollama installation and setup instructions.
AI
Zero-Cost Internal AI Chatbot with Ollama and OpenClaw
This article explains how to build an internal AI chatbot with zero API costs using Ollama and OpenClaw. We introduce implementation methods for cost reduction crucial to SMBs, integration with existing Slack and LINE, conversation memory, and FAQ automation. Centered in Shinagawa, Minato, Ota, and Meguro wards, we propose a zero-cost AI strategy that can start with existing Mac hardware.
AI
Hybrid AI Strategy Guide — Achieving 50% Cost Reduction with Cloud API + Local LLM [2026]
A practical guide to reducing AI operational costs by over 50% with a hybrid AI strategy combining cloud APIs and local LLMs. Learn optimal architecture design and implementation steps using local models like Qwen 3.5 and DeepSeek R1 with Claude, GPT, and Gemini.

Related Terms

Feel free to contact us

Contact Us