Close Menu

    Subscribe to Updates

    Get the latest news from tastytech.

    What's Hot

    Physical AI moves closer to factory floors as companies test humanoid robots

    May 14, 2026

    5 Small Language Models for Agentic Tool Calling

    May 14, 2026

    Windows Update Will Soon Revert Problematic Drivers Automatically

    May 14, 2026
    Facebook X (Twitter) Instagram
    Facebook X (Twitter) Instagram
    tastytech.intastytech.in
    Subscribe
    • AI News & Trends
    • Tech News
    • AI Tools
    • Business & Startups
    • Guides & Tutorials
    • Tech Reviews
    • Automobiles
    • Gaming
    • movies
    tastytech.intastytech.in
    Home»Business & Startups»5 Small Language Models for Agentic Tool Calling
    5 Small Language Models for Agentic Tool Calling
    Business & Startups

    5 Small Language Models for Agentic Tool Calling

    gvfx00@gmail.comBy gvfx00@gmail.comMay 14, 2026No Comments7 Mins Read
    Share
    Facebook Twitter LinkedIn Pinterest Email



     

    Table of Contents

    Toggle
    • # Introduction
    • # 1. SmolLM3-3B
    • # 2. Qwen3-4B-Instruct-2507
    • # 3. Phi-3-mini-4k-instruct
    • # 4. Gemma-4-E2B-it
    • # 5. Mistral-7B-Instruct-v0.3
    • # Wrapping Up
      • Related posts:
    • Qwen3.5-Omni Can Clone Your Voice, Whisper, Shout
    • The One Model That Codes, Reasons, and Chats
    • Staying Ahead of AI in Your Career

    # Introduction

     
    Agentic AI systems depend on a model’s ability to reliably call tools, selecting the right function, formatting arguments correctly, and integrating results into multi-step workflows. Large frontier models such as ChatGPT, Claude, and Gemini handle this well, but they come with tradeoffs in cost, latency, and hardware requirements that make them impractical for many real-world deployments. Small language models have done well to close that gap, and several compact, open-weight options now offer first-class tool-calling support without the need for a data center to run them.

    And now, in no particular order, here are 5 small language models for agentic tool calling. Note that, for convenience and consistency, all model links point to Hugging Face-hosted models.

     

    # 1. SmolLM3-3B

     

     

    Technical Aspect Details
    Parameters 3B
    Architecture Decoder-only transformer (GQA + NoPE, 3:1 ratio)
    Context Length 64K native; up to 128K with YaRN extrapolation
    Training Tokens 11.2T
    Multilingual Support 6 languages (EN, FR, ES, DE, IT, PT)
    Reasoning Mode Dual-mode (thinking / no-think toggle)
    Tool Calling Yes: JSON/XML (xml_tools) and Python (python_tools)
    License Apache 2.0

     

    SmolLM3 is a 3B parameter language model designed to push the boundaries of small models, supporting dual-mode reasoning, 6 languages, and long context. It is a decoder-only transformer using Grouped Query Attention (GQA) and No Positional Embeddings (NoPE) (with a 3:1 ratio), pretrained on 11.2T tokens with a staged curriculum of web, code, math, and reasoning data. Post-training included a mid-training phase on 140 billion reasoning tokens, followed by supervised fine-tuning and alignment via Anchored Preference Optimization (APO), HuggingFace’s off-policy approach to preference alignment. The model supports two distinct tool-calling interfaces, JSON/XML blobs via xml_tools and Python-style function calls via python_tools, making it highly flexible for agentic pipelines and RAG systems. As a fully open release, including weights, datasets, and training code, SmolLM3 is ideal for chatbots, RAG systems, and code assistants on constrained hardware such as edge devices or low-VRAM machines.

     

    # 2. Qwen3-4B-Instruct-2507

     

     

    Technical Aspect Details
    Parameters 4.0B (3.6B non-embedding)
    Architecture Causal LM, 36 layers, GQA (32 Q heads / 8 KV heads)
    Context Length 262,144 tokens (native)
    Reasoning Mode Non-thinking only (no blocks)
    Multilingual 100+ languages
    Tool Calling Yes: native, via Qwen-Agent / MCP
    License Apache 2.0

     

    Qwen3-4B-Instruct-2507 is an updated version of the Qwen3-4B non-thinking mode, featuring significant improvements in general capabilities including: instruction following, logical reasoning, text comprehension, mathematics, science, coding, and tool usage. It also possesses substantial gains in long-tail knowledge coverage across multiple languages. Both the Instruct and Thinking variants share 4 billion total parameters (3.6B excluding embeddings) built across 36 transformer layers, using GQA with 32 query heads and 8 key/value heads, enabling efficient memory management for very long contexts. This specific non-thinking variant is optimized for direct, fast-response use cases, such as delivering concise answers without explicit chain-of-thought traces, making it well-suited for chatbots, customer support, and tool-calling agents where low latency matters. Qwen3 excels in tool-calling capabilities, and Alibaba recommends using the Qwen-Agent framework, which encapsulates tool-calling templates and parsers internally, reducing coding complexity, with support for MCP server configuration files.

     

    # 3. Phi-3-mini-4k-instruct

     

     

    Technical Aspect Details
    Parameters 3.8B
    Architecture Decoder-only transformer
    Context Length 4K tokens
    Vocabulary Size 32,064 tokens
    Training Data Synthetic + filtered public web data
    Post-training SFT + DPO
    Tool Calling Yes: via chat template (requiring HF’s transformers ≥ 4.41.2)
    License MIT

     

    Phi-3-Mini-4K-Instruct is a 3.8B parameter, lightweight, state-of-the-art open model trained with the Phi-3 datasets that include both synthetic data and filtered publicly available web data, with a focus on high-quality and reasoning-dense properties. The model underwent a post-training process incorporating both Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO) for instruction following and safety. Microsoft’s flagship “small but smart” model, Phi-3-mini was notable at launch for its ability to run on-device, including smartphones, while rivaling GPT-3.5 in capability benchmarks. The model is primarily intended for memory- and compute-constrained environments, latency-bound scenarios, and tasks requiring strong reasoning, especially math and logic. While older than the other models in this list and limited to a 4K context window, the MIT license makes it one of the most permissively licensed options available, and its strong general reasoning has made it a popular base for fine-tuning in commercial applications.

     

    # 4. Gemma-4-E2B-it

     

     

    Technical Aspect Details
    Effective Parameters 2.3B (5.1B total with embeddings)
    Architecture Dense, hybrid attention (sliding window + global) + PLE
    Layers 35
    Sliding Window 512 tokens
    Context Length 128K tokens
    Vocabulary Size 262K
    Modalities Text, Image, Audio (≤30 sec), Video (as frames)
    Multilingual 35+ native, trained on 140+ languages
    Tool Calling Yes: native function calling
    License Apache 2.0

     

    Gemma-4-E2B is part of Google DeepMind’s Gemma 4 family, which features a hybrid attention mechanism, local sliding window attention with full global attention. This design delivers the processing speed and low memory footprint of a lightweight model without sacrificing the deep awareness required for complex, long-context tasks. The “E” in E2B stands for “effective” parameters, enabled by a key architectural innovation called Per-Layer Embeddings (PLE), which adds a dedicated conditioning vector at every decoder layer. This is the mechanism which allows the E2B to run in under 1.5 GB of memory with quantization and still produce valuable outputs. The model supports native function calling, enabling agentic workflows, and is optimized for on-device deployment on mobile and IoT devices, capable of handling text, image, audio, and video inputs. Released under Apache 2.0 (a change from earlier Gemma generations’ more restrictive custom license), Gemma 4 E2B is an attractive option for developers building multimodal agentic applications running entirely at the edge.

     

    # 5. Mistral-7B-Instruct-v0.3

     

     

    Technical Aspect Details
    Parameters 7.25B
    Architecture Transformer, GQA + SWA
    Context Length 32,768 tokens
    Vocabulary Size 32,768 tokens (extended from v0.2)
    Tokenizer v3 Mistral tokenizer
    Function Calling Yes: via TOOL_CALLS / AVAILABLE_TOOLS / TOOL_RESULTS tokens (see here)
    License Apache 2.0

     

    Mistral-7B-Instruct-v0.3 is an instruct fine-tuned version of Mistral-7B-v0.3, which introduced three key changes over v0.2: an extended vocabulary to 32,768 tokens, support for the v3 tokenizer, and support for function calling. The model employs grouped-query attention for faster inference and Sliding Window Attention (SWA) to handle long sequences efficiently, and function calling support is made possible through the extended vocabulary including dedicated tokens for TOOL_CALLS, AVAILABLE_TOOLS, and TOOL_RESULTS. As the largest model in this roundup at 7B parameters, Mistral-7B-Instruct-v0.3 offers the best general instruction-following performance of the group and has become an industry-standard workhorse, widely available through Ollama, vLLM, and most inference platforms.

     

    # Wrapping Up

     
    The five models covered here — SmolLM3-3B, Qwen3-4B-Instruct-2507, Phi-3-mini-4k-instruct, Gemma-4-E2B-it, and Mistral-7B-Instruct-v0.3 — span a range of architectures, parameter counts, context windows, and release dates, but share one important trait: they all support structured tool calling in a compact, open-weight package.

    From Hugging Face’s fully transparent SmolLM3 to Google DeepMind’s multimodal edge-optimized Gemma 4 E2B, the selection demonstrates that capable agentic models no longer require massive infrastructure and frontier models to deploy. Whether your priority is on-device inference, long-context handling, multilingual coverage, or the most permissive license possible, there is a model in this list worth exploring.

    Keep in mind that these aren’t the only small language models with tool-calling capabilities. They do, however, do a good job representing those with which I have direct experience, and which I feel comfortable including based on my results.
     
     

    Matthew Mayo (@mattmayo13) holds a master’s degree in computer science and a graduate diploma in data mining. As managing editor of KDnuggets & Statology, and contributing editor at Machine Learning Mastery, Matthew aims to make complex data science concepts accessible. His professional interests include natural language processing, language models, machine learning algorithms, and exploring emerging AI. He is driven by a mission to democratize knowledge in the data science community. Matthew has been coding since he was 6 years old.



    Related posts:

    40 Questions to Go from Beginner to Advanced

    7 High Paying Side Hustles for Students

    The Conditional Memory Revolution for LLMs

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleWindows Update Will Soon Revert Problematic Drivers Automatically
    Next Article Physical AI moves closer to factory floors as companies test humanoid robots
    gvfx00@gmail.com
    • Website

    Related Posts

    Business & Startups

    OpenAI’s New API Voice Models Will Change the Way You Use AI

    May 14, 2026
    Business & Startups

    How AI Agents Will Transform Data Science Work in 2026

    May 13, 2026
    Business & Startups

    10 GitHub Repositories to Master Self-Hosting

    May 13, 2026
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Black Swans in Artificial Intelligence — Dan Rose AI

    October 2, 2025152 Views

    Every Clue That Tony Stark Was Always Doctor Doom

    October 20, 202587 Views

    We let ChatGPT judge impossible superhero debates — here’s how it ruled

    December 31, 202578 Views
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram

    Subscribe to Updates

    Get the latest tech news from tastytech.

    About Us
    About Us

    TastyTech.in brings you the latest AI, tech news, cybersecurity tips, and gadget insights all in one place. Stay informed, stay secure, and stay ahead with us!

    Most Popular

    Black Swans in Artificial Intelligence — Dan Rose AI

    October 2, 2025152 Views

    Every Clue That Tony Stark Was Always Doctor Doom

    October 20, 202587 Views

    We let ChatGPT judge impossible superhero debates — here’s how it ruled

    December 31, 202578 Views

    Subscribe to Updates

    Get the latest news from tastytech.

    Facebook X (Twitter) Instagram Pinterest
    • Homepage
    • About Us
    • Contact Us
    • Privacy Policy
    © 2026 TastyTech. Designed by TastyTech.

    Type above and press Enter to search. Press Esc to cancel.

    Ad Blocker Enabled!
    Ad Blocker Enabled!
    Our website is made possible by displaying online advertisements to our visitors. Please support us by disabling your Ad Blocker.