Close Menu

    Subscribe to Updates

    Get the latest news from tastytech.

    What's Hot

    New Christmas Movies to Watch in 2025

    December 13, 2025

    Demand for EVs to drop as buyers return to combustion cars – report

    December 13, 2025

    Hailuo AI Video Generator App Review: Key Features, Pricing

    December 13, 2025
    Facebook X (Twitter) Instagram
    Facebook X (Twitter) Instagram
    tastytech.intastytech.in
    Subscribe
    • AI News & Trends
    • Tech News
    • AI Tools
    • Business & Startups
    • Guides & Tutorials
    • Tech Reviews
    • Automobiles
    • Gaming
    • movies
    tastytech.intastytech.in
    Home»Business & Startups»A Guide to Agentic Coding
    A Guide to Agentic Coding
    Business & Startups

    A Guide to Agentic Coding

    gvfx00@gmail.comBy gvfx00@gmail.comDecember 12, 2025No Comments10 Mins Read
    Share
    Facebook Twitter LinkedIn Pinterest Email


    Mistral AI’s latest announcement introduces DevStral 2 (123B parameters), DevStral Small 2 (24B), and the Mistral Vibe CLI, a terminal-native coding assistant built for agentic coding tasks. Both models are fully open source and tuned for production workflows, while the new Vibe CLI brings project-aware editing, code search, version control, and execution directly into the terminal.

    Together, these updates aim to speed up developer workflows by making large-scale code refactoring, bug fixes, and feature development more automated, and in this guide we’ll outline the technical capabilities of each tool and provide hands-on examples to get started.

    Table of Contents

    Toggle
    • What is DevStral 2?
      • Key Features of DevStral 2
    • Setup for DevStral 2
    • Hands-On: DevStral 2
      • Task 1: Calling DevStral 2 via the Mistral API (Python SDK) 
      • Task 2: Hugging Face Transformers with DevStral 2
    • What is DevStral Small 2?
      • Key Features of DevStral Small 2
    • Hands-On: DevStral Small 2
      • Task: Calling DevStral Small 2 via the Mistral API
    • What is Mistral Vibe CLI?
      • Key Features of Mistral Vibe CLI
    • Setup for Mistral Vibe CLI
    • Hands-On: Mistral Vibe CLI
      • Task: Run Vibe in Script and Programmatic Mode
    • Conclusion
    • Frequently Asked Questions
        • Login to continue reading and enjoy expert-curated content.
      • Related posts:
    • Tracing & Debugging LLM Apps
    • Top 5 Agentic Coding CLI Tools
    • Statistics at the Command Line for Beginner Data Scientists

    What is DevStral 2?

    DevStral 2 is a 123-billion-parameter dense transformer designed specifically for software engineering agents. It features a 256K-token context window, enabling it to analyze entire code repositories at once. Despite its size, it is much smaller than competitor models: for example, DevStral 2 is 5x smaller than DeepSeek v3.2 and 8x smaller than Kimi K2 yet matches or exceeds their performance. This compactness makes DevStral 2 practical for enterprise deployment.

    Key Features of DevStral 2

    The Key technical highlights of DevStral 2 include: 

    • SOTA coding performance: 72.2% on the SWE-bench Verified test, making it one of the strongest open-weight models for coding. 
    • Large context handling: With 256K tokens, it can track architecture-level context across many files. 
    • Agentic workflows: Built to “explore codebases and orchestrate changes across multiple files”, DevStral 2 can detect failures, retry with corrections, and handle tasks like multi-file refactoring, bug fixing, and modernizing legacy code. 

    These capabilities mean DevStral 2 is not just a powerful code completion model, but a true coding assistant that maintains state across an entire project. For developers, this translates to faster, more reliable automated changes: for example, DevStral 2 can understand a project’s file structure and dependencies, propose code modifications across many modules, and even apply fixes iteratively if tests fail.

    You can learn more about the pricing of DevStral 2 from their official page.

    Setup for DevStral 2

    1. Sign up or Login to the mistral platform via https://v2.auth.mistral.ai/login. 
    2. Create your organization by giving an appropriate name. 
    3. Go to API Keys section in the sidebar and choose an appropriate plan.
    Mistral AI Studio
    1. Once the plan is activated, generate an API Key. 
    Creating new API Key

    Hands-On: DevStral 2

    Task 1: Calling DevStral 2 via the Mistral API (Python SDK) 

    Utilize Mistral’s official SDK to submit coding requests. For example, if you want DevStral 2 to redo a Python function for better speed, you can type: 

    !pip install mistralai 
    
    from mistralai import Mistral 
    import os
    from getpass import getpass 
    
    api_key = getpass("Enter your Mistral API Key: ") 
    
    client = Mistral(api_key=api_key) 
    response = client.chat.complete( 
    
    model="devstral-2512", # correct model name 
    messages=[ 
    
    {"role": "system", "content": "You are a Python code assistant."}, 
    {"role": "user", "content": ( 
    "Refactor the following function to improve performance:\n" 
    "```python\ndef compute_numbers(n):\n" 
    " result = []\n" 
    " for i in range(n):\n" 
    " if i % 100 == 0:\n" 
    " result.append(i**2)\n" 
    " return result\n```" 
    )} 
    ] 
    ) 
    
    print(response.choices[0].message.content)

    The request is made to DevStral 2 to make a loop function faster. The AI will examine the function and give a reformed version (for instance, recommending using list comprehensions or vectorized libraries). Although the Python SDK makes it easier to interact with the model, you may also opt to make HTTP requests for direct API access if that is your choice. 

    Refactored function

    Task 2: Hugging Face Transformers with DevStral 2

    Hugging Face has DevStral 2 weights available meaning that it is possible to run the model locally (if your hardware is good enough) using the Transformers library. Just to give an example: 

    !pip install transformers # make sure you have transformers installed 
    
    # optionally: pip install git+https://github.com/huggingface/transformers if using bleeding-edge 
    from transformers import MistralForCausalLM, MistralCommonBackend 
    import torch 
    
    model_id = "mistralai/Devstral-2-123B-Instruct-2512" 
    
    # Load tokenizer and model 
    tokenizer = MistralCommonBackend.from_pretrained(model_id, trust_remote_code=True) 
    model = MistralForCausalLM.from_pretrained(model_id, device_map="auto", trust_remote_code=True) 
    
    # Optionally, set dtype for better memory usage (e.g. bfloat16 or float16) if you have GPU 
    model = model.to(torch.bfloat16) 
    
    prompt = ( 
    "Write a function to merge two sorted lists of integers into one sorted list:\n" 
    "```python\n" 
    "# Input: list1 and list2, both sorted\n"
    "```" 
    ) 
    
    inputs = tokenizer(prompt, return_tensors="pt").to(model.device) 
    outputs = model.generate(**inputs, max_new_tokens=100) 
    
    print(tokenizer.decode(outputs[0], skip_special_tokens=True))

    The displayed code snippet utilizes the “DevStral 2 Instruct” model to produce a complete Python function similar to the previous code.

    What is DevStral Small 2?

    DevStral Small 2 brings the same design principles to a much smaller model. It has 24 billion parameters and the same 256K context window but is sized to run on a single GPU or even a high-end consumer CPU.

    Key Features of DevStral Small 2

    The Key attributes of DevStral Small 2 include: 

    • Lightweight & local: At 24B parameters, DevStral Small 2 is optimized for on-premises use. Mistral notes it can run on one RTX 4090 GPU or a Mac with 32GB RAM. This means developers can iterate locally without requiring a data-center cluster. 
    • High performance: It scores 68.0% on SWE-bench Verified, placing it on par with models up to 5x its size. In practice this means Small 2 can handle complex code tasks almost as well as larger models for many use cases.  
    • Multimodal support: DevStral Small 2 adds vision capabilities, so it can analyze images or screenshots in prompts. For example, you could feed it a diagram or UI mockup and ask it to generate corresponding code. This makes it possible to build multimodal coding agents that reason about both code and visual artifacts. 
    • Apache 2.0 open license: Released under Apache 2.0, DevStral Small 2 is free for commercial and non-commercial use.

    From a developer’s perspective, DevStral Small 2 enables fast prototyping and on-device privacy. Because inference is quick (even running on CPU), you get tight feedback loops when testing changes. And since the runtime is local, sensitive code never has to leave your infrastructure. 

    Hands-On: DevStral Small 2

    Task: Calling DevStral Small 2 via the Mistral API

    Just like DevStral 2, the Small model is available via the Mistral API. In the Python SDK, you could do: 

    !pip install mistralai 
    
    from mistralai import Mistral 
    import os 
    from getpass import getpass 
    
    api_key = getpass("Enter your Mistral API Key: ") 
    
    client = Mistral(api_key=api_key) 
    response = client.chat.complete( 
    model="devstral-small-2507", # updated valid model name 
    messages=[ 
    {"role": "system", "content": "You are a Python code assistant."}, 
    {"role": "user", "content": ( 
    "Write a clean and efficient Python function to find the first " 
    "non-repeating character in a string. Return None if no such " 
    "character exists." 
    )} 
    ] 
    ) 
    
    print(response.choices[0].message.content)

    Output: 

    Method explanation and example usage

    What is Mistral Vibe CLI?

    Mistral Vibe CLI is an open-source, Python-based command-line interface that turns DevStral into an agent running in your terminal. It provides a conversational chat interface that understands your entire project. Vibe automatically scans your project’s directory and Git status to build context.

    You can reference files with @autocompletion, execute shell commands with exclamation(!) , and use slash commands ( /config, /theme, etc.) to adjust settings. Because Vibe can “understand your entire codebase and not just the file you’re editing”, it enables architecture-level reasoning (for example, suggesting consistent changes across modules).

    Key Features of Mistral Vibe CLI

    The main characteristics of Vibe CLI are the following: 

    • Interactive chat with the tools: Vibe allows you to give it a chat-like prompt where the natural language requests are issued. However, it has an assortment of tools for reading and writing files, code search (grep), version control, and running shell commands. For instance, it can read a file with the read_file command, apply a patch by writing it to the file with the write_file command, search for the repo using grep, etc. 
    • Project-aware context: Vibe, by default, keeps the repo indexed to ensure any query is rendered by the complete project structure and Git history. You need not instruct it to the files manually rather just say “Update the authentication code” and it will investigate the relevant modules. 
    • Smart references: Referring to specific files (with autocompletion) is possible by using @path/to/file in prompts, and commands can be executed directly using !ls or other shell prefixes. Furthermore, builtin commands (e.g. /config) can be used through /slash. This results in a seamless CLI experience, complete with persistent history and even customization of the theme.  
    • Scripting and permissions: Vibe offers non-interactive mode (through --prompt or piping) to script batch tasks for scripting. You can create a config.toml file to set the default models (e.g. pointing to DevStral 2 via API), switch --auto-approve on or off for tool execution, and limit risky operations in sensitive repos. 

    Setup for Mistral Vibe CLI

    1. You can install Mistral Vibe CLI using one of the following commands: 
    uv tool install mistral-vibe

    OR 

    curl -LsSf https://mistral.ai/vibe/install.sh | sh 

    OR 

    pip install mistral-vibe 
    1. To launch the CLI, navigate to your project directory and then run the following command: 
    Vibe 
    Installing vibe and vibe-acp for vibe coding
    1. In case you are using Vibe for the very first time, it will do the following: 
    • Generate a pre-set configuration file named config.toml located at ~/.vibe/. 
    • Ask you to input your API key if it’s not set up yet, in that case, you could refer to these steps to register an account and obtain an API key. 
    • Store the API key at ~/.vibe/.env for the future. 

    Hands-On: Mistral Vibe CLI

    Task: Run Vibe in Script and Programmatic Mode

    Prompt: vibe "Write a Python function to reverse a linked list" 

    Running Mistral Vibe on Script mode

    Prompt for programmatic mode: 

    vibe -p "Generate a SQL schema for an employee database"
    Creating SQL Schema
    Response in programmatic mode

    The response was satisfactory.

    Conclusion

    DevStral 2, its smaller variant, and the Mistral Vibe CLI push hard toward autonomous coding agents, giving developers faster iteration, better code insight, and lower compute costs. DevStral 2 handles multi-file code work at scale, DevStral Small 2 brings similar behavior to local setups, and Vibe CLI makes both models usable directly from your terminal with smart, context-aware tools.

    To try them out, grab a Mistral API key, test the models through the API or Hugging Face, and follow the recommended settings in the docs. Whether you’re building codebots, tightening CI, or speeding up daily coding, these tools offer a practical entry into AI-driven development. Whereas DevStral 2 model series is competing in the LLM competition that’s out there, Mistral Vibe CLI is there to offer an alternative to the other CLI alternatives out there.

    Frequently Asked Questions

    Q1. How do DevStral 2, DevStral Small 2, and Vibe CLI help developers?

    A. They speed up coding by enabling autonomous code navigation, refactoring, debugging, and project-aware assistance directly in the terminal.

    Q2. What’s the difference between DevStral 2 and DevStral Small 2?

    A. DevStral 2 is a larger, more powerful model, while Small 2 offers similar agentic behavior but is light enough for local use.

    Q3. How do I start using these tools?

    A. Get a Mistral API key, explore the models through the API or Hugging Face, and follow the recommended settings in the official documentation.


    Riya Bansal

    Data Science Trainee at Analytics Vidhya
    I am currently working as a Data Science Trainee at Analytics Vidhya, where I focus on building data-driven solutions and applying AI/ML techniques to solve real-world business problems. My work allows me to explore advanced analytics, machine learning, and AI applications that empower organizations to make smarter, evidence-based decisions.
    With a strong foundation in computer science, software development, and data analytics, I am passionate about leveraging AI to create impactful, scalable solutions that bridge the gap between technology and business.
    📩 You can also reach out to me at [email protected]

    Login to continue reading and enjoy expert-curated content.

    Related posts:

    5 AI-Assisted Coding Techniques Guaranteed to Save You Time

    Top 15 Artificial Intelligence Movies And TV Series To Watch In 2025

    A Complete Guide to Seaborn

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleFiber vs. Cable Internet 101: The Best Tips
    Next Article How people really use AI: The surprising truth from analysing billions of interactions
    gvfx00@gmail.com
    • Website

    Related Posts

    Business & Startups

    The Smooth Alternative to ReLU

    December 13, 2025
    Business & Startups

    This Is Where JavaScript Beats Python For Machine Learning

    December 12, 2025
    Business & Startups

    10 GitHub Repositories to Master Machine Learning Deployment

    December 12, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    BMW Will Put eFuel In Cars Made In Germany From 2028

    October 14, 202511 Views

    Ubiquiti UNAS 2 Review: A Pure Network-Attached Storage Server

    September 30, 20255 Views

    What is Fine-Tuning? Your Ultimate Guide to Tailoring AI Models in 2025

    October 14, 20254 Views
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram

    Subscribe to Updates

    Get the latest tech news from tastytech.

    About Us
    About Us

    TastyTech.in brings you the latest AI, tech news, cybersecurity tips, and gadget insights all in one place. Stay informed, stay secure, and stay ahead with us!

    Most Popular

    BMW Will Put eFuel In Cars Made In Germany From 2028

    October 14, 202511 Views

    Ubiquiti UNAS 2 Review: A Pure Network-Attached Storage Server

    September 30, 20255 Views

    What is Fine-Tuning? Your Ultimate Guide to Tailoring AI Models in 2025

    October 14, 20254 Views

    Subscribe to Updates

    Get the latest news from tastytech.

    Facebook X (Twitter) Instagram Pinterest
    • Homepage
    • About Us
    • Contact Us
    • Privacy Policy
    © 2025 TastyTech. Designed by TastyTech.

    Type above and press Enter to search. Press Esc to cancel.

    Ad Blocker Enabled!
    Ad Blocker Enabled!
    Our website is made possible by displaying online advertisements to our visitors. Please support us by disabling your Ad Blocker.