Close Menu
  • AI
  • Content Creation
  • Tech
  • Robotics
AI-trends.todayAI-trends.today
  • AI
  • Content Creation
  • Tech
  • Robotics
Trending
  • Anthropic releases Claude Opus 4.7, a major upgrade for agentic coding, high-resolution vision, and long-horizon autonomous tasks
  • The Coding Guide to Property Based Testing with Hypothesis and Stateful, Differential and Metamorphic Test Designs
  • Schematik Is ‘Cursor for Hardware.’ The Anthropics Want In
  • Hacking the EU’s new age-verification app takes only 2 minutes
  • Google AI Releases Google Auto-Diagnosis: A Large Language Model LLM Based System to Diagnose Integrity Test Failures At Scale
  • This is a complete guide to running OpenAI’s GPT-OSS open-weight models using advanced inference workflows.
  • The Huey Code Guide: Build a High-Performance Background Task Processor Using Scheduling with Retries and Pipelines.
  • Top 19 AI Red Teaming Tools (2026): Secure Your ML Models
AI-trends.todayAI-trends.today
Home»Tech»LocalCowork powered By LFM2-24BA2B Executes Privacy-First Agents Workflows Locally via Model Context Protocol.

LocalCowork powered By LFM2-24BA2B Executes Privacy-First Agents Workflows Locally via Model Context Protocol.

Tech By Gavin Wallace06/03/20264 Mins Read
Facebook Twitter LinkedIn Email
NVIDIA AI Releases Llama Nemotron Nano VL: A Compact Vision-Language
NVIDIA AI Releases Llama Nemotron Nano VL: A Compact Vision-Language
Share
Facebook Twitter LinkedIn Email

Liquid AI is now available. LFM2-24B-A2BThe model is designed for low latency local tool distribution. LocalCoworkOpen-source desktop agents are available on their Liquid4All GitHub Cookbook. It provides an architecture deployable for enterprise workflows that can be executed entirely on the device, removing API calls as well as data exit for sensitive environments.

Architectural and serving configuration

LFM2-24B A2B uses the Sparse MoE architecture to ensure low-latency performance on consumer hardware. Although the model contains 24 trillion parameters in total it activates only about 2 billion parameters for each token when inference is performed.

The model can maintain its broad knowledge base by using this structural design, while reducing computational overhead for every generation. Liquid AI stressed-tested the Liquid AI model with the following hardware/software stack:

  • Hardware: Apple M4 Max, 36 GB unified memory, 32 GPU cores.
  • Serving Engine llama-server Flash Attention is enabled.
  • Quantization: Q4_K_M format.
  • Memory Footprint The RAM is 14.5GB.
  • Hyperparameters: The temperature is set at 0.1. Top_p and max_tokens are both 0.1.

LocalCowork Tool Integration

LocalCowork, an offline AI desktop agent, uses the Model Context Protocol to run pre-built AI tools. It does not rely on cloud APIs and doesn’t compromise data privacy. Every action is recorded in a local audit log. It includes 75 tools on 14 MCP server capable of performing tasks like OCR and security scanning. The demo provided focuses instead on a subset of highly reliable and curated tools, spread across six servers. Each tool has been rigorously tested for over 80% accuracy in completing a single step, as well as verified participation in multi-step chains.

LocalCowork implements this model in practice. The system is available offline. The preconfigured tools include a full suite of business-grade applications:

  • File Operation: Browse, read, and search the filesystem of the host.
  • Security scanning: Local directories can be used to identify API keys, and personally identifiable information (PII).
  • Document processing: The OCR process, which includes parsing, diffing, and generating of PDFs.
  • The Audit Logging System To track compliance, it is important to record every call made by an instrument locally.

Performance Benchmarks

The Liquid AI Team evaluated the model using 100 one-step prompts for tool selection and 50 chains of multi-steps (requiring three to six discrete tool operations, including searching, OCRing, parsing, exporting, etc.).

Latency

This model is averaged ~385 ms per tool-selection response. The sub-second dispatch is ideal for applications that require immediate feedback.

Accuracy

  • Single-Step Executions: 80% accuracy.
  • Multi-Step Chains: Completion rate: 26%

What you need to know

  • Privacy First Local Execution LocalCowork is a mobile-only solution that does not rely on cloud APIs or data exfiltration. This makes it ideal for enterprise environments with strict privacy requirements.
  • Efficient MoE Architecture: LFM2-24B-A2B uses a Sparse Mixture-of-Experts design (MoE), activating 2 billion of the 24 billion parameters for each token. This allows it to comfortably fit within a 14.5GB RAM footprint. Q4_K_M quantization.
  • Hardware with a Sub-Second Lag: Benchmarked on the Apple M4 Max, this model has an average of 385 ms latency when executing tool selections, making it possible to have highly interactive real-time work flows.
  • Standardized MCP Tool Integration: The agent leverages the Model Context Protocol (MCP) to seamlessly connect with local tools—including filesystem operations, OCR, and security scanning—while automatically logging all actions to a local audit trail.
  • With Multi-Step limits, you can achieve high accuracy in single-step measurements. The model achieves 80% accuracy on single-step tool execution but drops to a 26% success rate on multi-step chains due to ‘sibling confusion’ (selecting a similar but incorrect tool), indicating it currently functions best in a guided, human-in-the-loop loop rather than as a fully autonomous agent.

Click here to find out more Repo The following are some examples of how to get started: Technical details. Also, feel free to follow us on Twitter Join our Facebook group! 120k+ ML SubReddit Subscribe now our Newsletter. Wait! Are you using Telegram? now you can join us on telegram as well.


privacy work x
Share. Facebook Twitter LinkedIn Email
Avatar
Gavin Wallace

Related Posts

Anthropic releases Claude Opus 4.7, a major upgrade for agentic coding, high-resolution vision, and long-horizon autonomous tasks

19/04/2026

The Coding Guide to Property Based Testing with Hypothesis and Stateful, Differential and Metamorphic Test Designs

19/04/2026

Google AI Releases Google Auto-Diagnosis: A Large Language Model LLM Based System to Diagnose Integrity Test Failures At Scale

18/04/2026

This is a complete guide to running OpenAI’s GPT-OSS open-weight models using advanced inference workflows.

18/04/2026
Top News

OpenAI’s Fidji Simo Plans to Make ChatGPT Means Extra Helpful—and Have You Pay For It

The Intelligence Age by Sam Altman • AI Blog

Meta receives a power and tax credit for Louisiana’s largest data centre

A startup is charging you for the chance to speak with AI-based versions of experts

Jack Dongarra: How supercomputing will evolve

Load More
AI-Trends.Today

Your daily source of AI news and trends. Stay up to date with everything AI and automation!

X (Twitter) Instagram
Top Insights

Baidu Releases ERNIE-4.5-21B-A3B-Thinking: A Compact MoE Model for Deep Reasoning

10/09/2025

Elon Musk Terafab partnership with Intel: 5 burning questions

08/04/2026
Latest News

Anthropic releases Claude Opus 4.7, a major upgrade for agentic coding, high-resolution vision, and long-horizon autonomous tasks

19/04/2026

The Coding Guide to Property Based Testing with Hypothesis and Stateful, Differential and Metamorphic Test Designs

19/04/2026
X (Twitter) Instagram
  • Privacy Policy
  • Contact Us
  • Terms and Conditions
© 2026 AI-Trends.Today

Type above and press Enter to search. Press Esc to cancel.