π Real-Time On-Device AI Agent with Polaris-4B β Run It Yourself, No Cloud, No Cost
We just deployed a real-time on-device AI agent using the Polaris-4B-Preview model β one of the top-performing <6B open LLMs on Hugging Face.
π± Whatβs remarkable? This model runs entirely on a mobile device, without cloud, and without any manual optimization. It was built using ZETIC.MLange, and the best part?
β‘οΈ Itβs totally automated, free to use, and anyone can do it. You donβt need to write deployment code, tweak backends, or touch device-specific SDKs. Just upload your model β and ZETIC.MLange handles the rest.
π§ About the Model - Model: Polaris-4B-Preview - Size: ~4B parameters - Ranking: Top 3 on Hugging Face LLM Leaderboard (<6B) - Tokenizer: Token-incremental inference supported - Modifications: None β stock weights, just optimized for mobile
βοΈ What ZETIC.MLange Does ZETIC.MLange is a fully automated deployment framework for On-Device AI, built for AI engineers who want to focus on models β not infrastructure.
Hereβs what it does in minutes: - π Analyzes model structure - βοΈ Converts to mobile-optimized format (e.g., GGUF, ONNX) - π¦ Generates a runnable runtime environment with pre/post-processing - π± Targets real mobile hardware (CPU, GPU, NPU β including Qualcomm, MediaTek, Apple) - π― Gives you a downloadable SDK or mobile app component β ready to run And yes β this is available now, for free, at https://mlange.zetic.ai
π§ͺ For AI Engineers Like You, If you want to: - Test LLMs directly on-device - Run models offline with no latency - Avoid cloud GPU costs - Deploy to mobile without writing app-side inference code
Then this is your moment. You can do exactly what we did, using your own models β all in a few clicks.
π Real-Time On-Device AI Agent with Polaris-4B β Run It Yourself, No Cloud, No Cost
We just deployed a real-time on-device AI agent using the Polaris-4B-Preview model β one of the top-performing <6B open LLMs on Hugging Face.
π± Whatβs remarkable? This model runs entirely on a mobile device, without cloud, and without any manual optimization. It was built using ZETIC.MLange, and the best part?
β‘οΈ Itβs totally automated, free to use, and anyone can do it. You donβt need to write deployment code, tweak backends, or touch device-specific SDKs. Just upload your model β and ZETIC.MLange handles the rest.
π§ About the Model - Model: Polaris-4B-Preview - Size: ~4B parameters - Ranking: Top 3 on Hugging Face LLM Leaderboard (<6B) - Tokenizer: Token-incremental inference supported - Modifications: None β stock weights, just optimized for mobile
βοΈ What ZETIC.MLange Does ZETIC.MLange is a fully automated deployment framework for On-Device AI, built for AI engineers who want to focus on models β not infrastructure.
Hereβs what it does in minutes: - π Analyzes model structure - βοΈ Converts to mobile-optimized format (e.g., GGUF, ONNX) - π¦ Generates a runnable runtime environment with pre/post-processing - π± Targets real mobile hardware (CPU, GPU, NPU β including Qualcomm, MediaTek, Apple) - π― Gives you a downloadable SDK or mobile app component β ready to run And yes β this is available now, for free, at https://mlange.zetic.ai
π§ͺ For AI Engineers Like You, If you want to: - Test LLMs directly on-device - Run models offline with no latency - Avoid cloud GPU costs - Deploy to mobile without writing app-side inference code
Then this is your moment. You can do exactly what we did, using your own models β all in a few clicks.
we've merged LightGlue keypoint matcher to Hugging Face transformers! it allows commercial use when paired with an open-source keypoint detector ππ»
π SmolAgents v1.19.0 is live! This release brings major improvements to agent flexibility, UI usability, streaming architecture, and developer experience: making it easier than ever to build smart, interactive AI agents. Here's what's new:
π§ Agent Upgrades - Support for managed agents in ToolCallingAgent - Context manager support for cleaner agent lifecycle handling - Output formatting now uses XML tags for consistency
π₯οΈ UI Enhancements - GradioUI now supports reset_agent_memory: perfect for fresh starts in dev & demos.
π Streaming Refactor - Streaming event aggregation moved off the Model class - β‘οΈ Better architecture & maintainability
π¦ Output Tracking - CodeAgent outputs are now stored in ActionStep - β More visibility and structure to agent decisions
π Bug Fixes - Smarter planning logic - Cleaner Docker logs - Better prompt formatting for additional_args - Safer internal functions and final answer matching
π Docs Improvements - Added quickstart examples with tool usage - One-click Colab launch buttons - Expanded reference docs (AgentMemory, GradioUI docstrings) - Fixed broken links and migrated to .md format
Hugging Face just wrapped 4 months of deep work with AMD to push kernel-level optimization on their MI300X GPUs. Now, it's time to share everything we learned.
Join us in Paris at STATION F for a hands-on weekend of workshops and a hackathon focused on making open-source LLMs faster and more efficient on AMD.
Prizes, amazing host speakers, ... if you want more details, navigate to https://lu.ma/fmvdjmur!
2 replies
Β·
reacted to their post with π€ππ§ 3 months ago
Iβve been running small language models (SLLMs) directly on smartphones β completely offline, with no cloud backend or server API calls.
I wanted to share: 1. β‘Β Tokens/sec performance across several SLLMs 2. π€Β Observations on hardware utilization (where the workload actually runs) 3. πΒ Trade-offs between model size, latency, and feasibility for mobile apps
There are reports for below models - QWEN3 0.6B - NVIDIA/Nemotron QWEN 1.5B - SimpleScaling S1 - TinyLlama - Unsloth tuned Llama 3.2 1B - Naver HyperClova 0.5B
πComparable Benchmark reports (no cloud, all on-device): Iβd really value your thoughts on: - Creative ideas to further optimize inference under these hardware constraints - Other compact LLMs worth testing on-device - Experiences youβve had trying to deploy LLMs at the edge
If thereβs interest, Iβm happy to share more details on the test setup, hardware specs, or the tooling we used for these comparisons.
Thanks for taking a look, and you can build your own through at "https://mlange.zetic.ai"!