Tagged "local-deployment"
- Qwen3.5-35B-A3B Emerges as Game-Changer for Agentic Coding Tasks
- PyTorch Foundation Announces New Members as Agentic AI Demand Grows
- Show HN: MCP-Enabled File Storage for AI Agents, Auth via Ethereum Wallet
- Show HN: A Human-Curated, CLI-Driven Context Layer for AI Agents
- No, Local LLMs Can't Replace ChatGPT or Gemini — I Tried
- Elastic Introduces Best-in-Class Embedding Models for High Performance Semantic Search
- Making Wolfram Technology Available as Foundation Tool for LLM Systems
- Wave Field LLM Achieves O(n log n) Scaling: 825M Model Trained to 1B Parameters in 13 Hours
- Custom Portable Workstation Optimized for Local AI Inference Builds
- nanollama: Open-Source Framework for Training Llama 3 from Scratch with One-Command GGUF Export
- A Tool to Tell You What LLMs Can Run on Your Machine
- GPT-OSS 20B Demonstrates Practical Agentic Capabilities Running Fully Locally
- GLM-5 Becomes Top Open-Weights Model on Extended NYT Connections Benchmark
- Elastic Introduces Best-in-Class Embedding Models for High Performance Semantic Search
- Ouro 2.6B Thinking Model GGUFs Released with Q8_0 and Q4_K_M Quantization
- Ollama 0.17 Released With Improved OpenClaw Onboarding
- How Slow Local LLMs Are on My Framework 13 AMD Strix Point
- Claude Code Open – AI Coding Platform with Web IDE and Agents
- 24 Simultaneous Claude Code Agents on Local Hardware
- LayerScale Launches Inference Engine Faster Than vLLM, SGLang, and TRT-LLM
- Why My Country's AI Scene Is Built on Sand
- Sarvam AI Launches Edge Model to Challenge Major AI Players with Local-First Approach
- AMD Announces Day 0 Support for Qwen 3.5 LLM on Instinct GPUs
- Meet Sarvam Edge: India's AI Model That Runs on Phones and Laptops With No Internet
- Open-Source Models Now Comprise 4 of Top 5 Most-Used Endpoints on OpenRouter
- Alibaba Unveils Major AI Model Upgrade Ahead of DeepSeek Release
- WinClaw: Windows-Native AI Assistant with Office Automation
- Simile AI Raises $100M Series A for Local AI Infrastructure
- MiniMax M2.5: 230B Parameter MoE Model Coming to HuggingFace
- Ming-flash-omni-2.0: 100B MoE Omni-Modal Model Released
- Qwen Coder Next Shows Specialized Agent Performance
- GLM-5 Released: 744B Parameter MoE Model Targeting Complex Tasks
- Nanbeige4.1-3B: A Small General Model that Reasons, Aligns, and Acts
- Energy-Based Models Compared Against Frontier AI for Sudoku Solving
- DeepSeek Launches Model Update with 1M Context Window
- Anthropic Releases Claude Opus 4.6 Sabotage Risk Assessment