Tagged "edge-device"
-
New Era of On-Device AI Driven by High-Speed UFS 5.0 Storage
-
PyTorch Foundation Announces New Members as Agentic AI Demand Grows
-
Mirai Announces $10M to Advance On-Device AI Performance for Consumer Devices
-
Show HN: 100% LLM Accuracy–No Fine-Tuning, JSON Only
-
How AI is Redefining Price and Performance in Modern Laptops
-
What Breaks When AI Agent Frameworks Are Forced Into <1MB RAM and Sub-ms Startup
-
Show HN: A Ground Up TLS 1.3 Client Written in C
-
Mirai Tech Raises $10 Million for On-Device AI Innovation
-
No, Local LLMs Can't Replace ChatGPT or Gemini — I Tried
-
Kioxia Sampling UFS 5.0 Embedded Flash Memory for Next-Generation Mobile Applications
-
Enhanced Interface Speed Enables High-Performance On-Device AI Features in Smartphones
-
Elastic Introduces Best-in-Class Embedding Models for High Performance Semantic Search
-
Enterprise Infrastructure Guide: Running Local LLMs for 70-150 Developers
-
Which Web Frameworks Are Most Token-Efficient for AI Agents?
-
Breaking the Speed Limit: Strategies for 17k Tokens/Sec Local Inference
-
South Korea to Launch $687 Million Project to Develop On-Device AI Semiconductors
-
Qwen3 Demonstrates Advanced Voice Cloning via Embeddings
-
Custom Portable Workstation Optimized for Local AI Inference Builds
-
Nvidia Could Launch Its First Laptops With Its Own Processors
-
Future of Mobile AI: What On-Device Intelligence Means for App Developers
-
Future of Mobile AI: What On-Device Intelligence Means for App Developers
-
The Complete Stack for Local Autonomous Agents: From GGML to Orchestration
-
Breaking the Speed Limit: Strategies for 17k Tokens/Sec Local Inference
-
Yet Another Fix Coming for Older AMD GPUs on Linux – Thanks to Valve Developer
-
AI-Powered Reverse-Engineering of Rosetta 2 for Linux
-
AI Is Stress Testing Processor Architectures and RISC-V Fits the Moment
-
How Slow Local LLMs Are on My Framework 13 AMD Strix Point
-
Google Open-Sources NPU IP, Synaptics Implements It for Hardware Acceleration
-
GGML Joins Hugging Face: What This Means for Local Model Optimization
-
DietPi Released a New Version v10.1
-
AI PCs Explained: 7 Critical Truths About NPUs and Privacy
-
Taalas Etches AI Models onto Transistors to Rocket Boost Inference
-
Strix Halo Performance Benchmarks: Minimax M2.5, Step 3.5 Flash, Qwen3 Coder
-
Qwen3 Coder Next Remains Effective at Aggressive Quantization Levels
-
[Release] Ouro-2.6B-Thinking: ByteDance's Recurrent Model Now Runnable Locally
-
GGML.AI Acquired by Hugging Face
-
Apple Researchers Develop On-Device AI Agent That Interacts With Apps for You
-
Sarvam Brings AI to Feature Phones, Cars, and Smart Glasses
-
Running Local LLMs and VLMs on Arduino UNO Q with yzma
-
Mihup and Qualcomm Collaborate to Advance Secure On-Device Voice AI for BFSI
-
Complete Offline AI System: Voice Control and Smart Home via Local LLM and Radio Without Internet
-
LayerScale Launches Inference Engine Faster Than vLLM, SGLang, and TRT-LLM
-
Kitten TTS V0.8 Released: State-of-the-Art Super-Tiny Text-to-Speech Model Under 25MB
-
Show HN: Shiro.computer Static Page, Unix/NPM Shimmed to Host Claude Code
-
Sarvam AI Launches Edge Model to Challenge Major AI Players with Local-First Approach
-
Qualcomm Ventures Positions India as Blueprint for Affordable On-Device AI Infrastructure
-
OpenClaw Refactored in Go, Runs on $10 Hardware
-
Same INT8 Model Shows 93% to 71% Accuracy Variance Across Snapdragon Chipsets
-
Cloudflare Releases Agents SDK v0.5.0 with Rust-Powered Infire Engine for Edge Inference
-
Can We Leverage AI/LLMs for Self-Learning?
-
Meet Sarvam Edge: India's AI Model That Runs on Phones and Laptops With No Internet
-
Cohere Releases Tiny Aya: Efficient 3.3B Multilingual Model for 70+ Languages
-
Chinese AI Chipmaker Axera Semiconductor Plans $379 Million Hong Kong IPO for Edge Inference Hardware
-
ASUS Zenbook 14 Launches in India with AI-Capable Hardware, Starting at Rs 1,15,990
-
Asus ExpertBook B3 G2 Laptop Features Ryzen AI 9 HX 470 CPU in 1.41kg Ultraportable Form Factor
-
GPU-Accelerated DataFrame Library for Local Inference Workloads
-
Alibaba Unveils Major AI Model Upgrade Ahead of DeepSeek Release
-
First Vibecoded AI Operating System for Local Deployment
-
Scaling llama.cpp On Neoverse N2: Solving Cross-NUMA Performance Issues
-
Scaling llama.cpp On Neoverse N2: Solving Cross-NUMA Performance Issues
-
Samsung's REAM: Alternative Model Compression Technique
-
Running Mistral-7B on Intel NPU Achieves 12.6 Tokens/Second
-
Memio Launches AI-Powered Knowledge Hub for Android with Local Processing
-
NAS System Achieves 18 tok/s with 80B LLM Using Only Integrated Graphics
-
Nanbeige4.1-3B: A Small General Model that Reasons, Aligns, and Acts
-
Energy-Based Models Compared Against Frontier AI for Sudoku Solving
-
Arm SME2 Technology Expands CPU Capabilities for On-Device AI