PXAI
Feed
Regions
DE
ES
FR
GR
IT
UK
US
View All
Viral
World
Politics
Technology
Daily Briefing
Sources
|
ToS
PXAI Audio Feed
+5
ΟΛΑ
11/04 01:26
dev.to
Claude Code install and config for Ollama, llama.cpp, pricing
Claude Code
agentic coding
project-level workflow
code completion vs. agentic
Ollama
llama.cpp
11/04 00:35
dev.to
Gemma4 Tool Calling Fixes in llama.cpp, RTX cuBLAS MatMul Bug, & Local Ollama + Whisper UI
Gemma4
llama.cpp
cuBLAS
RTX GPU
Whisper
Ollama
10/04 15:46
dev.to
~21 tok/s Gemma 4 on a Ryzen mini PC: llama.cpp, Vulkan, and the messy truth about local chat
Gemma 4
llama.cpp
Vulkan
local inference
Ubuntu 24.04
AMD iGPU
09/04 19:14
dev.to
I just open sourced Oryon, a local first desktop app for open source AI models
Oryon
open source AI
local-first
desktop app
Tauri
Rust
09/04 11:44
dev.to
Sick of API costs and rate limits? I turned my M1 Mac into a fully offline AI coding agent. No cloud. No API keys. Just raw local compute using Llama.cpp and a 26B model. Check out the architecture and build it yourself! 🚀👇
offline AI
M1 Mac
Llama.cpp
26B model
zero API cost
privacy
08/04 16:56
dev.to
Ollama, LM Studio, and GPT4All Are All Just llama.cpp — Here's Why Performance Still Differs
local LLM
llama.cpp
Ollama
LM Studio
inference speed
VRAM overhead
08/04 04:03
dev.to
LLMKube Now Deploys Any Inference Engine, Not Just llama.cpp
LLMKube
Kubernetes operator
inference engines
llama.cpp
vLLM
TGI
07/04 18:52
dev.to
Google Dropped TurboQuant Two Weeks Ago. The Community Already Made It Usable.
Google
TurboQuant
KV cache compression
llama.cpp
vLLM
memory reduction
07/04 04:34
dev.to
Running Gemma 2 27B Locally: MLX vs vLLM vs llama.cpp Performance Comparison
Gemma 2
MLX
vLLM
llama.cpp
inference harness
quantization
07/04 04:34
dev.to
Running Gemma 2 27B Locally: MLX vs vLLM vs llama.cpp Performance Comparison
Gemma 2
MLX
vLLM
llama.cpp
inference harness
quantization
Comments
Loading...
Send
Dev Changelog
v8.42
No logs found in database.
0
Display Settings
Size
Aa
Brightness
Theme
Dark
Comments