Coding
Best coding LLMs to run locally
Code-focused models for completion, refactors, and local agents.
33 models in this collection.
GLM 5
GLM · 744B
chatcodingreasoningtool-use+2
VRAM427.8 GB
RAM558 GB
QuantQ4_K_M
DeepSeek V3.2
DeepSeek · 685B
chatcodingreasoningtool-use+2
VRAM393.9 GB
RAM514 GB
QuantQ4_K_M
GLM 4.7
GLM · 355B
chatcodingreasoningtool-use+2
VRAM204.1 GB
RAM267 GB
QuantQ4_K_M
MiMo V2 Flash
MiMo · 309B
chatcodingreasoningtool-use+2
VRAM177.7 GB
RAM232 GB
QuantQ4_K_M
MiniMax M2.5
MiniMax · 230B
chatcodingreasoningtool-use+2
VRAM132.3 GB
RAM173 GB
QuantQ4_K_M
Step 3.5 Flash
Step · 196B
chatcodingreasoningtool-use+2
VRAM112.7 GB
RAM147 GB
QuantQ4_K_M
Devstral 2 123B
Mistral · 123B
codingtool-usereasoningfrontier
VRAM70.7 GB
RAM93 GB
QuantQ4_K_M
GPT-OSS 120B
GPT-OSS · 120B
chatcodingreasoningtool-use+2
VRAM69 GB
RAM90 GB
QuantQ4_K_M
Qwen3 Coder Next 80B A3B
Qwen · 80B
codingtool-usereasoningmoe+2
VRAM46 GB
RAM60 GB
QuantQ4_K_M
Llama 3.1 70B
Llama · 70B
chatgeneralcoding
VRAM27 GB
RAM32 GB
QuantQ2_K
Llama 3.3 70B
Llama · 70B
Best for High-end workstations
chatgeneralcodinginstruct
VRAM35 GB
RAM40 GB
QuantQ3_K_M
Qwen3.6 35B A3B
Qwen · 35B
chatcodingreasoningtool-use+2
VRAM20.1 GB
RAM27 GB
QuantQ4_K_M
CodeLlama 34B
CodeLlama · 34B
codinginstructchat
VRAM22 GB
RAM24 GB
QuantQ4_K_M
Qwen 2.5 Coder 32B
Qwen · 32B
codinginstructchat
VRAM21 GB
RAM24 GB
QuantQ4_K_M
Qwen3 Coder 30B A3B
Qwen · 30B
codingtool-usereasoningmoe+1
VRAM17.3 GB
RAM23 GB
QuantQ4_K_M
Devstral Small 2 24B
Mistral · 24B
codingtool-usereasoning
VRAM13.8 GB
RAM18 GB
QuantQ4_K_M
Codestral 22B
Mistral · 22B
Best for Advanced coding assistants
codingfim
VRAM15 GB
RAM20 GB
QuantQ4_K_M
GPT-OSS 20B
GPT-OSS · 20B
chatcodingreasoningtool-use+1
VRAM11.5 GB
RAM15 GB
QuantQ4_K_M
DeepSeek Coder V2 Lite 16B
DeepSeek · 16B
codingmoechat
VRAM11 GB
RAM16 GB
QuantQ4_K_M
StarCoder2 15B
StarCoder · 15B
codingfimchat
VRAM10.5 GB
RAM12 GB
QuantQ4_K_M
Phi-3 Medium 14B
Phi · 14B
chatgeneralcoding
VRAM9.5 GB
RAM12 GB
QuantQ4_K_M
Phi-4 14B
Phi · 14B
Best for Analysis-heavy assistants
chatgeneralreasoningcoding
VRAM9.5 GB
RAM12 GB
QuantQ4_K_M
Phi-4 Reasoning 14B
Phi · 14B
chatreasoningcoding
VRAM8 GB
RAM11 GB
QuantQ4_K_M
Phi-4 Reasoning Plus 14B
Phi · 14B
chatreasoningcodingfrontier
VRAM8 GB
RAM11 GB
QuantQ4_K_M
CodeLlama 13B
CodeLlama · 13B
codinginstructchat
VRAM9 GB
RAM12 GB
QuantQ4_K_M
Llama 3.1 8B
Llama · 8B
Best for General use
chatgeneralcoding
VRAM4.5 GB
RAM6 GB
QuantQ3_K_M
CodeLlama 7B
CodeLlama · 7B
codinginstructchat
VRAM5 GB
RAM8 GB
QuantQ4_K_M
Qwen 2.5 Coder 7B
Qwen · 7B
Best for Code generation
codinginstructchat
VRAM5.5 GB
RAM8 GB
QuantQ4_K_M
StarCoder2 7B
StarCoder · 7B
codingfimchat
VRAM5 GB
RAM8 GB
QuantQ4_K_M
Phi-3 Mini 3.8B
Phi · 3.8B
chatsmallcoding
VRAM3 GB
RAM4 GB
QuantQ4_K_M
Phi-4 Mini 3.8B
Phi · 3.8B
Best for Low-VRAM devices
chatsmallreasoningcoding
VRAM3 GB
RAM4 GB
QuantQ4_K_M
Stable Code 3B
StableCode · 3B
codingsmallchat
VRAM2.5 GB
RAM4 GB
QuantQ4_K_M
StarCoder2 3B
StarCoder · 3B
codingsmallfimchat
VRAM2.5 GB
RAM4 GB
QuantQ4_K_M