Skip to main content

33 models in this collection.

GLM 5

GLM · 744B

372 GB
chatcodingreasoningtool-use+2
VRAM427.8 GB
RAM558 GB
QuantQ4_K_M

DeepSeek V3.2

DeepSeek · 685B

342.5 GB
chatcodingreasoningtool-use+2
VRAM393.9 GB
RAM514 GB
QuantQ4_K_M

GLM 4.7

GLM · 355B

177.5 GB
chatcodingreasoningtool-use+2
VRAM204.1 GB
RAM267 GB
QuantQ4_K_M

MiMo V2 Flash

MiMo · 309B

154.5 GB
chatcodingreasoningtool-use+2
VRAM177.7 GB
RAM232 GB
QuantQ4_K_M

MiniMax M2.5

MiniMax · 230B

115 GB
chatcodingreasoningtool-use+2
VRAM132.3 GB
RAM173 GB
QuantQ4_K_M

Step 3.5 Flash

Step · 196B

98 GB
chatcodingreasoningtool-use+2
VRAM112.7 GB
RAM147 GB
QuantQ4_K_M

Devstral 2 123B

Mistral · 123B

61.5 GB
codingtool-usereasoningfrontier
VRAM70.7 GB
RAM93 GB
QuantQ4_K_M

GPT-OSS 120B

GPT-OSS · 120B

60 GB
chatcodingreasoningtool-use+2
VRAM69 GB
RAM90 GB
QuantQ4_K_M

Qwen3 Coder Next 80B A3B

Qwen · 80B

40 GB
codingtool-usereasoningmoe+2
VRAM46 GB
RAM60 GB
QuantQ4_K_M

Llama 3.1 70B

Llama · 70B

25 GB
chatgeneralcoding
VRAM27 GB
RAM32 GB
QuantQ2_K

Llama 3.3 70B

Llama · 70B

33 GB

Best for High-end workstations

chatgeneralcodinginstruct
VRAM35 GB
RAM40 GB
QuantQ3_K_M

Qwen3.6 35B A3B

Qwen · 35B

17.5 GB
chatcodingreasoningtool-use+2
VRAM20.1 GB
RAM27 GB
QuantQ4_K_M

CodeLlama 34B

CodeLlama · 34B

20 GB
codinginstructchat
VRAM22 GB
RAM24 GB
QuantQ4_K_M

Qwen 2.5 Coder 32B

Qwen · 32B

19 GB
codinginstructchat
VRAM21 GB
RAM24 GB
QuantQ4_K_M

Qwen3 Coder 30B A3B

Qwen · 30B

15 GB
codingtool-usereasoningmoe+1
VRAM17.3 GB
RAM23 GB
QuantQ4_K_M

Devstral Small 2 24B

Mistral · 24B

12 GB
codingtool-usereasoning
VRAM13.8 GB
RAM18 GB
QuantQ4_K_M

Codestral 22B

Mistral · 22B

13 GB

Best for Advanced coding assistants

codingfim
VRAM15 GB
RAM20 GB
QuantQ4_K_M

GPT-OSS 20B

GPT-OSS · 20B

10 GB
chatcodingreasoningtool-use+1
VRAM11.5 GB
RAM15 GB
QuantQ4_K_M

DeepSeek Coder V2 Lite 16B

DeepSeek · 16B

9.5 GB
codingmoechat
VRAM11 GB
RAM16 GB
QuantQ4_K_M

StarCoder2 15B

StarCoder · 15B

9 GB
codingfimchat
VRAM10.5 GB
RAM12 GB
QuantQ4_K_M

Phi-3 Medium 14B

Phi · 14B

8.2 GB
chatgeneralcoding
VRAM9.5 GB
RAM12 GB
QuantQ4_K_M

Phi-4 14B

Phi · 14B

8.2 GB

Best for Analysis-heavy assistants

chatgeneralreasoningcoding
VRAM9.5 GB
RAM12 GB
QuantQ4_K_M

Phi-4 Reasoning 14B

Phi · 14B

7 GB
chatreasoningcoding
VRAM8 GB
RAM11 GB
QuantQ4_K_M

Phi-4 Reasoning Plus 14B

Phi · 14B

7 GB
chatreasoningcodingfrontier
VRAM8 GB
RAM11 GB
QuantQ4_K_M

CodeLlama 13B

CodeLlama · 13B

7.9 GB
codinginstructchat
VRAM9 GB
RAM12 GB
QuantQ4_K_M

Llama 3.1 8B

Llama · 8B

3.9 GB

Best for General use

chatgeneralcoding
VRAM4.5 GB
RAM6 GB
QuantQ3_K_M

CodeLlama 7B

CodeLlama · 7B

4.2 GB
codinginstructchat
VRAM5 GB
RAM8 GB
QuantQ4_K_M

Qwen 2.5 Coder 7B

Qwen · 7B

4.7 GB

Best for Code generation

codinginstructchat
VRAM5.5 GB
RAM8 GB
QuantQ4_K_M

StarCoder2 7B

StarCoder · 7B

4.2 GB
codingfimchat
VRAM5 GB
RAM8 GB
QuantQ4_K_M

Phi-3 Mini 3.8B

Phi · 3.8B

2.3 GB
chatsmallcoding
VRAM3 GB
RAM4 GB
QuantQ4_K_M

Phi-4 Mini 3.8B

Phi · 3.8B

2.3 GB

Best for Low-VRAM devices

chatsmallreasoningcoding
VRAM3 GB
RAM4 GB
QuantQ4_K_M

Stable Code 3B

StableCode · 3B

1.8 GB
codingsmallchat
VRAM2.5 GB
RAM4 GB
QuantQ4_K_M

StarCoder2 3B

StarCoder · 3B

1.8 GB
codingsmallfimchat
VRAM2.5 GB
RAM4 GB
QuantQ4_K_M