24GB VRAM Models for Ollama
High-value tier for large quantized models and long context tests.
Recommended models
| Model | Fit | Expected tok/s |
|---|---|---|
| Qwen3 32B Q4 | good | 10-16 |
| Llama 3.3 70B Q4 | edge | 6-9 |
High-value tier for large quantized models and long context tests.
| Model | Fit | Expected tok/s |
|---|---|---|
| Qwen3 32B Q4 | good | 10-16 |
| Llama 3.3 70B Q4 | edge | 6-9 |