Ollama Models Api
Read the docs
Toggle theme
llama
nsheth/llama-3-lumimaid-8b-v0.1-iq-imatrix
It uses this one Q4_K_M-imat (4.89 BPW) quant for up to 12288 context sizes. for less than 8gb vram
281 Pulls
2 months ago
1 Tags
Back to models