AI/Ollama: Difference between revisions
Jump to navigation
Jump to search
| Line 132: | Line 132: | ||
!scope='row' style='text-align:left' | <code>OLLAMA_NUM_GPU</code> | !scope='row' style='text-align:left' | <code>OLLAMA_NUM_GPU</code> | ||
| <code>999</code> || Forces Ollama to offload as many layers as possible to your Intel Arc iGPU instead of the slower CPU. | | <code>999</code> || Forces Ollama to offload as many layers as possible to your Intel Arc iGPU instead of the slower CPU. | ||
|- | |||
!scope='row' style='text-align:left' | <code>OLLAMA_NUM_GPU</code> | |||
| <code>0</code> || Forces Ollama CPU-only for all models | |||
|} | |} | ||
|} | |} | ||
Revision as of 10:10, 1 March 2026
curl -fsSL https://ollama.com/install.sh | sh
ollama pull gpt-oss:20b
ollama --version
ollama ls
curl -fsSL https://claude.ai/install.sh | bash
ollama launch claude --model gpt-oss:20b
|
export ANTHROPIC_BASE_URL=http://localhost:11434
export ANTHROPIC_AUTH_TOKEN=ollama
export ANTHROPIC_API_KEY=""
export OLLAMA_NUM_CTX=32768
export OLLAMA_KEEP_ALIVE=5m
claude --model gpt-oss:20b
| ||||
| |||||
Optimization
|
Optimization | ||||||||||||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
|
Installation
|
Installation | ||||
|---|---|---|---|---|
|
References
|
References | ||
|---|---|---|