| #2453 |
Add support for older AMD GPU gfx803, gfx802, gfx805 (e.g. Radeon RX 580, FirePro W7100)
feature requestamdbuild
|
2024-02-11 |
220 |
| #1016 |
Support AMD GPUs on Intel Macs
feature requestamdmacos
|
2023-11-06 |
169 |
| #5872 |
[Ascend ] add ascend npu support
|
2024-07-23 |
119 |
| #941 |
`digest mismatch` on download
bug
|
2023-10-28 |
155 |
| #1736 |
Download slows to a crawl at 99%
bugnetworkingregistry
|
2023-12-29 |
124 |
| #1730 |
MLX backend
feature request
|
2023-12-27 |
95 |
| #11691 |
Structured output with OpenAI SDK and gpt-oss:20b not working
buggpt-oss
|
2025-08-05 |
87 |
| #5245 |
Allow importing multi-file GGUF models
bug
|
2024-06-23 |
87 |
| #3368 |
Reranking models
|
2024-03-27 |
86 |
| #2006 |
Rate limit download speed on pulling new models
networking
|
2024-01-15 |
79 |
| #1590 |
Add support for Intel Arc GPUs
feature requestintel
|
2023-12-18 |
72 |
| #3504 |
I can't pull any models
bug
|
2024-04-05 |
61 |
| #162 |
Don't automatically start on startup / have an option to disable this
feature request
|
2023-07-21 |
61 |
| #10430 |
Adding support for amd new GPUS 9070 and 9070 XT
feature requestwindows
|
2025-04-27 |
59 |
| #7865 |
Model Context Protocol (MCP) support
feature request
|
2024-11-27 |
58 |
| #1102 |
Ollama on FreeBSD
feature request
|
2023-11-12 |
58 |
| #5186 |
AMD Ryzen NPU support
feature requestamd
|
2024-06-20 |
57 |
| #10844 |
discover/gpu.go: Add Support for Distributed Inferencing (continued)
|
2025-05-24 |
56 |
| #8618 |
Support Janus-Pro-7b for vision models
feature request
|
2025-01-27 |
56 |
| #5800 |
Enable speculative decoding
feature requestperformance
|
2024-07-19 |
56 |
| #9187 |
🪟 Windows 11 24H2 | Smart App Control Blocking Ollama | Unsafe
bugwindowsinstall
|
2025-02-18 |
52 |
| #3185 |
ollama doesn't distribute notice licenses in its release artifacts
bug
|
2024-03-16 |
50 |
| #9674 |
Error: POST predict: Post "http://127.0.0.1:62622/completion": read tcp 127.0.0.1:62627->127.0.0.1:62622: wsarecv: The remote host has closed a connection.
bugneeds more info
|
2025-03-12 |
49 |
| #11714 |
gpt-oss 20b gguf model fail to run
bug
|
2025-08-06 |
48 |
| #11199 |
Request for Support of AMD Ryzen AI Platform NPU
feature request
|
2025-06-25 |
48 |
| #2169 |
Inference with OpenVINO on Intel
feature request
|
2024-01-24 |
47 |
| #12197 |
Some requests get processed on CPU, even though model is loaded in GPU (GPT-OSS)
bug
|
2025-09-06 |
46 |
| #3144 |
add /metrics endpoint
feature requestapi
|
2024-03-14 |
45 |
| #3243 |
Support Steam Deck Docker amdgpu - gfx1033
feature requestamd
|
2024-03-19 |
43 |
| #4643 |
Llama.cpp now supports distributed inference across multiple machines.
feature request
|
2024-05-26 |
42 |
| #6922 |
Support for jinaai/jina-embeddings-v3 embedding model
model
|
2024-09-23 |
40 |
| #5360 |
Support for Snapdragon X Elite NPU & GPU
feature requestwindows
|
2024-06-28 |
40 |
| #2503 |
Support Radeon RX 5700 XT (gfx1010)
documentationamd
|
2024-02-14 |
40 |
| #12962 |
feat: Add Support for Qwen3-vl and Qwen2.5-vl Video Mode
|
2025-11-05 |
38 |
| #10792 |
Gemma 3n
model
|
2025-05-21 |
39 |
| #9506 |
Ollama errors on older versions of Linux/GLIBC on 0.5.13
bug
|
2025-03-05 |
39 |
| #7650 |
AMD Radeon 780M GPU (Pop OS !) System 76
buglinuxamdgpu
|
2024-11-13 |
37 |
| #6364 |
docker container can't detect Nvidia GPU - intermittent "cuda driver library failed to get device context 801"
bugnvidianeeds more infodocker
|
2024-08-14 |
37 |
| #2929 |
Ollama only using half of available CPU cores with NUMA multi-socket systems
buglinuxperformance
|
2024-03-05 |
37 |
| #12606 |
GPT-OSS:20b reasoning loop when reasoning==high
bug
|
2025-10-14 |
36 |
| #12187 |
GPT-OSS not completing tool calls
bug
|
2025-09-04 |
36 |
| #10989 |
support for qwen3-embedding and qwen3-reranker models
model
|
2025-06-05 |
36 |
| #9387 |
phi4 multimodal and mini instruct support
model
|
2025-02-27 |
36 |
| #4072 |
Ollama should prevent sleep when working.
feature requestgood first issuewindows
|
2024-05-01 |
36 |
| #228 |
Please don’t clutter the user home directory
feature request
|
2023-07-27 |
36 |
| #12149 |
[Model Request] Support new Apertus model
model
|
2025-09-02 |
34 |
| #14116 |
Tiered context length can exhaust VRAM
bug
|
2026-02-06 |
33 |
| #9503 |
NVIDIA GPU drivers not loaded on Jeston Orin Nano
buglinuxnvidia
|
2025-03-04 |
33 |
| #7288 |
embedding generation failed. wsarecv: An existing connection was forcibly closed by the remote host.
bug
|
2024-10-21 |
33 |
| #6958 |
molmo by allen ai support
model
|
2024-09-25 |
32 |
| #3113 |
Integrated Intel GPU support
feature requestintel
|
2024-03-13 |
33 |
| #335 |
Model import/export
feature request
|
2023-08-11 |
33 |
| #9639 |
Unsupported Value NaN in Ollama log
bug
|
2025-03-11 |
32 |
| #3107 |
Windows Rocm: HSA_OVERRIDE_GFX_VERSION doesn´t work
bugwindowsamd
|
2024-03-13 |
32 |
| #3004 |
Does ollama support accelerated running on npu?
feature request
|
2024-03-08 |
32 |
| #10458 |
Qwen3 MoE 30b-a3b, poor performance and Low GPU utilization issue
bug
|
2025-04-29 |
31 |
| #13576 |
feat: Add support for remote providers and OpenAI integration
|
2025-12-27 |
4 |
| #10030 |
Deepseek R1, 671b is faster than 70b
performance
|
2025-03-28 |
30 |
| #8843 |
Ollama current/stable (or built from source) appears broken on AMD MI300x ROCm gfx942
bugamdgpu
|
2025-02-05 |
30 |
| #6262 |
Batch embeddings get progressively worse with larger batches
bug
|
2024-08-08 |
30 |
| #10956 |
Garbage output when running llama3.2-vision:11b
bugneeds more info
|
2025-06-03 |
29 |
| #5747 |
Support to Intel NPU by Intel NPU Acceleration Library
feature requestintel
|
2024-07-17 |
29 |
| #3078 |
Ollama is not using the 100% of RTX4000 VRAM (18 of 20GB)
performancenvidia
|
2024-03-12 |
29 |
| #10331 |
Client2 Feedback
|
2025-04-18 |
28 |
| #9678 |
Unusually high VRAM usage of Gemma 3 27B
bug
|
2025-03-12 |
28 |
| #8605 |
Error fetching ANY model locally
bugnetworking
|
2025-01-27 |
28 |
| #8495 |
Why do I keep getting "@@@@" as responses?
bug
|
2025-01-20 |
28 |
| #8105 |
Digest mismatch for llama3.3
bug
|
2024-12-15 |
28 |
| #12976 |
Performance Regression on Apple Silicon M1: GPU → CPU Fallback in v0.12.9 (works correctly in v0.12.5)
bugperformancemacosneeds more info
|
2025-11-05 |
27 |
| #12064 |
Tool call parsing errors
bug
|
2025-08-25 |
27 |
| #3184 |
Add Video-LLaVA
feature request
|
2024-03-16 |
27 |
| #13543 |
ggml: Fix PowerPC build and enable MMA Optimizations
|
2025-12-22 |
14 |
| #7130 |
GPU VRAM Usage Timeout Warnings on Embeddings Model Load
bugmemory
|
2024-10-08 |
26 |
| #12209 |
Version 11 bombing out and responds with GGGGGGGGGGGGGGG
bugnvidianeeds more info
|
2025-09-07 |
25 |
| #10970 |
Support for MedGemma
model
|
2025-06-04 |
25 |
| #7956 |
Low GPU usage on second GPU
bug
|
2024-12-05 |
25 |
| #6230 |
Add Generate Embedding for Sparse vector
feature request
|
2024-08-07 |
25 |
| #990 |
TPU backend support
feature request
|
2023-11-03 |
25 |
| #914 |
Locally-hosted library
question
|
2023-10-26 |
25 |
| #796 |
Support `ppc64le` architecture
feature request
|
2023-10-16 |
25 |
| #14118 |
MLX Error
bug
|
2026-02-06 |
24 |
| #13547 |
REGRESSION: NVIDIA-Nemotron-Nano-9B-v2 not working.
bug
|
2025-12-23 |
24 |
| #13054 |
Ollama 0.12.10 embedding crash (nomic-embed-text-v1.5 on macOS)
bugmacosembeddings
|
2025-11-11 |
24 |
| #8423 |
save with OLLAMA_MODELS set doesn't work anymore in 0.5.5
bug
|
2025-01-14 |
24 |
| #6537 |
Add metrics endpoint and basic request metrics otel based
|
2024-08-27 |
19 |
| #6294 |
AirLLM integration?
feature request
|
2024-08-09 |
24 |
| #3898 |
Apple Silicone Neural Engine: Core ML model package format
feature request
|
2024-04-25 |
24 |
| #3222 |
Support Grok
model
|
2024-03-18 |
24 |
| #14073 |
New default context lengths will break
|
2026-02-04 |
23 |
| #11772 |
use cpu to offload moe weights to reduce the VRAM usage.
feature request
|
2025-08-07 |
23 |
| #11160 |
Enable Intel GPU support with SYCL backend
|
2025-06-22 |
23 |
| #5629 |
Crashing or gibberish output on 3x Radeon GPUs
bugamdneeds more info
|
2024-07-11 |
23 |
| #13763 |
Getting nonsense text responses using Intel ARC 750 8GB card (running on TrueNas)
bug
|
2026-01-17 |
22 |
| #13297 |
Downloading from HuggingFace asks me for a token
bug
|
2025-12-02 |
22 |
| #12940 |
Ollama Runner Fails with “Exit Status 2” and Random Non-Responsive Behavior on Windows
bug
|
2025-11-04 |
22 |
| #12600 |
Continue support for AMD gfx906
feature requestamd
|
2025-10-13 |
22 |
| #9727 |
Please support vision models
model
|
2025-03-13 |
22 |
| #7167 |
Fine-tuned Llama 3.2 1B safe_serialized: Error: json: cannot unmarshal array into Go struct field .model.merges of type string
bugcreate
|
2024-10-10 |
22 |
| #5321 |
Llama3: Generated outputs inconsistent despite seed and temperature
bug
|
2024-06-27 |
22 |
| #3606 |
multilingual-e5-large and multilingual-e5-base Embedding Model Support
model
|
2024-04-11 |
22 |