url
stringlengths
51
54
repository_url
stringclasses
1 value
labels_url
stringlengths
65
68
comments_url
stringlengths
60
63
events_url
stringlengths
58
61
html_url
stringlengths
39
44
id
int64
1.78B
2.82B
node_id
stringlengths
18
19
number
int64
1
8.69k
title
stringlengths
1
382
user
dict
labels
listlengths
0
5
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
2
milestone
null
comments
int64
0
323
created_at
timestamp[s]
updated_at
timestamp[s]
closed_at
timestamp[s]
author_association
stringclasses
4 values
sub_issues_summary
dict
active_lock_reason
null
draft
bool
2 classes
pull_request
dict
body
stringlengths
2
118k
closed_by
dict
reactions
dict
timeline_url
stringlengths
60
63
performed_via_github_app
null
state_reason
stringclasses
4 values
is_pull_request
bool
2 classes
https://api.github.com/repos/ollama/ollama/issues/5614
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5614/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5614/comments
https://api.github.com/repos/ollama/ollama/issues/5614/events
https://github.com/ollama/ollama/pull/5614
2,401,766,631
PR_kwDOJ0Z1Ps51BVsT
5,614
OpenAI: Support Tools
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjha...
[]
closed
false
null
[]
null
1
2024-07-10T21:21:37
2024-07-23T23:02:27
2024-07-17T03:52:59
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5614", "html_url": "https://github.com/ollama/ollama/pull/5614", "diff_url": "https://github.com/ollama/ollama/pull/5614.diff", "patch_url": "https://github.com/ollama/ollama/pull/5614.patch", "merged_at": "2024-07-17T03:52:59" }
``` curl -s 127.0.0.1:11434/v1/chat/completions -d '{ "model": "mike/mistral", "messages": [ { "role": "user", "content": "What'\''s the weather like today in Paris?" }, { "role": "assistant", "tool_calls": [ { "id": "89a1e453-0bce-4de3-a456-c54bed09...
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjha...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5614/reactions", "total_count": 8, "+1": 0, "-1": 0, "laugh": 0, "hooray": 8, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5614/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/700
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/700/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/700/comments
https://api.github.com/repos/ollama/ollama/issues/700/events
https://github.com/ollama/ollama/pull/700
1,926,883,924
PR_kwDOJ0Z1Ps5b77h7
700
rename server subprocess
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
[]
closed
false
null
[]
null
0
2023-10-04T19:53:55
2023-10-06T14:15:43
2023-10-06T14:15:42
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/700", "html_url": "https://github.com/ollama/ollama/pull/700", "diff_url": "https://github.com/ollama/ollama/pull/700.diff", "patch_url": "https://github.com/ollama/ollama/pull/700.patch", "merged_at": "2023-10-06T14:15:42" }
rename llama.cpp `server.exe` to `ollama-runner`. This makes it easier to see that the subprocess is associated with ollama.
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/700/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/700/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8647
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8647/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8647/comments
https://api.github.com/repos/ollama/ollama/issues/8647/events
https://github.com/ollama/ollama/issues/8647
2,817,099,829
I_kwDOJ0Z1Ps6n6Yg1
8,647
ollama[2075]: cuda driver library failed to get device context 999time=2025-01-29T03:21:43.859+01:00 level=WARN source=gpu.go:449 msg="error looking up nvidia GPU memory"
{ "login": "kurdo2000", "id": 137941417, "node_id": "U_kgDOCDjRqQ", "avatar_url": "https://avatars.githubusercontent.com/u/137941417?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kurdo2000", "html_url": "https://github.com/kurdo2000", "followers_url": "https://api.github.com/users/kurdo2...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg...
closed
false
null
[]
null
2
2025-01-29T02:34:23
2025-01-30T03:09:02
2025-01-30T03:09:01
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Hi, is my GPU not supported or what's the error? +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 550.120 Driver Version: 550.120 CUDA Version: 12.4 | |-----------------------------------------+-----------------...
{ "login": "kurdo2000", "id": 137941417, "node_id": "U_kgDOCDjRqQ", "avatar_url": "https://avatars.githubusercontent.com/u/137941417?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kurdo2000", "html_url": "https://github.com/kurdo2000", "followers_url": "https://api.github.com/users/kurdo2...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8647/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8647/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8067
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8067/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8067/comments
https://api.github.com/repos/ollama/ollama/issues/8067/events
https://github.com/ollama/ollama/issues/8067
2,735,177,482
I_kwDOJ0Z1Ps6jB38K
8,067
have a ques
{ "login": "17692019663", "id": 57002321, "node_id": "MDQ6VXNlcjU3MDAyMzIx", "avatar_url": "https://avatars.githubusercontent.com/u/57002321?v=4", "gravatar_id": "", "url": "https://api.github.com/users/17692019663", "html_url": "https://github.com/17692019663", "followers_url": "https://api.github.com/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg...
closed
false
null
[]
null
2
2024-12-12T08:17:34
2024-12-23T08:12:10
2024-12-23T08:12:10
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? C:\Users\Administrator> docker exec -it ollama ollama run llama3.2:1b pulling manifest pulling 74701a8c35f6... 100% ▕████████████████████████████████████████████████████████▏ 1.3 GB pulling 966de95ca8a6... 100% ▕████████████████████████████████████████████████████████▏ 1.4 KB pulling fcc5a6b...
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8067/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8067/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/134
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/134/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/134/comments
https://api.github.com/repos/ollama/ollama/issues/134/events
https://github.com/ollama/ollama/pull/134
1,813,267,553
PR_kwDOJ0Z1Ps5V9on7
134
Update images.go
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
0
2023-07-20T06:34:19
2023-07-20T06:46:19
2023-07-20T06:46:02
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/134", "html_url": "https://github.com/ollama/ollama/pull/134", "diff_url": "https://github.com/ollama/ollama/pull/134.diff", "patch_url": "https://github.com/ollama/ollama/pull/134.patch", "merged_at": "2023-07-20T06:46:02" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/134/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/134/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3226
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3226/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3226/comments
https://api.github.com/repos/ollama/ollama/issues/3226/events
https://github.com/ollama/ollama/issues/3226
2,192,612,235
I_kwDOJ0Z1Ps6CsJuL
3,226
Cannot create Modelfile on Windows Shell
{ "login": "worras2019", "id": 52268635, "node_id": "MDQ6VXNlcjUyMjY4NjM1", "avatar_url": "https://avatars.githubusercontent.com/u/52268635?v=4", "gravatar_id": "", "url": "https://api.github.com/users/worras2019", "html_url": "https://github.com/worras2019", "followers_url": "https://api.github.com/use...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-03-18T15:57:03
2024-03-19T08:50:15
2024-03-19T08:50:15
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? The following is not working in WIndows Shell. Windows Shell doesn't recognize FROM. -------------------------------------------------- FROM llama2 # set the temperature to 1 [higher is more creative, lower is more coherent] PARAMETER temperature 1 # set the system message SYSTEM """ ...
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3226/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3226/timeline
null
not_planned
false
https://api.github.com/repos/ollama/ollama/issues/2788
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2788/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2788/comments
https://api.github.com/repos/ollama/ollama/issues/2788/events
https://github.com/ollama/ollama/issues/2788
2,157,663,506
I_kwDOJ0Z1Ps6Am1US
2,788
Bug: LLaVA 1.6 34b not respecting initial user prompt
{ "login": "mobilemike", "id": 30590, "node_id": "MDQ6VXNlcjMwNTkw", "avatar_url": "https://avatars.githubusercontent.com/u/30590?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mobilemike", "html_url": "https://github.com/mobilemike", "followers_url": "https://api.github.com/users/mobilem...
[]
closed
false
null
[]
null
0
2024-02-27T21:47:53
2024-02-29T19:30:15
2024-02-29T19:30:15
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
M2 Max MBP 96GB RAM Ollama 0.1.27 Compared against llama.cpp CLI @b11a93d (same as Ollama version) Problem: When using the Ollama CLI or API with an image, the initial text prompt isn't respected. Examples like the one on the [model page](https://ollama.com/library/llava) typically show prompts like "What is in t...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2788/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2788/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4205
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4205/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4205/comments
https://api.github.com/repos/ollama/ollama/issues/4205/events
https://github.com/ollama/ollama/issues/4205
2,281,206,136
I_kwDOJ0Z1Ps6H-HF4
4,205
please support deepseek-ai/DeepSeek-V2-Chat
{ "login": "taozhiyuai", "id": 146583103, "node_id": "U_kgDOCLyuPw", "avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4", "gravatar_id": "", "url": "https://api.github.com/users/taozhiyuai", "html_url": "https://github.com/taozhiyuai", "followers_url": "https://api.github.com/users/tao...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
1
2024-05-06T15:54:12
2024-05-08T06:12:39
2024-05-08T06:12:39
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
please support deepseek-ai/DeepSeek-V2-Chat https://hf-mirror.com/deepseek-ai/DeepSeek-V2-Chat/tree/main
{ "login": "taozhiyuai", "id": 146583103, "node_id": "U_kgDOCLyuPw", "avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4", "gravatar_id": "", "url": "https://api.github.com/users/taozhiyuai", "html_url": "https://github.com/taozhiyuai", "followers_url": "https://api.github.com/users/tao...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4205/reactions", "total_count": 3, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 3, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4205/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2058
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2058/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2058/comments
https://api.github.com/repos/ollama/ollama/issues/2058/events
https://github.com/ollama/ollama/issues/2058
2,089,071,525
I_kwDOJ0Z1Ps58hLOl
2,058
Prompt Eval Count is 1 when image is included in multimodal request
{ "login": "Dillon-Yun", "id": 40805060, "node_id": "MDQ6VXNlcjQwODA1MDYw", "avatar_url": "https://avatars.githubusercontent.com/u/40805060?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Dillon-Yun", "html_url": "https://github.com/Dillon-Yun", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
1
2024-01-18T21:24:42
2024-10-23T17:46:39
2024-10-23T17:46:38
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
When I hit the `/api/generate` endpoint for a multimodal model, such as bakllava, I am seeing that `prompt_eval_count == 1` when an image is included. Is this correct behavior? How can I retrieve the total amount of tokens in the request for a multimodal request?
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2058/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2058/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5880
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5880/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5880/comments
https://api.github.com/repos/ollama/ollama/issues/5880/events
https://github.com/ollama/ollama/issues/5880
2,425,575,429
I_kwDOJ0Z1Ps6Qk1gF
5,880
Make Ollama available on Chocolatey
{ "login": "b-a0", "id": 25707742, "node_id": "MDQ6VXNlcjI1NzA3NzQy", "avatar_url": "https://avatars.githubusercontent.com/u/25707742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/b-a0", "html_url": "https://github.com/b-a0", "followers_url": "https://api.github.com/users/b-a0/followers"...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 5860134234, "node_id": ...
open
false
null
[]
null
2
2024-07-23T16:04:09
2024-09-05T19:52:37
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
The Windows installers works great, but requires that I manually download/update it from the ollama website. Would it be possible, probably when the Windows version has matured some more, to add ollama to [Chocolatey](https://community.chocolatey.org/)? That will probabably aid in making this an (even) easier to instal...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5880/reactions", "total_count": 4, "+1": 2, "-1": 2, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5880/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2377
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2377/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2377/comments
https://api.github.com/repos/ollama/ollama/issues/2377/events
https://github.com/ollama/ollama/pull/2377
2,121,583,534
PR_kwDOJ0Z1Ps5mL7sV
2,377
Bump llama.cpp to b2081
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
1
2024-02-06T20:09:01
2024-02-07T20:04:41
2024-02-07T20:04:38
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2377", "html_url": "https://github.com/ollama/ollama/pull/2377", "diff_url": "https://github.com/ollama/ollama/pull/2377.diff", "patch_url": "https://github.com/ollama/ollama/pull/2377.patch", "merged_at": "2024-02-07T20:04:38" }
null
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2377/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2377/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3513
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3513/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3513/comments
https://api.github.com/repos/ollama/ollama/issues/3513/events
https://github.com/ollama/ollama/issues/3513
2,229,191,000
I_kwDOJ0Z1Ps6E3sFY
3,513
Stream: false is ignored in as of c863c6a96d01afda37b65f86eb14a1f04a3c7c47
{ "login": "eusebiu", "id": 414589, "node_id": "MDQ6VXNlcjQxNDU4OQ==", "avatar_url": "https://avatars.githubusercontent.com/u/414589?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eusebiu", "html_url": "https://github.com/eusebiu", "followers_url": "https://api.github.com/users/eusebiu/fo...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
4
2024-04-06T10:57:21
2024-04-13T07:22:27
2024-04-13T07:22:26
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Stream: false is ignored ### What did you expect to see? When stream is set to false, the response should contain the full result. ### Steps to reproduce Pull the latest code, compile, run (pull some model), run /api/generate with stream: false -> the result has no content (just so...
{ "login": "eusebiu", "id": 414589, "node_id": "MDQ6VXNlcjQxNDU4OQ==", "avatar_url": "https://avatars.githubusercontent.com/u/414589?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eusebiu", "html_url": "https://github.com/eusebiu", "followers_url": "https://api.github.com/users/eusebiu/fo...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3513/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3513/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8117
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8117/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8117/comments
https://api.github.com/repos/ollama/ollama/issues/8117/events
https://github.com/ollama/ollama/pull/8117
2,742,219,356
PR_kwDOJ0Z1Ps6FWP5c
8,117
Examples(privategpt): update outdated langchain dependencies and add UnstructuredRSTLoader
{ "login": "KaymeKaydex", "id": 43151027, "node_id": "MDQ6VXNlcjQzMTUxMDI3", "avatar_url": "https://avatars.githubusercontent.com/u/43151027?v=4", "gravatar_id": "", "url": "https://api.github.com/users/KaymeKaydex", "html_url": "https://github.com/KaymeKaydex", "followers_url": "https://api.github.com/...
[]
closed
false
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/...
[ { "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "htt...
null
3
2024-12-16T12:12:05
2025-01-13T19:26:24
2025-01-13T19:26:24
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8117", "html_url": "https://github.com/ollama/ollama/pull/8117", "diff_url": "https://github.com/ollama/ollama/pull/8117.diff", "patch_url": "https://github.com/ollama/ollama/pull/8117.patch", "merged_at": null }
This commit updates the langchain dependencies to the latest stable versions and adds the UnstructuredRSTLoader for loading RST files. - Update langchain dependencies to the latest stable versions. - Add UnstructuredRSTLoader for loading RST files.
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8117/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8117/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3074
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3074/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3074/comments
https://api.github.com/repos/ollama/ollama/issues/3074/events
https://github.com/ollama/ollama/issues/3074
2,181,146,552
I_kwDOJ0Z1Ps6CAae4
3,074
CUDA error: an illegal memory access was encountered
{ "login": "lizhichao999", "id": 34128722, "node_id": "MDQ6VXNlcjM0MTI4NzIy", "avatar_url": "https://avatars.githubusercontent.com/u/34128722?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lizhichao999", "html_url": "https://github.com/lizhichao999", "followers_url": "https://api.github.c...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg...
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
6
2024-03-12T09:35:45
2024-07-24T23:14:51
2024-07-24T23:14:51
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
ollama run llama2 When I executed the command: ollama run llama2, an error occurred related to the data information. ![1710236043402](https://github.com/ollama/ollama/assets/34128722/098fb024-f6f6-419e-99a0-354d29eecd2d) I only have a 128G GPU. Do I need more than 2 GPUs? I can run it on my local laptop with 2 GPU...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3074/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3074/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7704
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7704/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7704/comments
https://api.github.com/repos/ollama/ollama/issues/7704/events
https://github.com/ollama/ollama/issues/7704
2,665,123,889
I_kwDOJ0Z1Ps6e2pAx
7,704
On Windows 11 pro, it does work to right click "restart to update"
{ "login": "jack-liew", "id": 102201538, "node_id": "U_kgDOBhd4wg", "avatar_url": "https://avatars.githubusercontent.com/u/102201538?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jack-liew", "html_url": "https://github.com/jack-liew", "followers_url": "https://api.github.com/users/jack-l...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg...
closed
false
null
[]
null
1
2024-11-17T00:32:03
2024-11-18T22:39:53
2024-11-18T22:39:53
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? It does work to right click "restart to update". Thanks fo great work. ---------------------- ### OS Windows ### GPU Intel ### CPU Intel ### Ollama version 0.4.1
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7704/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7704/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8642
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8642/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8642/comments
https://api.github.com/repos/ollama/ollama/issues/8642/events
https://github.com/ollama/ollama/pull/8642
2,816,895,956
PR_kwDOJ0Z1Ps6JSssP
8,642
Update README with DeepSeek model in the examples
{ "login": "harsha-s", "id": 29120299, "node_id": "MDQ6VXNlcjI5MTIwMjk5", "avatar_url": "https://avatars.githubusercontent.com/u/29120299?v=4", "gravatar_id": "", "url": "https://api.github.com/users/harsha-s", "html_url": "https://github.com/harsha-s", "followers_url": "https://api.github.com/users/har...
[]
open
false
null
[]
null
0
2025-01-28T23:10:31
2025-01-28T23:15:57
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8642", "html_url": "https://github.com/ollama/ollama/pull/8642", "diff_url": "https://github.com/ollama/ollama/pull/8642.diff", "patch_url": "https://github.com/ollama/ollama/pull/8642.patch", "merged_at": null }
Updated the README with the an example of DeepSeek model
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8642/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8642/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7306
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7306/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7306/comments
https://api.github.com/repos/ollama/ollama/issues/7306/events
https://github.com/ollama/ollama/pull/7306
2,603,930,359
PR_kwDOJ0Z1Ps5_Y0rC
7,306
integration: harden embedding test
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
0
2024-10-22T00:15:14
2024-10-22T22:25:25
2024-10-22T22:25:22
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7306", "html_url": "https://github.com/ollama/ollama/pull/7306", "diff_url": "https://github.com/ollama/ollama/pull/7306.diff", "patch_url": "https://github.com/ollama/ollama/pull/7306.patch", "merged_at": "2024-10-22T22:25:22" }
Use cosine similarity to make the embeddings tests more robust
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7306/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7306/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1671
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1671/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1671/comments
https://api.github.com/repos/ollama/ollama/issues/1671/events
https://github.com/ollama/ollama/issues/1671
2,053,904,792
I_kwDOJ0Z1Ps56bBmY
1,671
Error: connect ECONNREFUSED 127.0.0.1:11434
{ "login": "LTtt456c", "id": 53040718, "node_id": "MDQ6VXNlcjUzMDQwNzE4", "avatar_url": "https://avatars.githubusercontent.com/u/53040718?v=4", "gravatar_id": "", "url": "https://api.github.com/users/LTtt456c", "html_url": "https://github.com/LTtt456c", "followers_url": "https://api.github.com/users/LTt...
[]
closed
false
null
[]
null
8
2023-12-22T13:12:25
2024-06-14T10:53:14
2023-12-29T13:49:09
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hello everyone! My ollama in My docker docker Start ollama command is docker run -e OLLAMA_HOST=0.0.0.0:11434 -d -v ollama serve -p 11434:11434 --name ollama ollama/ollama Then I in vscode open chatbot-ollama And then input npm run dev And then Report an error ↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓ Here is the error lo...
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1671/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1671/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/756
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/756/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/756/comments
https://api.github.com/repos/ollama/ollama/issues/756/events
https://github.com/ollama/ollama/issues/756
1,938,044,665
I_kwDOJ0Z1Ps5zhDb5
756
Mistral - Failed To Load Model
{ "login": "mattdavenport", "id": 1127393, "node_id": "MDQ6VXNlcjExMjczOTM=", "avatar_url": "https://avatars.githubusercontent.com/u/1127393?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mattdavenport", "html_url": "https://github.com/mattdavenport", "followers_url": "https://api.github....
[]
closed
false
null
[]
null
2
2023-10-11T15:08:01
2023-10-11T15:40:14
2023-10-11T15:10:58
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I'm running macOS (Ventura 13.0.1) 16in. M1 2021. I am able to run all of the llama2 models just fine, but the following occurs when attempting to run the mistral model: ``` ~ % ollama pull mistral:latest ...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/756/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/756/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2363
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2363/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2363/comments
https://api.github.com/repos/ollama/ollama/issues/2363/events
https://github.com/ollama/ollama/pull/2363
2,119,594,984
PR_kwDOJ0Z1Ps5mFH5O
2,363
Move hub auth out to new package
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
1
2024-02-05T22:39:49
2024-02-15T17:42:21
2024-02-15T17:42:17
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2363", "html_url": "https://github.com/ollama/ollama/pull/2363", "diff_url": "https://github.com/ollama/ollama/pull/2363.diff", "patch_url": "https://github.com/ollama/ollama/pull/2363.patch", "merged_at": null }
null
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2363/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2363/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2514
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2514/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2514/comments
https://api.github.com/repos/ollama/ollama/issues/2514/events
https://github.com/ollama/ollama/issues/2514
2,136,760,146
I_kwDOJ0Z1Ps5_XF9S
2,514
Support for safetensors
{ "login": "gracehubai", "id": 159539863, "node_id": "U_kgDOCYJilw", "avatar_url": "https://avatars.githubusercontent.com/u/159539863?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gracehubai", "html_url": "https://github.com/gracehubai", "followers_url": "https://api.github.com/users/gra...
[]
closed
false
null
[]
null
3
2024-02-15T15:02:55
2024-02-21T05:38:43
2024-02-18T06:49:05
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Do we already support inferencing safetensors?
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2514/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2514/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2214
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2214/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2214/comments
https://api.github.com/repos/ollama/ollama/issues/2214/events
https://github.com/ollama/ollama/pull/2214
2,102,794,214
PR_kwDOJ0Z1Ps5lMKh3
2,214
Detect lack of AVX and fallback to CPU mode
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
0
2024-01-26T19:42:11
2024-01-26T20:06:47
2024-01-26T20:06:44
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2214", "html_url": "https://github.com/ollama/ollama/pull/2214", "diff_url": "https://github.com/ollama/ollama/pull/2214.diff", "patch_url": "https://github.com/ollama/ollama/pull/2214.patch", "merged_at": "2024-01-26T20:06:44" }
We build the GPU libraries with AVX enabled to ensure that if not all layers fit on the GPU we get better performance in a mixed mode. If the user is using a virtualization/emulation system that lacks AVX this used to result in an illegal instruction error and crash before this fix. Now we will report a warning in the...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2214/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2214/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8431
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8431/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8431/comments
https://api.github.com/repos/ollama/ollama/issues/8431/events
https://github.com/ollama/ollama/issues/8431
2,788,732,623
I_kwDOJ0Z1Ps6mOK7P
8,431
Embedding with bge-large:latest fails with "GGML_ASSERT(i01 >= 0 && i01 < ne01) failed"
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/...
[]
open
false
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[ { "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.g...
null
3
2025-01-15T03:07:10
2025-01-15T06:35:56
null
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
> @rick-github Thanks,I tried the "options":{"num_ctx":512} ,It works well! 💯 > but the logs always shows that > > ![Image](https://github.com/user-attachments/assets/ac0c0a66-aa0e-44df-bb25-fc387a09565f) _Originally posted by @Justin-12138 in [#7441](https://github.com/ollama/ollama/issues/7441#issuecomment-2591...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8431/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8431/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/1128
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1128/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1128/comments
https://api.github.com/repos/ollama/ollama/issues/1128/events
https://github.com/ollama/ollama/pull/1128
1,993,443,855
PR_kwDOJ0Z1Ps5fcrpD
1,128
FAQ: answer a few faq questions
{ "login": "technovangelist", "id": 633681, "node_id": "MDQ6VXNlcjYzMzY4MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4", "gravatar_id": "", "url": "https://api.github.com/users/technovangelist", "html_url": "https://github.com/technovangelist", "followers_url": "https://api.git...
[]
closed
false
null
[]
null
0
2023-11-14T19:43:51
2023-11-15T23:05:14
2023-11-15T23:05:13
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1128", "html_url": "https://github.com/ollama/ollama/pull/1128", "diff_url": "https://github.com/ollama/ollama/pull/1128.diff", "patch_url": "https://github.com/ollama/ollama/pull/1128.patch", "merged_at": "2023-11-15T23:05:13" }
null
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/us...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1128/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1128/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4247
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4247/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4247/comments
https://api.github.com/repos/ollama/ollama/issues/4247/events
https://github.com/ollama/ollama/pull/4247
2,284,495,403
PR_kwDOJ0Z1Ps5u0qMP
4,247
skip hidden files in list models handler
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
[]
closed
false
null
[]
null
0
2024-05-08T01:34:13
2024-05-08T02:01:46
2024-05-08T02:01:45
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4247", "html_url": "https://github.com/ollama/ollama/pull/4247", "diff_url": "https://github.com/ollama/ollama/pull/4247.diff", "patch_url": "https://github.com/ollama/ollama/pull/4247.patch", "merged_at": "2024-05-08T02:01:45" }
Hidden files on MacOS (ex: `.DS_Store`) cause the list command to file when not skipped: ``` ❯ ollama ls Error: unqualified name: ```
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4247/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4247/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4025
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4025/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4025/comments
https://api.github.com/repos/ollama/ollama/issues/4025/events
https://github.com/ollama/ollama/pull/4025
2,269,092,269
PR_kwDOJ0Z1Ps5uBD92
4,025
Update linux.md, suggest containerized install
{ "login": "ericcurtin", "id": 1694275, "node_id": "MDQ6VXNlcjE2OTQyNzU=", "avatar_url": "https://avatars.githubusercontent.com/u/1694275?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ericcurtin", "html_url": "https://github.com/ericcurtin", "followers_url": "https://api.github.com/users...
[]
open
false
null
[]
null
2
2024-04-29T13:43:19
2024-05-06T22:54:56
null
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4025", "html_url": "https://github.com/ollama/ollama/pull/4025", "diff_url": "https://github.com/ollama/ollama/pull/4025.diff", "patch_url": "https://github.com/ollama/ollama/pull/4025.patch", "merged_at": null }
It's the simplest install process, once you run the one liner, everything just works.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4025/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4025/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5860
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5860/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5860/comments
https://api.github.com/repos/ollama/ollama/issues/5860/events
https://github.com/ollama/ollama/pull/5860
2,423,755,415
PR_kwDOJ0Z1Ps52I6Yu
5,860
auth: update auth
{ "login": "joshyan1", "id": 76125168, "node_id": "MDQ6VXNlcjc2MTI1MTY4", "avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joshyan1", "html_url": "https://github.com/joshyan1", "followers_url": "https://api.github.com/users/jos...
[]
closed
false
null
[]
null
2
2024-07-22T20:55:20
2024-09-04T13:43:34
2024-09-04T13:43:33
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5860", "html_url": "https://github.com/ollama/ollama/pull/5860", "diff_url": "https://github.com/ollama/ollama/pull/5860.diff", "patch_url": "https://github.com/ollama/ollama/pull/5860.patch", "merged_at": null }
pre cursor to setting up tests for https://github.com/ollama/ollama/pull/5441 moved initialize keypair to auth use initialize keypair implicitly when reading pubkey
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5860/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5860/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/511
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/511/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/511/comments
https://api.github.com/repos/ollama/ollama/issues/511/events
https://github.com/ollama/ollama/issues/511
1,891,545,545
I_kwDOJ0Z1Ps5wvrHJ
511
Support InternLM
{ "login": "vansinhu", "id": 60632596, "node_id": "MDQ6VXNlcjYwNjMyNTk2", "avatar_url": "https://avatars.githubusercontent.com/u/60632596?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vansinhu", "html_url": "https://github.com/vansinhu", "followers_url": "https://api.github.com/users/van...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
6
2023-09-12T02:09:05
2024-06-24T08:28:12
2024-03-11T19:32:53
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Dear ollama developer, Greetings! I am vansinhu, a community developer and volunteer at InternLM. [InternLM](https://github.com/InternLM/InternLM) is a large language model similar to llama2, and we look forward to InternLM being supported in ollama. If there are any challenges or inquiries regarding support for Inter...
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/511/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/511/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8383
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8383/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8383/comments
https://api.github.com/repos/ollama/ollama/issues/8383/events
https://github.com/ollama/ollama/pull/8383
2,781,700,019
PR_kwDOJ0Z1Ps6Hah72
8,383
fix: run container as non-root user
{ "login": "chaturvedi-kna", "id": 63336082, "node_id": "MDQ6VXNlcjYzMzM2MDgy", "avatar_url": "https://avatars.githubusercontent.com/u/63336082?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chaturvedi-kna", "html_url": "https://github.com/chaturvedi-kna", "followers_url": "https://api.gi...
[]
open
false
null
[]
null
1
2025-01-11T08:28:55
2025-01-19T02:54:33
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8383", "html_url": "https://github.com/ollama/ollama/pull/8383", "diff_url": "https://github.com/ollama/ollama/pull/8383.diff", "patch_url": "https://github.com/ollama/ollama/pull/8383.patch", "merged_at": null }
**Enable non-root user Container Deployment Support** **Description** This PR adds support for running ollama in rootless containers by properly setting up permissions for the `.ollama` directory. This change enables ollama to run in environments that enforce non-root execution such as: * OpenShift * Kubernet...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8383/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8383/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5306
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5306/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5306/comments
https://api.github.com/repos/ollama/ollama/issues/5306/events
https://github.com/ollama/ollama/pull/5306
2,375,887,192
PR_kwDOJ0Z1Ps5zqhZW
5,306
Do not reinstall the CLI tools if they are already installed on macOS
{ "login": "seanchristians", "id": 25487785, "node_id": "MDQ6VXNlcjI1NDg3Nzg1", "avatar_url": "https://avatars.githubusercontent.com/u/25487785?v=4", "gravatar_id": "", "url": "https://api.github.com/users/seanchristians", "html_url": "https://github.com/seanchristians", "followers_url": "https://api.gi...
[]
closed
false
null
[]
null
2
2024-06-26T17:17:30
2024-09-06T17:49:40
2024-09-05T04:41:02
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5306", "html_url": "https://github.com/ollama/ollama/pull/5306", "diff_url": "https://github.com/ollama/ollama/pull/5306.diff", "patch_url": "https://github.com/ollama/ollama/pull/5306.patch", "merged_at": null }
Check `installed()` inside of the `install` function before running the cli setup commands. Fix #5305.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5306/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5306/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6165
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6165/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6165/comments
https://api.github.com/repos/ollama/ollama/issues/6165/events
https://github.com/ollama/ollama/issues/6165
2,447,319,803
I_kwDOJ0Z1Ps6R3yL7
6,165
No devices found using AMD gpus
{ "login": "cinglish", "id": 40436106, "node_id": "MDQ6VXNlcjQwNDM2MTA2", "avatar_url": "https://avatars.githubusercontent.com/u/40436106?v=4", "gravatar_id": "", "url": "https://api.github.com/users/cinglish", "html_url": "https://github.com/cinglish", "followers_url": "https://api.github.com/users/cin...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5755339642, "node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg...
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
2
2024-08-04T22:23:38
2024-09-05T19:03:53
2024-09-05T19:03:53
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Getting the following error when loading models with AMD gpus (Instinct MI60s): ``` rocBLAS error: Could not initialize Tensile host: No devices found ``` I have 4 devices allocated to the container and it seems to still discover them at startup: ``` time=2024-08-04T22:17:48.256Z level=I...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6165/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6165/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2989
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2989/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2989/comments
https://api.github.com/repos/ollama/ollama/issues/2989/events
https://github.com/ollama/ollama/pull/2989
2,174,611,866
PR_kwDOJ0Z1Ps5pAaFv
2,989
CI updates
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
1
2024-03-07T19:26:30
2024-03-09T21:20:27
2024-03-09T21:20:24
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2989", "html_url": "https://github.com/ollama/ollama/pull/2989", "diff_url": "https://github.com/ollama/ollama/pull/2989.diff", "patch_url": "https://github.com/ollama/ollama/pull/2989.patch", "merged_at": null }
Not ready for review yet...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2989/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2989/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/476
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/476/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/476/comments
https://api.github.com/repos/ollama/ollama/issues/476/events
https://github.com/ollama/ollama/pull/476
1,884,716,024
PR_kwDOJ0Z1Ps5ZuHR7
476
tighten up the error string for `ollama show` flags
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
[]
closed
false
null
[]
null
0
2023-09-06T20:37:08
2023-09-06T20:38:50
2023-09-06T20:38:49
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/476", "html_url": "https://github.com/ollama/ollama/pull/476", "diff_url": "https://github.com/ollama/ollama/pull/476.diff", "patch_url": "https://github.com/ollama/ollama/pull/476.patch", "merged_at": "2023-09-06T20:38:49" }
null
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/476/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/476/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3018
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3018/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3018/comments
https://api.github.com/repos/ollama/ollama/issues/3018/events
https://github.com/ollama/ollama/pull/3018
2,177,073,503
PR_kwDOJ0Z1Ps5pIy3-
3,018
Add allowed host middleware and remove `workDir` middleware
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[]
closed
false
null
[]
null
0
2024-03-09T06:14:34
2024-03-09T06:23:48
2024-03-09T06:23:47
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3018", "html_url": "https://github.com/ollama/ollama/pull/3018", "diff_url": "https://github.com/ollama/ollama/pull/3018.diff", "patch_url": "https://github.com/ollama/ollama/pull/3018.patch", "merged_at": "2024-03-09T06:23:47" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3018/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3018/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/638
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/638/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/638/comments
https://api.github.com/repos/ollama/ollama/issues/638/events
https://github.com/ollama/ollama/pull/638
1,918,206,635
PR_kwDOJ0Z1Ps5beytn
638
add cuda to windows generation
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
[]
closed
false
null
[]
null
0
2023-09-28T20:15:47
2023-10-20T16:44:24
2023-09-29T16:53:14
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/638", "html_url": "https://github.com/ollama/ollama/pull/638", "diff_url": "https://github.com/ollama/ollama/pull/638.diff", "patch_url": "https://github.com/ollama/ollama/pull/638.patch", "merged_at": null }
Still a bunch of stuff to do for packaging this for distribution, but I tested building and running this locally and it works along with #637 so may as well get it in.
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/638/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/638/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3616
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3616/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3616/comments
https://api.github.com/repos/ollama/ollama/issues/3616/events
https://github.com/ollama/ollama/issues/3616
2,240,686,081
I_kwDOJ0Z1Ps6FjigB
3,616
Richer grammars
{ "login": "tezlm", "id": 56274831, "node_id": "MDQ6VXNlcjU2Mjc0ODMx", "avatar_url": "https://avatars.githubusercontent.com/u/56274831?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tezlm", "html_url": "https://github.com/tezlm", "followers_url": "https://api.github.com/users/tezlm/follow...
[]
closed
false
null
[]
null
9
2024-04-12T18:23:12
2024-12-05T00:37:28
2024-12-05T00:37:28
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What are you trying to do? Being able to specify grammars is great, but it seems a bit underutilized at the moment. This is mostly a thought dump on how it could be improved from experimentation... ### How should we solve this? - Using llama.cpp grammar directly would be pretty powerful and nice to have - Speci...
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3616/reactions", "total_count": 9, "+1": 9, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3616/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8624
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8624/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8624/comments
https://api.github.com/repos/ollama/ollama/issues/8624/events
https://github.com/ollama/ollama/issues/8624
2,814,669,000
I_kwDOJ0Z1Ps6nxHDI
8,624
Deepseek 80% size reduction
{ "login": "gileneusz", "id": 34601970, "node_id": "MDQ6VXNlcjM0NjAxOTcw", "avatar_url": "https://avatars.githubusercontent.com/u/34601970?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gileneusz", "html_url": "https://github.com/gileneusz", "followers_url": "https://api.github.com/users/...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
1
2025-01-28T05:27:54
2025-01-28T22:28:24
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
New quants done by unsloth.ai: | MoE Bits | Disk Size | Type | Quality | Link | Down_proj | |-----------|-----------|----------|---------|---------------------------------...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8624/reactions", "total_count": 8, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 8, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8624/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/59
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/59/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/59/comments
https://api.github.com/repos/ollama/ollama/issues/59/events
https://github.com/ollama/ollama/issues/59
1,795,166,966
I_kwDOJ0Z1Ps5rABL2
59
generate pauses after about 50 tokens
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2023-07-08T23:42:53
2023-07-13T02:15:17
2023-07-13T02:15:17
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Generation will get paused after about 50 tokens being provided ``` % ollama run orca >>> Write a review of the restaurant "five guys" As an AI assistant, I cannot write a biased or subjective review, but I can provide you with some general information about the restaurant "Five Guys". Five Guys is an American f...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/59/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/59/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5278
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5278/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5278/comments
https://api.github.com/repos/ollama/ollama/issues/5278/events
https://github.com/ollama/ollama/issues/5278
2,373,277,484
I_kwDOJ0Z1Ps6NdVcs
5,278
Is it possible to start llama server through dynamic dependency library?
{ "login": "leeyiding", "id": 49098278, "node_id": "MDQ6VXNlcjQ5MDk4Mjc4", "avatar_url": "https://avatars.githubusercontent.com/u/49098278?v=4", "gravatar_id": "", "url": "https://api.github.com/users/leeyiding", "html_url": "https://github.com/leeyiding", "followers_url": "https://api.github.com/users/...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-06-25T17:43:21
2024-07-03T02:36:33
2024-07-03T02:36:33
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi, I'm trying to run Ollama in [Nanos Unikernel](https://github.com/nanovms/nanos), Unikernel a single-process operating system. I found that in https://github.com/ollama/ollama/commit/58d95cc9bd446a8209e7388a96c70367cbafd653, the running mode of llama server changed from loading dynamic dependency libraries to run...
{ "login": "leeyiding", "id": 49098278, "node_id": "MDQ6VXNlcjQ5MDk4Mjc4", "avatar_url": "https://avatars.githubusercontent.com/u/49098278?v=4", "gravatar_id": "", "url": "https://api.github.com/users/leeyiding", "html_url": "https://github.com/leeyiding", "followers_url": "https://api.github.com/users/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5278/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5278/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1579
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1579/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1579/comments
https://api.github.com/repos/ollama/ollama/issues/1579/events
https://github.com/ollama/ollama/issues/1579
2,046,156,422
I_kwDOJ0Z1Ps559d6G
1,579
Error while running ollama locally.
{ "login": "nehalmathew1996", "id": 82964954, "node_id": "MDQ6VXNlcjgyOTY0OTU0", "avatar_url": "https://avatars.githubusercontent.com/u/82964954?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nehalmathew1996", "html_url": "https://github.com/nehalmathew1996", "followers_url": "https://api...
[]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
11
2023-12-18T09:28:17
2024-04-05T21:14:06
2024-02-01T23:19:46
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
ConnectionError: HTTPConnectionPool(host='localhost', port=11434): Max retries exceeded with url: /api/generate/ (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x000002298AE1EF50>: Failed to establish a new connection: [WinError 10061] No connection could be made because the target machine ...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1579/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1579/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7394
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7394/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7394/comments
https://api.github.com/repos/ollama/ollama/issues/7394/events
https://github.com/ollama/ollama/issues/7394
2,617,770,363
I_kwDOJ0Z1Ps6cCAF7
7,394
The Open WebUI generate unsense text, but the cli terminal can chat normally with the same hosted model
{ "login": "QiuJYWX", "id": 68810978, "node_id": "MDQ6VXNlcjY4ODEwOTc4", "avatar_url": "https://avatars.githubusercontent.com/u/68810978?v=4", "gravatar_id": "", "url": "https://api.github.com/users/QiuJYWX", "html_url": "https://github.com/QiuJYWX", "followers_url": "https://api.github.com/users/QiuJYW...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q...
closed
false
null
[]
null
9
2024-10-28T09:06:43
2024-11-13T22:12:46
2024-11-13T22:12:46
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? The results generated by cli terminal: ![cli](https://github.com/user-attachments/assets/aa1ea3c5-bcd0-45e0-b9e4-480aa81036ec) The results generated by open webui: ![webui](https://github.com/user-attachments/assets/81f236c9-517c-4fc5-a3e3-6c9f284c7555) ![webui2](https://github.com/user-at...
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7394/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7394/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8316
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8316/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8316/comments
https://api.github.com/repos/ollama/ollama/issues/8316/events
https://github.com/ollama/ollama/issues/8316
2,770,077,246
I_kwDOJ0Z1Ps6lHAY-
8,316
Feature Request: Limit Ollama to use only GPU RAM (VRAM), not system RAM
{ "login": "mlibre", "id": 8473036, "node_id": "MDQ6VXNlcjg0NzMwMzY=", "avatar_url": "https://avatars.githubusercontent.com/u/8473036?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mlibre", "html_url": "https://github.com/mlibre", "followers_url": "https://api.github.com/users/mlibre/foll...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
4
2025-01-06T08:02:41
2025-01-24T09:45:19
2025-01-24T09:45:19
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### Feature Request **Description**: I would like to request a feature that allows Ollama to be configured to use only GPU RAM, without utilizing CPU or system RAM. This would be beneficial for systems with limited RAM resources and high GPU RAM (VRAM) memory availability. **Benefits**: - Optimized performance ...
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8316/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8316/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8562
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8562/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8562/comments
https://api.github.com/repos/ollama/ollama/issues/8562/events
https://github.com/ollama/ollama/issues/8562
2,809,089,096
I_kwDOJ0Z1Ps6nb0xI
8,562
None value being returned from with_structured_output request
{ "login": "jonmach", "id": 7843429, "node_id": "MDQ6VXNlcjc4NDM0Mjk=", "avatar_url": "https://avatars.githubusercontent.com/u/7843429?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jonmach", "html_url": "https://github.com/jonmach", "followers_url": "https://api.github.com/users/jonmach/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
6
2025-01-24T10:26:05
2025-01-27T10:15:42
2025-01-24T13:27:20
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Python versions are: langchain 0.3.15 langchain-community 0.3.15 langchain-core 0.3.31 langchain-ollama 0.2.2 ollama 0.4.7 Running ollama 0.5.7 (pip install -U ollama did not increase the version beyond 0.4.7) --- Using **with_structured_output()** seems to work for a very simple example suc...
{ "login": "jonmach", "id": 7843429, "node_id": "MDQ6VXNlcjc4NDM0Mjk=", "avatar_url": "https://avatars.githubusercontent.com/u/7843429?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jonmach", "html_url": "https://github.com/jonmach", "followers_url": "https://api.github.com/users/jonmach/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8562/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8562/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/130
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/130/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/130/comments
https://api.github.com/repos/ollama/ollama/issues/130/events
https://github.com/ollama/ollama/pull/130
1,812,847,886
PR_kwDOJ0Z1Ps5V8NI1
130
vendor in progress bar and change to bytes instead of bibytes
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
[]
closed
false
null
[]
null
0
2023-07-19T22:54:28
2023-07-20T00:24:03
2023-07-20T00:24:03
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/130", "html_url": "https://github.com/ollama/ollama/pull/130", "diff_url": "https://github.com/ollama/ollama/pull/130.diff", "patch_url": "https://github.com/ollama/ollama/pull/130.patch", "merged_at": "2023-07-20T00:24:03" }
null
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/130/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/130/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4611
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4611/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4611/comments
https://api.github.com/repos/ollama/ollama/issues/4611/events
https://github.com/ollama/ollama/issues/4611
2,314,861,010
I_kwDOJ0Z1Ps6J-fnS
4,611
User comments on personal model page
{ "login": "razvanab", "id": 2854730, "node_id": "MDQ6VXNlcjI4NTQ3MzA=", "avatar_url": "https://avatars.githubusercontent.com/u/2854730?v=4", "gravatar_id": "", "url": "https://api.github.com/users/razvanab", "html_url": "https://github.com/razvanab", "followers_url": "https://api.github.com/users/razva...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 6573197867, "node_id": ...
open
false
null
[]
null
0
2024-05-24T09:05:07
2024-07-08T17:22:09
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I would like to suggest a section for user comments on the personal Ollama model page. This will help with better tuning the parameters for your version of the model.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4611/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4611/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/4287
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4287/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4287/comments
https://api.github.com/repos/ollama/ollama/issues/4287/events
https://github.com/ollama/ollama/pull/4287
2,287,796,918
PR_kwDOJ0Z1Ps5u_yAg
4,287
suppress ollama_llama_server.exe blank command window popup
{ "login": "ashokgelal", "id": 401055, "node_id": "MDQ6VXNlcjQwMTA1NQ==", "avatar_url": "https://avatars.githubusercontent.com/u/401055?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ashokgelal", "html_url": "https://github.com/ashokgelal", "followers_url": "https://api.github.com/users/a...
[]
closed
false
null
[]
null
10
2024-05-09T14:19:20
2025-01-29T16:31:20
2024-11-23T21:29:08
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4287", "html_url": "https://github.com/ollama/ollama/pull/4287", "diff_url": "https://github.com/ollama/ollama/pull/4287.diff", "patch_url": "https://github.com/ollama/ollama/pull/4287.patch", "merged_at": null }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4287/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4287/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1827
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1827/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1827/comments
https://api.github.com/repos/ollama/ollama/issues/1827/events
https://github.com/ollama/ollama/issues/1827
2,068,731,697
I_kwDOJ0Z1Ps57Tlcx
1,827
Massive slowdown on v 0.1.18 vs 0.1.17 with same model on Intel Mac
{ "login": "pjv", "id": 327716, "node_id": "MDQ6VXNlcjMyNzcxNg==", "avatar_url": "https://avatars.githubusercontent.com/u/327716?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pjv", "html_url": "https://github.com/pjv", "followers_url": "https://api.github.com/users/pjv/followers", "fol...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
8
2024-01-06T17:18:27
2024-04-15T01:15:54
2024-01-07T05:46:19
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
<img width="255" alt="SCR-20240106-kfri" src="https://github.com/jmorganca/ollama/assets/327716/c1b5ce5b-acd5-4c81-a59a-6db39ff6a257"> I don’t have exact timings but the same model (in this case, `deepseek-coder:6.7b-instruct-q4_K_S`) generates tokens roughly 5 times faster on 0.1.17 than on 0.1.18 on my Intel Mac. ...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1827/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1827/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8108
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8108/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8108/comments
https://api.github.com/repos/ollama/ollama/issues/8108/events
https://github.com/ollama/ollama/issues/8108
2,740,794,394
I_kwDOJ0Z1Ps6jXTQa
8,108
Error: llama runner process has terminated: error:/opt/rocm/lib/libhipblas.so.2: undefined symbol: rocblas_sgbmv_64
{ "login": "dernikolas", "id": 150927066, "node_id": "U_kgDOCP722g", "avatar_url": "https://avatars.githubusercontent.com/u/150927066?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dernikolas", "html_url": "https://github.com/dernikolas", "followers_url": "https://api.github.com/users/der...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5755339642, "node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg...
open
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
2
2024-12-15T17:57:17
2025-01-06T17:54:55
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? When I start llama3.2 with `ollama run llama3.2` this error spawns: `Error: llama runner process has terminated: error:/opt/rocm/lib/libhipblas.so.2: undefined symbol: rocblas_sgbmv_64` I am on Cachy OS with ollama and ollama-rocm installed from extra repo and a AMD RX6800 is the GPU ### OS...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8108/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8108/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6571
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6571/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6571/comments
https://api.github.com/repos/ollama/ollama/issues/6571/events
https://github.com/ollama/ollama/issues/6571
2,497,977,180
I_kwDOJ0Z1Ps6U5Btc
6,571
Impossible to connect to ollama locally from another pc
{ "login": "Wilnox23", "id": 82610998, "node_id": "MDQ6VXNlcjgyNjEwOTk4", "avatar_url": "https://avatars.githubusercontent.com/u/82610998?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Wilnox23", "html_url": "https://github.com/Wilnox23", "followers_url": "https://api.github.com/users/Wil...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
4
2024-08-30T18:45:36
2024-08-30T21:12:58
2024-08-30T21:11:07
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Hello, I'm using ollama on windows but I can't connect to the server from another PC on the local network. I have already changed the following system environment variables. `OLLAMA_HOST=0.0.0.0` `OLLAMA_ORIGINS=*` Thank you for your assistance. ### OS Windows ### GPU Nvidia ### CPU In...
{ "login": "Wilnox23", "id": 82610998, "node_id": "MDQ6VXNlcjgyNjEwOTk4", "avatar_url": "https://avatars.githubusercontent.com/u/82610998?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Wilnox23", "html_url": "https://github.com/Wilnox23", "followers_url": "https://api.github.com/users/Wil...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6571/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6571/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2693
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2693/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2693/comments
https://api.github.com/repos/ollama/ollama/issues/2693/events
https://github.com/ollama/ollama/issues/2693
2,149,835,982
I_kwDOJ0Z1Ps6AI-TO
2,693
Ollama windows version not working
{ "login": "sdrshn-nmbr", "id": 122426494, "node_id": "U_kgDOB0wUfg", "avatar_url": "https://avatars.githubusercontent.com/u/122426494?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sdrshn-nmbr", "html_url": "https://github.com/sdrshn-nmbr", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
2
2024-02-22T19:55:55
2024-02-22T20:24:32
2024-02-22T20:24:12
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
It was working fine even yesterday, but I got an update notification and it hasn't been working since. I even tried deleting and reinstalling the installer exe, but it seems the app shows up for a few seconds and then disappears again, but powershell still recognizes the command - it just says ollama not running.
{ "login": "sdrshn-nmbr", "id": 122426494, "node_id": "U_kgDOB0wUfg", "avatar_url": "https://avatars.githubusercontent.com/u/122426494?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sdrshn-nmbr", "html_url": "https://github.com/sdrshn-nmbr", "followers_url": "https://api.github.com/users/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2693/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2693/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5758
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5758/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5758/comments
https://api.github.com/repos/ollama/ollama/issues/5758/events
https://github.com/ollama/ollama/pull/5758
2,414,681,278
PR_kwDOJ0Z1Ps51sq18
5,758
marshal json automatically for some template values
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
0
2024-07-17T22:02:46
2024-07-17T22:35:13
2024-07-17T22:35:11
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5758", "html_url": "https://github.com/ollama/ollama/pull/5758", "diff_url": "https://github.com/ollama/ollama/pull/5758.diff", "patch_url": "https://github.com/ollama/ollama/pull/5758.patch", "merged_at": "2024-07-17T22:35:11" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5758/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5758/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3678
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3678/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3678/comments
https://api.github.com/repos/ollama/ollama/issues/3678/events
https://github.com/ollama/ollama/pull/3678
2,246,645,169
PR_kwDOJ0Z1Ps5s1fe1
3,678
darwin: no partial offloading if required memory greater than system
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
0
2024-04-16T18:23:02
2024-04-16T19:05:57
2024-04-16T19:05:57
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3678", "html_url": "https://github.com/ollama/ollama/pull/3678", "diff_url": "https://github.com/ollama/ollama/pull/3678.diff", "patch_url": "https://github.com/ollama/ollama/pull/3678.patch", "merged_at": "2024-04-16T19:05:56" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3678/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3678/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4381
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4381/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4381/comments
https://api.github.com/repos/ollama/ollama/issues/4381/events
https://github.com/ollama/ollama/issues/4381
2,291,487,916
I_kwDOJ0Z1Ps6IlVSs
4,381
Yi 1.5 was released
{ "login": "ddpasa", "id": 112642920, "node_id": "U_kgDOBrbLaA", "avatar_url": "https://avatars.githubusercontent.com/u/112642920?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ddpasa", "html_url": "https://github.com/ddpasa", "followers_url": "https://api.github.com/users/ddpasa/follower...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
2
2024-05-12T20:23:39
2024-05-14T03:35:51
2024-05-13T21:34:04
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
https://github.com/01-ai/Yi-1.5 https://huggingface.co/collections/01-ai/yi-15-2024-05-663f3ecab5f815a3eaca7ca8
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/us...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4381/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4381/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/783
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/783/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/783/comments
https://api.github.com/repos/ollama/ollama/issues/783/events
https://github.com/ollama/ollama/pull/783
1,942,570,643
PR_kwDOJ0Z1Ps5cxVPh
783
fix: offloading on low end GPUs
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
0
2023-10-13T20:08:46
2023-10-13T21:36:45
2023-10-13T21:36:44
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/783", "html_url": "https://github.com/ollama/ollama/pull/783", "diff_url": "https://github.com/ollama/ollama/pull/783.diff", "patch_url": "https://github.com/ollama/ollama/pull/783.patch", "merged_at": "2023-10-13T21:36:44" }
Fixes two issues when using low end GPUs: GPUs with low VRAM are disproportionately affected by overhead when offloading so any device that has less than 2GB VRAM will be exclusively CPU unless overwritten by num_gpu. A CUDA-enabled runner will still offload to GPU even if num_gpu is 0. This is problematic when t...
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/783/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/783/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/861
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/861/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/861/comments
https://api.github.com/repos/ollama/ollama/issues/861/events
https://github.com/ollama/ollama/pull/861
1,954,903,919
PR_kwDOJ0Z1Ps5dakP4
861
Added /clear command to clear the buffer
{ "login": "pepperoni21", "id": 29759371, "node_id": "MDQ6VXNlcjI5NzU5Mzcx", "avatar_url": "https://avatars.githubusercontent.com/u/29759371?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pepperoni21", "html_url": "https://github.com/pepperoni21", "followers_url": "https://api.github.com/...
[]
closed
false
null
[]
null
3
2023-10-20T19:25:06
2023-10-20T20:10:22
2023-10-20T20:05:19
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/861", "html_url": "https://github.com/ollama/ollama/pull/861", "diff_url": "https://github.com/ollama/ollama/pull/861.diff", "patch_url": "https://github.com/ollama/ollama/pull/861.patch", "merged_at": null }
Hey, I thought a /clear command could enhance user experience. It should work for both windows and linux/mac.
{ "login": "pepperoni21", "id": 29759371, "node_id": "MDQ6VXNlcjI5NzU5Mzcx", "avatar_url": "https://avatars.githubusercontent.com/u/29759371?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pepperoni21", "html_url": "https://github.com/pepperoni21", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/861/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/861/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4830
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4830/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4830/comments
https://api.github.com/repos/ollama/ollama/issues/4830/events
https://github.com/ollama/ollama/issues/4830
2,335,295,422
I_kwDOJ0Z1Ps6LMce-
4,830
model import to Ollama
{ "login": "fingerfree77", "id": 124861936, "node_id": "U_kgDOB3E98A", "avatar_url": "https://avatars.githubusercontent.com/u/124861936?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fingerfree77", "html_url": "https://github.com/fingerfree77", "followers_url": "https://api.github.com/use...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
3
2024-06-05T08:54:55
2024-09-04T13:31:21
2024-09-04T13:31:21
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Is it possible to import model to Ollama from HuggingFace which is not listed in the script convert-hf-to-gguf.py? In my case model: https://huggingface.co/ai-forever/ruGPT-3.5-13B When I try import got error: ``` Traceback (most recent call last): File "llm/llama.cpp/convert-hf-to-gguf.py", line 2865, in <modu...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4830/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4830/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3569
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3569/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3569/comments
https://api.github.com/repos/ollama/ollama/issues/3569/events
https://github.com/ollama/ollama/issues/3569
2,234,581,592
I_kwDOJ0Z1Ps6FMQJY
3,569
Could anyone tell me the purpose of "ollama-linux-amd64-rocm.tgz" in the released files?
{ "login": "airforce011", "id": 36444699, "node_id": "MDQ6VXNlcjM2NDQ0Njk5", "avatar_url": "https://avatars.githubusercontent.com/u/36444699?v=4", "gravatar_id": "", "url": "https://api.github.com/users/airforce011", "html_url": "https://github.com/airforce011", "followers_url": "https://api.github.com/...
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
1
2024-04-10T01:29:39
2024-04-23T00:02:08
2024-04-23T00:02:08
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What are you trying to do? I'm trying to run Ollama on Linux (Ubuntu 22.04) with ROCm, but I'm unable to use it even after following the installation instructions in the documentation. Do I need to use the "ollama-linux-amd64-rocm.tgz" file? However, I'm not sure how to use this file, and there are no instructions...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3569/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3569/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4963
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4963/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4963/comments
https://api.github.com/repos/ollama/ollama/issues/4963/events
https://github.com/ollama/ollama/issues/4963
2,344,382,940
I_kwDOJ0Z1Ps6LvHHc
4,963
Ollama in private EC2
{ "login": "arpitdish", "id": 123781157, "node_id": "U_kgDOB2DAJQ", "avatar_url": "https://avatars.githubusercontent.com/u/123781157?v=4", "gravatar_id": "", "url": "https://api.github.com/users/arpitdish", "html_url": "https://github.com/arpitdish", "followers_url": "https://api.github.com/users/arpitd...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-06-10T16:40:51
2024-07-10T23:01:21
2024-07-10T23:01:21
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? How to run ollama on private EC2 ### OS Linux ### GPU Nvidia ### CPU _No response_ ### Ollama version _No response_
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4963/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4963/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4290
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4290/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4290/comments
https://api.github.com/repos/ollama/ollama/issues/4290/events
https://github.com/ollama/ollama/pull/4290
2,288,039,651
PR_kwDOJ0Z1Ps5vAm1e
4,290
fix(server): prune files
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
1
2024-05-09T16:25:11
2024-11-23T21:28:14
2024-11-23T21:28:14
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4290", "html_url": "https://github.com/ollama/ollama/pull/4290", "diff_url": "https://github.com/ollama/ollama/pull/4290.diff", "patch_url": "https://github.com/ollama/ollama/pull/4290.patch", "merged_at": null }
follow up to #4269
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4290/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4290/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5627
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5627/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5627/comments
https://api.github.com/repos/ollama/ollama/issues/5627/events
https://github.com/ollama/ollama/issues/5627
2,402,504,053
I_kwDOJ0Z1Ps6PM011
5,627
unknown model architecture: 'gemma2'
{ "login": "mouse33333", "id": 4304174, "node_id": "MDQ6VXNlcjQzMDQxNzQ=", "avatar_url": "https://avatars.githubusercontent.com/u/4304174?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mouse33333", "html_url": "https://github.com/mouse33333", "followers_url": "https://api.github.com/users...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-07-11T07:47:19
2024-07-11T08:02:42
2024-07-11T07:55:05
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I'm trying to run models on my internal Intel GPU. Here's the software env: 1. activate a virtual env in Anaconda 2. already installed ipex-llm[cpp], ipex-llm[xpu] 3. start the Ollama serve 4. start gemma2 model Then the log said: unknown model architecture: 'gemma2' Here's the sre...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5627/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5627/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4011
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4011/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4011/comments
https://api.github.com/repos/ollama/ollama/issues/4011/events
https://github.com/ollama/ollama/issues/4011
2,267,922,429
I_kwDOJ0Z1Ps6HLb_9
4,011
When my ollama has both the gemma and llama large models, how can I enable them at the same time?Thank you.
{ "login": "joylijoy", "id": 167687386, "node_id": "U_kgDOCf602g", "avatar_url": "https://avatars.githubusercontent.com/u/167687386?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joylijoy", "html_url": "https://github.com/joylijoy", "followers_url": "https://api.github.com/users/joylijoy/...
[]
closed
false
null
[]
null
2
2024-04-29T00:54:44
2024-05-01T04:56:32
2024-05-01T04:56:32
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
When my ollama has both the gemma and llama large models, how can I enable them at the same time? Does pressing CTRL+D mean to exit the large model? Thank you.
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4011/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4011/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8457
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8457/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8457/comments
https://api.github.com/repos/ollama/ollama/issues/8457/events
https://github.com/ollama/ollama/issues/8457
2,792,336,253
I_kwDOJ0Z1Ps6mb6t9
8,457
model wanted in ollama please:Qwen2.5-Math-PRM-7B
{ "login": "twythebest", "id": 89891289, "node_id": "MDQ6VXNlcjg5ODkxMjg5", "avatar_url": "https://avatars.githubusercontent.com/u/89891289?v=4", "gravatar_id": "", "url": "https://api.github.com/users/twythebest", "html_url": "https://github.com/twythebest", "followers_url": "https://api.github.com/use...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
0
2025-01-16T10:36:00
2025-01-16T10:36:00
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
model wanted in ollama please:Qwen2.5-Math-PRM-7B
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8457/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8457/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6544
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6544/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6544/comments
https://api.github.com/repos/ollama/ollama/issues/6544/events
https://github.com/ollama/ollama/issues/6544
2,492,824,185
I_kwDOJ0Z1Ps6UlXp5
6,544
Specifying options via openai client extra_body are not handled by ollama
{ "login": "gaardhus", "id": 46934916, "node_id": "MDQ6VXNlcjQ2OTM0OTE2", "avatar_url": "https://avatars.githubusercontent.com/u/46934916?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gaardhus", "html_url": "https://github.com/gaardhus", "followers_url": "https://api.github.com/users/gaa...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
5
2024-08-28T19:19:11
2024-09-05T21:47:15
2024-09-04T13:25:17
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? So I've been trying to set the num_ctx for the mistral-nemo through the openai api client, however it does not seem to have an effect. ```python # Ollama client client = OllamaAsyncClient(host=base_url, **kwargs) chat_completion = await client.chat( mes...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6544/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6544/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7106
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7106/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7106/comments
https://api.github.com/repos/ollama/ollama/issues/7106/events
https://github.com/ollama/ollama/issues/7106
2,568,514,399
I_kwDOJ0Z1Ps6ZGGtf
7,106
Update /api/show endpoint to return a boolean describing if the model supports FIM
{ "login": "tomasz-stefaniak", "id": 7911227, "node_id": "MDQ6VXNlcjc5MTEyMjc=", "avatar_url": "https://avatars.githubusercontent.com/u/7911227?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tomasz-stefaniak", "html_url": "https://github.com/tomasz-stefaniak", "followers_url": "https://ap...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-10-06T07:32:55
2024-11-17T14:49:48
2024-11-17T14:49:48
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi, We support Ollama as a [model provider in Continue.dev](https://docs.continue.dev/customize/model-providers/ollama) and have recently added FIM (Fill-in-the-Middle) support. Currently, there is no API we are aware of that can be used to determine if a model supports FIM, so we guess its capabilities by checking ...
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7106/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7106/timeline
null
not_planned
false
https://api.github.com/repos/ollama/ollama/issues/6749
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6749/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6749/comments
https://api.github.com/repos/ollama/ollama/issues/6749/events
https://github.com/ollama/ollama/issues/6749
2,519,216,246
I_kwDOJ0Z1Ps6WKDB2
6,749
Add version when the docker container is starting
{ "login": "svaningelgem", "id": 199434, "node_id": "MDQ6VXNlcjE5OTQzNA==", "avatar_url": "https://avatars.githubusercontent.com/u/199434?v=4", "gravatar_id": "", "url": "https://api.github.com/users/svaningelgem", "html_url": "https://github.com/svaningelgem", "followers_url": "https://api.github.com/u...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-09-11T09:46:48
2024-09-12T03:50:22
2024-09-11T23:56:59
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
It'd be nice to know what version is running when sharing logs. So my feature request is to add the current version when you start the docker container.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6749/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6749/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4271
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4271/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4271/comments
https://api.github.com/repos/ollama/ollama/issues/4271/events
https://github.com/ollama/ollama/issues/4271
2,286,720,880
I_kwDOJ0Z1Ps6ITJdw
4,271
Partial pruning does not wrok
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
0
2024-05-09T01:17:56
2024-05-09T01:17:56
null
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? ``` time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't get file path for 'sha256:cfcf93119280c4a10c1df57335bad341e000cabbc4faff125531d941a5b0befa:partial:13': invalid digest format" time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't g...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4271/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4271/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/8649
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8649/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8649/comments
https://api.github.com/repos/ollama/ollama/issues/8649/events
https://github.com/ollama/ollama/issues/8649
2,817,207,537
I_kwDOJ0Z1Ps6n6yzx
8,649
Short run response duration calculations are off
{ "login": "NerdyShawn", "id": 16088118, "node_id": "MDQ6VXNlcjE2MDg4MTE4", "avatar_url": "https://avatars.githubusercontent.com/u/16088118?v=4", "gravatar_id": "", "url": "https://api.github.com/users/NerdyShawn", "html_url": "https://github.com/NerdyShawn", "followers_url": "https://api.github.com/use...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
3
2025-01-29T04:22:55
2025-01-29T14:03:03
2025-01-29T14:03:01
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Running the smaller `deepseek-r1:1.5b` model it seems like the very short time and duration how it is being calculated is off given the duration in the response. It seems since its close to zero it throws off how the time is measured. --- ![Image](https://github.com/user-attachments/assets/aba...
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8649/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8649/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2864
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2864/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2864/comments
https://api.github.com/repos/ollama/ollama/issues/2864/events
https://github.com/ollama/ollama/issues/2864
2,163,502,283
I_kwDOJ0Z1Ps6A9GzL
2,864
Read and write files (To allow for more types of AI models in Ollama)
{ "login": "trymeouteh", "id": 31172274, "node_id": "MDQ6VXNlcjMxMTcyMjc0", "avatar_url": "https://avatars.githubusercontent.com/u/31172274?v=4", "gravatar_id": "", "url": "https://api.github.com/users/trymeouteh", "html_url": "https://github.com/trymeouteh", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
4
2024-03-01T14:12:41
2024-03-08T14:09:08
2024-03-07T18:54:15
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
To my understanding it will seem Ollama can only support text chat based models and to my understand it could support more kinds of models by adding a reading files and writing files feature... - A way to have a model read one or more file on your system and be able to send it to the AI model to process. - A way fo...
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2864/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2864/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/502
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/502/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/502/comments
https://api.github.com/repos/ollama/ollama/issues/502/events
https://github.com/ollama/ollama/issues/502
1,889,051,436
I_kwDOJ0Z1Ps5wmKMs
502
Better Document /api/embeddings
{ "login": "FairyTail2000", "id": 22645621, "node_id": "MDQ6VXNlcjIyNjQ1NjIx", "avatar_url": "https://avatars.githubusercontent.com/u/22645621?v=4", "gravatar_id": "", "url": "https://api.github.com/users/FairyTail2000", "html_url": "https://github.com/FairyTail2000", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
5
2023-09-10T08:37:43
2023-10-26T00:12:03
2023-10-26T00:12:03
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
It's currently not very clear on how to use is and how it's relation is to the EMBED Modelfile instruction. Can you enlighten me and/or update the wiki and examples?
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/502/reactions", "total_count": 4, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 2 }
https://api.github.com/repos/ollama/ollama/issues/502/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7634
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7634/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7634/comments
https://api.github.com/repos/ollama/ollama/issues/7634/events
https://github.com/ollama/ollama/issues/7634
2,653,045,312
I_kwDOJ0Z1Ps6eIkJA
7,634
Custom tool are not getting called in tool call
{ "login": "Vikneshkumarmohan", "id": 71374469, "node_id": "MDQ6VXNlcjcxMzc0NDY5", "avatar_url": "https://avatars.githubusercontent.com/u/71374469?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Vikneshkumarmohan", "html_url": "https://github.com/Vikneshkumarmohan", "followers_url": "https...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-11-12T18:51:44
2024-12-03T03:55:02
2024-12-02T08:01:16
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? when using ChatOllama in Langgraph tool call is not happening for the custom tools , but it is working as expected with tools like tavily search , https://langchain-ai.github.io/langgraph/tutorials/multi_agent/multi-agent-collaboration/#invoke above is the example for reference code , when ...
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7634/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/ollama/ollama/issues/7634/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6186
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6186/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6186/comments
https://api.github.com/repos/ollama/ollama/issues/6186/events
https://github.com/ollama/ollama/pull/6186
2,449,377,760
PR_kwDOJ0Z1Ps53fc2u
6,186
Implement linux NUMA detection
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
0
2024-08-05T20:01:08
2024-08-05T22:20:11
2024-08-05T22:20:07
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6186", "html_url": "https://github.com/ollama/ollama/pull/6186", "diff_url": "https://github.com/ollama/ollama/pull/6186.diff", "patch_url": "https://github.com/ollama/ollama/pull/6186.patch", "merged_at": "2024-08-05T22:20:07" }
If the system has multiple numa nodes, enable numa support in llama.cpp If we detect `numactl` in the path, use that, else use the basic "distribute" mode. This also removes the `use_numa` setting as the bool type is no longer useful given this parameter now requires a mode. Fixes #6093 Fixes #2496
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6186/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6186/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5715
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5715/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5715/comments
https://api.github.com/repos/ollama/ollama/issues/5715/events
https://github.com/ollama/ollama/issues/5715
2,410,125,004
I_kwDOJ0Z1Ps6Pp5bM
5,715
Allow using `"""` in TEMPLATE Modelfile command
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-07-16T03:33:31
2024-07-16T03:33:31
null
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
null
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5715/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5715/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/7121
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7121/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7121/comments
https://api.github.com/repos/ollama/ollama/issues/7121/events
https://github.com/ollama/ollama/pull/7121
2,571,261,705
PR_kwDOJ0Z1Ps5925z4
7,121
llama: refine developer docs
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
0
2024-10-07T19:18:10
2024-10-07T19:43:49
2024-10-07T19:43:46
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7121", "html_url": "https://github.com/ollama/ollama/pull/7121", "diff_url": "https://github.com/ollama/ollama/pull/7121.diff", "patch_url": "https://github.com/ollama/ollama/pull/7121.patch", "merged_at": "2024-10-07T19:43:46" }
null
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7121/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7121/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8183
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8183/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8183/comments
https://api.github.com/repos/ollama/ollama/issues/8183/events
https://github.com/ollama/ollama/issues/8183
2,752,749,059
I_kwDOJ0Z1Ps6kE54D
8,183
How do I specify specific GPUs when running a model?
{ "login": "any35", "id": 2082210, "node_id": "MDQ6VXNlcjIwODIyMTA=", "avatar_url": "https://avatars.githubusercontent.com/u/2082210?v=4", "gravatar_id": "", "url": "https://api.github.com/users/any35", "html_url": "https://github.com/any35", "followers_url": "https://api.github.com/users/any35/follower...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-12-20T13:30:03
2025-01-13T01:43:05
2025-01-13T01:43:04
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I have a server with 16 RTX4090 graphics cards installed. Now I need to run different models at the same time and specify the corresponding GPUs. How can I do this? For example: run llama 3.3 on GPU0,1; llama 3.2 on gpu 2;...
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8183/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8183/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1306
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1306/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1306/comments
https://api.github.com/repos/ollama/ollama/issues/1306/events
https://github.com/ollama/ollama/pull/1306
2,014,904,973
PR_kwDOJ0Z1Ps5glFGZ
1,306
Fix Docker image not using GPU
{ "login": "bwest2397", "id": 73724178, "node_id": "MDQ6VXNlcjczNzI0MTc4", "avatar_url": "https://avatars.githubusercontent.com/u/73724178?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bwest2397", "html_url": "https://github.com/bwest2397", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
11
2023-11-28T17:06:07
2023-12-12T19:34:56
2023-11-28T21:25:23
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1306", "html_url": "https://github.com/ollama/ollama/pull/1306", "diff_url": "https://github.com/ollama/ollama/pull/1306.diff", "patch_url": "https://github.com/ollama/ollama/pull/1306.patch", "merged_at": null }
As it currently stands, the Docker image that gets built is seemingly unable to use the GPU despite the initial "build" stage of the Dockerfile ostensibly being built with CUDA support (i.e. built using the `nvidia/cuda` base image). As reported in https://github.com/jmorganca/ollama/issues/797, it seems that simply se...
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1306/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1306/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5667
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5667/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5667/comments
https://api.github.com/repos/ollama/ollama/issues/5667/events
https://github.com/ollama/ollama/pull/5667
2,406,750,057
PR_kwDOJ0Z1Ps51SJUt
5,667
Enhance list command
{ "login": "kaichen", "id": 4048, "node_id": "MDQ6VXNlcjQwNDg=", "avatar_url": "https://avatars.githubusercontent.com/u/4048?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kaichen", "html_url": "https://github.com/kaichen", "followers_url": "https://api.github.com/users/kaichen/followers"...
[]
closed
false
null
[]
null
2
2024-07-13T07:21:13
2024-11-22T00:49:46
2024-11-22T00:49:46
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5667", "html_url": "https://github.com/ollama/ollama/pull/5667", "diff_url": "https://github.com/ollama/ollama/pull/5667.diff", "patch_url": "https://github.com/ollama/ollama/pull/5667.patch", "merged_at": null }
changes - add missing filter description for list command - add sort option to list command
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5667/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5667/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/24
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/24/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/24/comments
https://api.github.com/repos/ollama/ollama/issues/24/events
https://github.com/ollama/ollama/pull/24
1,781,626,501
PR_kwDOJ0Z1Ps5USMRS
24
remove add cmd
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
0
2023-06-29T23:20:02
2023-06-29T23:29:30
2023-06-29T23:29:23
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/24", "html_url": "https://github.com/ollama/ollama/pull/24", "diff_url": "https://github.com/ollama/ollama/pull/24.diff", "patch_url": "https://github.com/ollama/ollama/pull/24.patch", "merged_at": "2023-06-29T23:29:23" }
🧹
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/24/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/24/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2417
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2417/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2417/comments
https://api.github.com/repos/ollama/ollama/issues/2417/events
https://github.com/ollama/ollama/issues/2417
2,126,159,957
I_kwDOJ0Z1Ps5-uqBV
2,417
Tools support in ChatCompletion endpoints
{ "login": "pamelafox", "id": 297042, "node_id": "MDQ6VXNlcjI5NzA0Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/297042?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pamelafox", "html_url": "https://github.com/pamelafox", "followers_url": "https://api.github.com/users/pame...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 6657611864, "node_id": ...
closed
false
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers"...
[ { "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/...
null
10
2024-02-08T22:17:27
2024-06-25T04:30:24
2024-06-25T04:30:24
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
We'd love tools support so we can use ollama with our existing OpenAI-using apps. Not sure if that's possible across the board with all models.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2417/reactions", "total_count": 40, "+1": 40, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2417/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3265
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3265/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3265/comments
https://api.github.com/repos/ollama/ollama/issues/3265/events
https://github.com/ollama/ollama/issues/3265
2,197,028,474
I_kwDOJ0Z1Ps6C8_56
3,265
Does ollama also plan to support the sound models?
{ "login": "insooneelife", "id": 8437769, "node_id": "MDQ6VXNlcjg0Mzc3Njk=", "avatar_url": "https://avatars.githubusercontent.com/u/8437769?v=4", "gravatar_id": "", "url": "https://api.github.com/users/insooneelife", "html_url": "https://github.com/insooneelife", "followers_url": "https://api.github.com...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
2
2024-03-20T09:18:59
2024-12-25T15:57:35
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What are you trying to do? Currently, ollama supports most llms and I know that it also supports vision model. I thought it would be nice if the sound model could also be used through ollama. I wonder if there are any plans like this. ### How should we solve this? I know that there are currently sound models rel...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3265/reactions", "total_count": 7, "+1": 7, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3265/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6098
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6098/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6098/comments
https://api.github.com/repos/ollama/ollama/issues/6098/events
https://github.com/ollama/ollama/issues/6098
2,440,088,309
I_kwDOJ0Z1Ps6RcMr1
6,098
Why is the llama3 model missing after I restart Ollama? When I run “ollama run llama3”, it re-pulls the manifest.
{ "login": "fanjikang", "id": 146172937, "node_id": "U_kgDOCLZsCQ", "avatar_url": "https://avatars.githubusercontent.com/u/146172937?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fanjikang", "html_url": "https://github.com/fanjikang", "followers_url": "https://api.github.com/users/fanjik...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
3
2024-07-31T13:38:24
2024-08-27T10:25:25
2024-08-27T10:25:25
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Why is the llama3 model missing after I restart Ollama? When I run “ollama run llama3”, it re-pulls the manifest. ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version v0.2.8
{ "login": "fanjikang", "id": 146172937, "node_id": "U_kgDOCLZsCQ", "avatar_url": "https://avatars.githubusercontent.com/u/146172937?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fanjikang", "html_url": "https://github.com/fanjikang", "followers_url": "https://api.github.com/users/fanjik...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6098/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6098/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7150
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7150/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7150/comments
https://api.github.com/repos/ollama/ollama/issues/7150/events
https://github.com/ollama/ollama/issues/7150
2,576,056,905
I_kwDOJ0Z1Ps6Zi4JJ
7,150
CORS (Cross-Origin Resource Sharing)
{ "login": "Apex-Predator-786", "id": 94287474, "node_id": "U_kgDOBZ62cg", "avatar_url": "https://avatars.githubusercontent.com/u/94287474?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Apex-Predator-786", "html_url": "https://github.com/Apex-Predator-786", "followers_url": "https://api.g...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 7706482389, "node_id": "LA_kwDOJ0Z1Ps8AAAABy1eW1Q...
open
false
null
[]
null
0
2024-10-09T14:08:10
2024-11-06T00:01:14
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? please enable CORS (Cross-Origin Resource Sharing) in rest api
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7150/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7150/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2771
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2771/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2771/comments
https://api.github.com/repos/ollama/ollama/issues/2771/events
https://github.com/ollama/ollama/pull/2771
2,155,434,878
PR_kwDOJ0Z1Ps5n-7UH
2,771
Bump llama.cpp to b2276
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
0
2024-02-27T00:51:50
2024-02-27T19:29:56
2024-02-27T19:29:53
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2771", "html_url": "https://github.com/ollama/ollama/pull/2771", "diff_url": "https://github.com/ollama/ollama/pull/2771.diff", "patch_url": "https://github.com/ollama/ollama/pull/2771.patch", "merged_at": "2024-02-27T19:29:53" }
Fixes #2758
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2771/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2771/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3271
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3271/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3271/comments
https://api.github.com/repos/ollama/ollama/issues/3271/events
https://github.com/ollama/ollama/issues/3271
2,197,650,730
I_kwDOJ0Z1Ps6C_X0q
3,271
Cloudflare Tunnels Not Work After Update 1.2.9
{ "login": "projetosfsi", "id": 117665999, "node_id": "U_kgDOBwNwzw", "avatar_url": "https://avatars.githubusercontent.com/u/117665999?v=4", "gravatar_id": "", "url": "https://api.github.com/users/projetosfsi", "html_url": "https://github.com/projetosfsi", "followers_url": "https://api.github.com/users/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
9
2024-03-20T14:15:13
2024-11-07T18:46:03
2024-03-20T16:49:15
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I'm using Cloudfalre Tunnels to connect Ollama to a web domain. Everything was working fine until the update to version 1.2.9 I started getting the 403 Access Denied error. Regardless of how I run it, with administrative access or as a regular user, nothing changes. After downgrading...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3271/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3271/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1730
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1730/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1730/comments
https://api.github.com/repos/ollama/ollama/issues/1730/events
https://github.com/ollama/ollama/issues/1730
2,057,739,752
I_kwDOJ0Z1Ps56pp3o
1,730
MLX backend
{ "login": "ageorgios", "id": 3126351, "node_id": "MDQ6VXNlcjMxMjYzNTE=", "avatar_url": "https://avatars.githubusercontent.com/u/3126351?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ageorgios", "html_url": "https://github.com/ageorgios", "followers_url": "https://api.github.com/users/ag...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
57
2023-12-27T20:10:50
2025-01-24T14:12:55
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Can ollama be converted to use MLX from Apple as backend for the models ?
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1730/reactions", "total_count": 204, "+1": 164, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 40, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1730/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/4407
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4407/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4407/comments
https://api.github.com/repos/ollama/ollama/issues/4407/events
https://github.com/ollama/ollama/issues/4407
2,293,538,684
I_kwDOJ0Z1Ps6ItJ98
4,407
Discrepancy between FAQ and program
{ "login": "FrederikSchack", "id": 47929714, "node_id": "MDQ6VXNlcjQ3OTI5NzE0", "avatar_url": "https://avatars.githubusercontent.com/u/47929714?v=4", "gravatar_id": "", "url": "https://api.github.com/users/FrederikSchack", "html_url": "https://github.com/FrederikSchack", "followers_url": "https://api.gi...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-05-13T18:34:20
2024-05-14T01:02:27
2024-05-14T01:02:27
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? According to your FAQ, there should be a control panel in the Windows version for setting the environment variables: https://github.com/ollama/ollama/blob/main/docs/faq.md There isn't any control panel when you start the program, no apparent control panel or configuration file in the Ollama ...
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4407/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4407/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/852
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/852/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/852/comments
https://api.github.com/repos/ollama/ollama/issues/852/events
https://github.com/ollama/ollama/issues/852
1,954,491,700
I_kwDOJ0Z1Ps50fy00
852
Compress with xz for macOS and Linux GitHub Releases
{ "login": "coolaj86", "id": 122831, "node_id": "MDQ6VXNlcjEyMjgzMQ==", "avatar_url": "https://avatars.githubusercontent.com/u/122831?v=4", "gravatar_id": "", "url": "https://api.github.com/users/coolaj86", "html_url": "https://github.com/coolaj86", "followers_url": "https://api.github.com/users/coolaj8...
[]
closed
false
null
[]
null
3
2023-10-20T15:08:55
2023-10-26T05:14:35
2023-10-26T00:49:39
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I am willing to help with this. ## Why `xz`? `xz` is a standard compression tool for binary files (just like `gzip` is a standard tool for compressing text files). It comes preinstalled on macOS and _most_ Linuxes (at least the ones used for desktop and server use, though not always in containers). Double c...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/852/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/852/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/572
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/572/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/572/comments
https://api.github.com/repos/ollama/ollama/issues/572/events
https://github.com/ollama/ollama/issues/572
1,908,543,284
I_kwDOJ0Z1Ps5xwg80
572
How to prevent model answer runaway?
{ "login": "FairyTail2000", "id": 22645621, "node_id": "MDQ6VXNlcjIyNjQ1NjIx", "avatar_url": "https://avatars.githubusercontent.com/u/22645621?v=4", "gravatar_id": "", "url": "https://api.github.com/users/FairyTail2000", "html_url": "https://github.com/FairyTail2000", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
9
2023-09-22T09:30:55
2023-11-09T21:16:28
2023-11-09T21:16:27
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I've encountered the following issue with some models: The model just keeps answering, providing question in the training format and answering them. This just keeps on going or it stops after one or two extra answers. Example: medllama2:7b-q5_0 docker gpu Question from me (fictional): > A male 23 year old pers...
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/572/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/572/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2658
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2658/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2658/comments
https://api.github.com/repos/ollama/ollama/issues/2658/events
https://github.com/ollama/ollama/issues/2658
2,147,927,324
I_kwDOJ0Z1Ps6ABsUc
2,658
Ollama should clear temp files on exit.
{ "login": "amnweb", "id": 16545063, "node_id": "MDQ6VXNlcjE2NTQ1MDYz", "avatar_url": "https://avatars.githubusercontent.com/u/16545063?v=4", "gravatar_id": "", "url": "https://api.github.com/users/amnweb", "html_url": "https://github.com/amnweb", "followers_url": "https://api.github.com/users/amnweb/fo...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
2
2024-02-22T00:12:32
2024-03-20T15:28:06
2024-03-20T15:28:06
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Found that upon exiting, Ollama does not delete temporary files, but upon starting, Ollama creates new identical files again. in temp folder ''..AppData\Local\Temp" OS: Windows 11 ![image](https://github.com/ollama/ollama/assets/16545063/8831fff1-d684-4217-bc39-a6aaac5624e9)
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2658/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2658/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8304
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8304/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8304/comments
https://api.github.com/repos/ollama/ollama/issues/8304/events
https://github.com/ollama/ollama/issues/8304
2,768,646,500
I_kwDOJ0Z1Ps6lBjFk
8,304
Unable to bind 0.0.0.0 port to all interfaces, except localhost
{ "login": "drigio", "id": 22986282, "node_id": "MDQ6VXNlcjIyOTg2Mjgy", "avatar_url": "https://avatars.githubusercontent.com/u/22986282?v=4", "gravatar_id": "", "url": "https://api.github.com/users/drigio", "html_url": "https://github.com/drigio", "followers_url": "https://api.github.com/users/drigio/fo...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
6
2025-01-04T08:36:06
2025-01-24T09:46:59
2025-01-24T09:46:59
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I'm trying to `ollama serve` by setting env variable `OLLAMA_HOST=0.0.0.0` on windows. However I am only able to access it via localhost:11434. Unable to access it via my host ip, nor via 127.0.0.1. There is no proxy enabled, no vpn. I have disabled firewall in my windows machine. ollam...
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8304/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8304/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7425
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7425/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7425/comments
https://api.github.com/repos/ollama/ollama/issues/7425/events
https://github.com/ollama/ollama/issues/7425
2,624,721,955
I_kwDOJ0Z1Ps6cchQj
7,425
Model loaded to RAM and not VRAM
{ "login": "johanesalxd", "id": 17249308, "node_id": "MDQ6VXNlcjE3MjQ5MzA4", "avatar_url": "https://avatars.githubusercontent.com/u/17249308?v=4", "gravatar_id": "", "url": "https://api.github.com/users/johanesalxd", "html_url": "https://github.com/johanesalxd", "followers_url": "https://api.github.com/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg...
closed
false
null
[]
null
1
2024-10-30T16:52:20
2024-10-30T16:55:15
2024-10-30T16:54:58
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? ![image](https://github.com/user-attachments/assets/0b0cad75-7e33-4b6b-8f76-8dc5f9f95b7c) [server.log](https://github.com/user-attachments/files/17576314/server.log) [lmstudio-issues](https://discord.com/channels/1110598183144399058/1294935199892639834) [complete-specs](https://pcpartpicker...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7425/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7425/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7136
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7136/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7136/comments
https://api.github.com/repos/ollama/ollama/issues/7136/events
https://github.com/ollama/ollama/issues/7136
2,573,570,199
I_kwDOJ0Z1Ps6ZZZCX
7,136
Tool binding prevent streaming
{ "login": "pyautoml", "id": 147545771, "node_id": "U_kgDOCMteqw", "avatar_url": "https://avatars.githubusercontent.com/u/147545771?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pyautoml", "html_url": "https://github.com/pyautoml", "followers_url": "https://api.github.com/users/pyautoml/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
3
2024-10-08T15:39:32
2024-10-16T01:41:11
2024-10-16T01:41:11
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Hi there. I discovered recently that onces a model is delivered tool via bind_tools is stops recognizing streeaming, and each time it waits for the data to be fully loaded. This snippet below works perfectly fine until tools are added: print("AI:", end=" ", flush=True) for chunk in self....
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7136/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/ollama/ollama/issues/7136/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8118
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8118/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8118/comments
https://api.github.com/repos/ollama/ollama/issues/8118/events
https://github.com/ollama/ollama/pull/8118
2,742,578,632
PR_kwDOJ0Z1Ps6FXfka
8,118
Change the update policy for OLLAMA_ORIGINS to Override
{ "login": "yaoxi-std", "id": 57950207, "node_id": "MDQ6VXNlcjU3OTUwMjA3", "avatar_url": "https://avatars.githubusercontent.com/u/57950207?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yaoxi-std", "html_url": "https://github.com/yaoxi-std", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
1
2024-12-16T14:43:16
2024-12-23T15:41:52
2024-12-23T15:41:51
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8118", "html_url": "https://github.com/ollama/ollama/pull/8118", "diff_url": "https://github.com/ollama/ollama/pull/8118.diff", "patch_url": "https://github.com/ollama/ollama/pull/8118.patch", "merged_at": null }
So that I will be able to remove some of the default origins.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8118/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8118/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7557
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7557/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7557/comments
https://api.github.com/repos/ollama/ollama/issues/7557/events
https://github.com/ollama/ollama/issues/7557
2,641,000,580
I_kwDOJ0Z1Ps6daniE
7,557
我的文件明明只有12GB,为什么加载的时候,却显示需要22GB显存?
{ "login": "fg2501", "id": 164639270, "node_id": "U_kgDOCdAyJg", "avatar_url": "https://avatars.githubusercontent.com/u/164639270?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fg2501", "html_url": "https://github.com/fg2501", "followers_url": "https://api.github.com/users/fg2501/follower...
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
null
[]
null
4
2024-11-07T13:13:24
2024-11-08T03:21:17
2024-11-07T21:50:01
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? ![3f5ba904-d558-4770-b3e3-ee7783694b2f](https://github.com/user-attachments/assets/1a2ef30a-766c-4f00-873c-0b3468238754) ![1111](https://github.com/user-attachments/assets/03dba386-b3cb-4267-9d02-07b048c3a503) 我用的qwen2.5:14b模型,为什么我加载的时候,显示却是需要22GB显存呢? ### OS Windows ### GPU Nvidia ### CPU...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7557/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7557/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4825
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4825/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4825/comments
https://api.github.com/repos/ollama/ollama/issues/4825/events
https://github.com/ollama/ollama/issues/4825
2,334,841,826
I_kwDOJ0Z1Ps6LKtvi
4,825
Adapter doesnt work
{ "login": "namogg", "id": 108812102, "node_id": "U_kgDOBnxXRg", "avatar_url": "https://avatars.githubusercontent.com/u/108812102?v=4", "gravatar_id": "", "url": "https://api.github.com/users/namogg", "html_url": "https://github.com/namogg", "followers_url": "https://api.github.com/users/namogg/follower...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
[ { "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/us...
null
6
2024-06-05T03:54:33
2024-07-10T19:38:32
2024-07-10T18:30:41
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I have converted my adapter to GGML format and create ollama model sucessfully with this comman ADAPTER downloads/ggml-adapter-model.bin However when testing the model, the adapter have no affect on the model. ### OS Linux ### GPU Nvidia ### CPU Intel, AMD ### Ollama version 0.1.29
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4825/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4825/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1641
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1641/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1641/comments
https://api.github.com/repos/ollama/ollama/issues/1641/events
https://github.com/ollama/ollama/issues/1641
2,051,188,544
I_kwDOJ0Z1Ps56QqdA
1,641
Ollama hangs sometimes if it runs out of VRAM
{ "login": "nick-tonjum", "id": 81722223, "node_id": "MDQ6VXNlcjgxNzIyMjIz", "avatar_url": "https://avatars.githubusercontent.com/u/81722223?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nick-tonjum", "html_url": "https://github.com/nick-tonjum", "followers_url": "https://api.github.com/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
3
2023-12-20T20:04:18
2024-01-21T01:54:47
2024-01-21T01:54:47
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi! I just have been having an issue with models that cause the system to run out of VRAM. It usually does the following: 1. (attempt to run a model via api, for example Llama2 70b) 1. ollama-runner tries to load the model into VRAM 2. ollama-runner runs out of VRAM and the process kills 3. the API hangs indefini...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1641/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1641/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5384
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5384/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5384/comments
https://api.github.com/repos/ollama/ollama/issues/5384/events
https://github.com/ollama/ollama/issues/5384
2,381,814,914
I_kwDOJ0Z1Ps6N95yC
5,384
dolphin-phi3 and dolphin-qwen2
{ "login": "olumolu", "id": 162728301, "node_id": "U_kgDOCbMJbQ", "avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4", "gravatar_id": "", "url": "https://api.github.com/users/olumolu", "html_url": "https://github.com/olumolu", "followers_url": "https://api.github.com/users/olumolu/foll...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
4
2024-06-29T15:25:22
2024-07-03T17:42:14
2024-07-03T17:42:13
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
dolphin-phi3 and qwen2 dolphin https://huggingface.co/cognitivecomputations can we have this models so people can us them like dolphin-llama3
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5384/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5384/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/44
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/44/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/44/comments
https://api.github.com/repos/ollama/ollama/issues/44/events
https://github.com/ollama/ollama/pull/44
1,792,041,442
PR_kwDOJ0Z1Ps5U1i1u
44
add ollama system tray
{ "login": "hoyyeva", "id": 63033505, "node_id": "MDQ6VXNlcjYzMDMzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoyyeva", "html_url": "https://github.com/hoyyeva", "followers_url": "https://api.github.com/users/hoyyev...
[]
closed
false
null
[]
null
1
2023-07-06T18:33:01
2023-07-07T17:51:57
2023-07-06T19:22:55
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/44", "html_url": "https://github.com/ollama/ollama/pull/44", "diff_url": "https://github.com/ollama/ollama/pull/44.diff", "patch_url": "https://github.com/ollama/ollama/pull/44.patch", "merged_at": null }
null
{ "login": "hoyyeva", "id": 63033505, "node_id": "MDQ6VXNlcjYzMDMzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoyyeva", "html_url": "https://github.com/hoyyeva", "followers_url": "https://api.github.com/users/hoyyev...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/44/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/44/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5741
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5741/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5741/comments
https://api.github.com/repos/ollama/ollama/issues/5741/events
https://github.com/ollama/ollama/issues/5741
2,412,692,108
I_kwDOJ0Z1Ps6PzsKM
5,741
ROCm Memory Issues with Long Contexts
{ "login": "ProjectMoon", "id": 183856, "node_id": "MDQ6VXNlcjE4Mzg1Ng==", "avatar_url": "https://avatars.githubusercontent.com/u/183856?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ProjectMoon", "html_url": "https://github.com/ProjectMoon", "followers_url": "https://api.github.com/user...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
2
2024-07-17T05:37:09
2024-12-19T11:53:09
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Similar (?) to #1952. I've been noticing that ollama will crash when using long context lengths on ROCm. In particular, the most noticeable thing is that I can continue large conversations with ollama from start, while the model remains loaded in memory. But after coming back later and the model...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5741/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5741/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/8354
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8354/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8354/comments
https://api.github.com/repos/ollama/ollama/issues/8354/events
https://github.com/ollama/ollama/issues/8354
2,776,575,597
I_kwDOJ0Z1Ps6lfy5t
8,354
Dynamic context size in OpenAI API compatibility.
{ "login": "x0wllaar", "id": 10964379, "node_id": "MDQ6VXNlcjEwOTY0Mzc5", "avatar_url": "https://avatars.githubusercontent.com/u/10964379?v=4", "gravatar_id": "", "url": "https://api.github.com/users/x0wllaar", "html_url": "https://github.com/x0wllaar", "followers_url": "https://api.github.com/users/x0w...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2025-01-09T00:08:41
2025-01-16T00:01:18
2025-01-16T00:01:18
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I noticed that the issue (#5356 ) regarding dynamically setting the context size (num_ctx) in OpenAI API was closed with a note saying it wasn't possible due to limitations of the API. However, I'd like to reopen this discussion as there seems to be a using the extra_body parameter available in the OpenAI API clients. ...
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8354/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8354/timeline
null
duplicate
false
https://api.github.com/repos/ollama/ollama/issues/4529
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4529/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4529/comments
https://api.github.com/repos/ollama/ollama/issues/4529/events
https://github.com/ollama/ollama/issues/4529
2,304,933,652
I_kwDOJ0Z1Ps6JYn8U
4,529
error loading model: error loading model vocabulary: unknown pre-tokenizer type: 'qwen2'
{ "login": "Anorid", "id": 139095718, "node_id": "U_kgDOCEpupg", "avatar_url": "https://avatars.githubusercontent.com/u/139095718?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Anorid", "html_url": "https://github.com/Anorid", "followers_url": "https://api.github.com/users/Anorid/follower...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
22
2024-05-20T02:14:35
2024-07-15T09:32:18
2024-06-24T16:48:15
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I carefully read the contents of the readme's documentation to try and found that something went wrong time=2024-05-20T10:06:02.688+08:00 level=INFO source=server.go:320 msg="starting llama server" cmd="/tmp/ollama2132883000/runners/cuda_v11/ollama_llama_server --model /root/autodl-tmp/models...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4529/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4529/timeline
null
completed
false