url stringlengths 51 54 | repository_url stringclasses 1
value | labels_url stringlengths 65 68 | comments_url stringlengths 60 63 | events_url stringlengths 58 61 | html_url stringlengths 39 44 | id int64 1.78B 2.82B | node_id stringlengths 18 19 | number int64 1 8.69k | title stringlengths 1 382 | user dict | labels listlengths 0 5 | state stringclasses 2
values | locked bool 1
class | assignee dict | assignees listlengths 0 2 | milestone null | comments int64 0 323 | created_at timestamp[s] | updated_at timestamp[s] | closed_at timestamp[s] | author_association stringclasses 4
values | sub_issues_summary dict | active_lock_reason null | draft bool 2
classes | pull_request dict | body stringlengths 2 118k ⌀ | closed_by dict | reactions dict | timeline_url stringlengths 60 63 | performed_via_github_app null | state_reason stringclasses 4
values | is_pull_request bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/ollama/ollama/issues/5614 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5614/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5614/comments | https://api.github.com/repos/ollama/ollama/issues/5614/events | https://github.com/ollama/ollama/pull/5614 | 2,401,766,631 | PR_kwDOJ0Z1Ps51BVsT | 5,614 | OpenAI: Support Tools | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjha... | [] | closed | false | null | [] | null | 1 | 2024-07-10T21:21:37 | 2024-07-23T23:02:27 | 2024-07-17T03:52:59 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5614",
"html_url": "https://github.com/ollama/ollama/pull/5614",
"diff_url": "https://github.com/ollama/ollama/pull/5614.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5614.patch",
"merged_at": "2024-07-17T03:52:59"
} | ```
curl -s 127.0.0.1:11434/v1/chat/completions -d '{
"model": "mike/mistral",
"messages": [
{
"role": "user",
"content": "What'\''s the weather like today in Paris?"
},
{
"role": "assistant",
"tool_calls": [
{
"id": "89a1e453-0bce-4de3-a456-c54bed09... | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjha... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5614/reactions",
"total_count": 8,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 8,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5614/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/700 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/700/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/700/comments | https://api.github.com/repos/ollama/ollama/issues/700/events | https://github.com/ollama/ollama/pull/700 | 1,926,883,924 | PR_kwDOJ0Z1Ps5b77h7 | 700 | rename server subprocess | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 0 | 2023-10-04T19:53:55 | 2023-10-06T14:15:43 | 2023-10-06T14:15:42 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/700",
"html_url": "https://github.com/ollama/ollama/pull/700",
"diff_url": "https://github.com/ollama/ollama/pull/700.diff",
"patch_url": "https://github.com/ollama/ollama/pull/700.patch",
"merged_at": "2023-10-06T14:15:42"
} | rename llama.cpp `server.exe` to `ollama-runner`. This makes it easier to see that the subprocess is associated with ollama.
| {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/700/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/700/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8647 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8647/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8647/comments | https://api.github.com/repos/ollama/ollama/issues/8647/events | https://github.com/ollama/ollama/issues/8647 | 2,817,099,829 | I_kwDOJ0Z1Ps6n6Yg1 | 8,647 | ollama[2075]: cuda driver library failed to get device context 999time=2025-01-29T03:21:43.859+01:00 level=WARN source=gpu.go:449 msg="error looking up nvidia GPU memory" | {
"login": "kurdo2000",
"id": 137941417,
"node_id": "U_kgDOCDjRqQ",
"avatar_url": "https://avatars.githubusercontent.com/u/137941417?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kurdo2000",
"html_url": "https://github.com/kurdo2000",
"followers_url": "https://api.github.com/users/kurdo2... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg... | closed | false | null | [] | null | 2 | 2025-01-29T02:34:23 | 2025-01-30T03:09:02 | 2025-01-30T03:09:01 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Hi,
is my GPU not supported or what's the error?
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 550.120 Driver Version: 550.120 CUDA Version: 12.4 |
|-----------------------------------------+-----------------... | {
"login": "kurdo2000",
"id": 137941417,
"node_id": "U_kgDOCDjRqQ",
"avatar_url": "https://avatars.githubusercontent.com/u/137941417?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kurdo2000",
"html_url": "https://github.com/kurdo2000",
"followers_url": "https://api.github.com/users/kurdo2... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8647/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8647/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8067 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8067/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8067/comments | https://api.github.com/repos/ollama/ollama/issues/8067/events | https://github.com/ollama/ollama/issues/8067 | 2,735,177,482 | I_kwDOJ0Z1Ps6jB38K | 8,067 | have a ques | {
"login": "17692019663",
"id": 57002321,
"node_id": "MDQ6VXNlcjU3MDAyMzIx",
"avatar_url": "https://avatars.githubusercontent.com/u/57002321?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/17692019663",
"html_url": "https://github.com/17692019663",
"followers_url": "https://api.github.com/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg... | closed | false | null | [] | null | 2 | 2024-12-12T08:17:34 | 2024-12-23T08:12:10 | 2024-12-23T08:12:10 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
C:\Users\Administrator> docker exec -it ollama ollama run llama3.2:1b
pulling manifest
pulling 74701a8c35f6... 100% ▕████████████████████████████████████████████████████████▏ 1.3 GB
pulling 966de95ca8a6... 100% ▕████████████████████████████████████████████████████████▏ 1.4 KB
pulling fcc5a6b... | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8067/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8067/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/134 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/134/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/134/comments | https://api.github.com/repos/ollama/ollama/issues/134/events | https://github.com/ollama/ollama/pull/134 | 1,813,267,553 | PR_kwDOJ0Z1Ps5V9on7 | 134 | Update images.go | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-07-20T06:34:19 | 2023-07-20T06:46:19 | 2023-07-20T06:46:02 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/134",
"html_url": "https://github.com/ollama/ollama/pull/134",
"diff_url": "https://github.com/ollama/ollama/pull/134.diff",
"patch_url": "https://github.com/ollama/ollama/pull/134.patch",
"merged_at": "2023-07-20T06:46:02"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/134/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/134/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3226 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3226/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3226/comments | https://api.github.com/repos/ollama/ollama/issues/3226/events | https://github.com/ollama/ollama/issues/3226 | 2,192,612,235 | I_kwDOJ0Z1Ps6CsJuL | 3,226 | Cannot create Modelfile on Windows Shell | {
"login": "worras2019",
"id": 52268635,
"node_id": "MDQ6VXNlcjUyMjY4NjM1",
"avatar_url": "https://avatars.githubusercontent.com/u/52268635?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/worras2019",
"html_url": "https://github.com/worras2019",
"followers_url": "https://api.github.com/use... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2024-03-18T15:57:03 | 2024-03-19T08:50:15 | 2024-03-19T08:50:15 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
The following is not working in WIndows Shell. Windows Shell doesn't recognize FROM.
--------------------------------------------------
FROM llama2
# set the temperature to 1 [higher is more creative, lower is more coherent]
PARAMETER temperature 1
# set the system message
SYSTEM """
... | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3226/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3226/timeline | null | not_planned | false |
https://api.github.com/repos/ollama/ollama/issues/2788 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2788/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2788/comments | https://api.github.com/repos/ollama/ollama/issues/2788/events | https://github.com/ollama/ollama/issues/2788 | 2,157,663,506 | I_kwDOJ0Z1Ps6Am1US | 2,788 | Bug: LLaVA 1.6 34b not respecting initial user prompt | {
"login": "mobilemike",
"id": 30590,
"node_id": "MDQ6VXNlcjMwNTkw",
"avatar_url": "https://avatars.githubusercontent.com/u/30590?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mobilemike",
"html_url": "https://github.com/mobilemike",
"followers_url": "https://api.github.com/users/mobilem... | [] | closed | false | null | [] | null | 0 | 2024-02-27T21:47:53 | 2024-02-29T19:30:15 | 2024-02-29T19:30:15 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | M2 Max MBP 96GB RAM
Ollama 0.1.27
Compared against llama.cpp CLI @b11a93d (same as Ollama version)
Problem:
When using the Ollama CLI or API with an image, the initial text prompt isn't respected. Examples like the one on the [model page](https://ollama.com/library/llava) typically show prompts like "What is in t... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2788/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2788/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4205 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4205/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4205/comments | https://api.github.com/repos/ollama/ollama/issues/4205/events | https://github.com/ollama/ollama/issues/4205 | 2,281,206,136 | I_kwDOJ0Z1Ps6H-HF4 | 4,205 | please support deepseek-ai/DeepSeek-V2-Chat | {
"login": "taozhiyuai",
"id": 146583103,
"node_id": "U_kgDOCLyuPw",
"avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/taozhiyuai",
"html_url": "https://github.com/taozhiyuai",
"followers_url": "https://api.github.com/users/tao... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 1 | 2024-05-06T15:54:12 | 2024-05-08T06:12:39 | 2024-05-08T06:12:39 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null |
please support deepseek-ai/DeepSeek-V2-Chat
https://hf-mirror.com/deepseek-ai/DeepSeek-V2-Chat/tree/main | {
"login": "taozhiyuai",
"id": 146583103,
"node_id": "U_kgDOCLyuPw",
"avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/taozhiyuai",
"html_url": "https://github.com/taozhiyuai",
"followers_url": "https://api.github.com/users/tao... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4205/reactions",
"total_count": 3,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 3,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4205/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2058 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2058/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2058/comments | https://api.github.com/repos/ollama/ollama/issues/2058/events | https://github.com/ollama/ollama/issues/2058 | 2,089,071,525 | I_kwDOJ0Z1Ps58hLOl | 2,058 | Prompt Eval Count is 1 when image is included in multimodal request | {
"login": "Dillon-Yun",
"id": 40805060,
"node_id": "MDQ6VXNlcjQwODA1MDYw",
"avatar_url": "https://avatars.githubusercontent.com/u/40805060?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Dillon-Yun",
"html_url": "https://github.com/Dillon-Yun",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 1 | 2024-01-18T21:24:42 | 2024-10-23T17:46:39 | 2024-10-23T17:46:38 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | When I hit the `/api/generate` endpoint for a multimodal model, such as bakllava, I am seeing that `prompt_eval_count == 1` when an image is included. Is this correct behavior? How can I retrieve the total amount of tokens in the request for a multimodal request? | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2058/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2058/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5880 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5880/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5880/comments | https://api.github.com/repos/ollama/ollama/issues/5880/events | https://github.com/ollama/ollama/issues/5880 | 2,425,575,429 | I_kwDOJ0Z1Ps6Qk1gF | 5,880 | Make Ollama available on Chocolatey | {
"login": "b-a0",
"id": 25707742,
"node_id": "MDQ6VXNlcjI1NzA3NzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/25707742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/b-a0",
"html_url": "https://github.com/b-a0",
"followers_url": "https://api.github.com/users/b-a0/followers"... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 5860134234,
"node_id": ... | open | false | null | [] | null | 2 | 2024-07-23T16:04:09 | 2024-09-05T19:52:37 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | The Windows installers works great, but requires that I manually download/update it from the ollama website. Would it be possible, probably when the Windows version has matured some more, to add ollama to [Chocolatey](https://community.chocolatey.org/)? That will probabably aid in making this an (even) easier to instal... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5880/reactions",
"total_count": 4,
"+1": 2,
"-1": 2,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5880/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/2377 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2377/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2377/comments | https://api.github.com/repos/ollama/ollama/issues/2377/events | https://github.com/ollama/ollama/pull/2377 | 2,121,583,534 | PR_kwDOJ0Z1Ps5mL7sV | 2,377 | Bump llama.cpp to b2081 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 1 | 2024-02-06T20:09:01 | 2024-02-07T20:04:41 | 2024-02-07T20:04:38 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2377",
"html_url": "https://github.com/ollama/ollama/pull/2377",
"diff_url": "https://github.com/ollama/ollama/pull/2377.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2377.patch",
"merged_at": "2024-02-07T20:04:38"
} | null | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2377/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2377/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3513 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3513/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3513/comments | https://api.github.com/repos/ollama/ollama/issues/3513/events | https://github.com/ollama/ollama/issues/3513 | 2,229,191,000 | I_kwDOJ0Z1Ps6E3sFY | 3,513 | Stream: false is ignored in as of c863c6a96d01afda37b65f86eb14a1f04a3c7c47 | {
"login": "eusebiu",
"id": 414589,
"node_id": "MDQ6VXNlcjQxNDU4OQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/414589?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/eusebiu",
"html_url": "https://github.com/eusebiu",
"followers_url": "https://api.github.com/users/eusebiu/fo... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 4 | 2024-04-06T10:57:21 | 2024-04-13T07:22:27 | 2024-04-13T07:22:26 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Stream: false is ignored
### What did you expect to see?
When stream is set to false, the response should contain the full result.
### Steps to reproduce
Pull the latest code, compile, run (pull some model), run /api/generate with stream: false -> the result has no content (just so... | {
"login": "eusebiu",
"id": 414589,
"node_id": "MDQ6VXNlcjQxNDU4OQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/414589?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/eusebiu",
"html_url": "https://github.com/eusebiu",
"followers_url": "https://api.github.com/users/eusebiu/fo... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3513/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3513/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8117 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8117/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8117/comments | https://api.github.com/repos/ollama/ollama/issues/8117/events | https://github.com/ollama/ollama/pull/8117 | 2,742,219,356 | PR_kwDOJ0Z1Ps6FWP5c | 8,117 | Examples(privategpt): update outdated langchain dependencies and add UnstructuredRSTLoader | {
"login": "KaymeKaydex",
"id": 43151027,
"node_id": "MDQ6VXNlcjQzMTUxMDI3",
"avatar_url": "https://avatars.githubusercontent.com/u/43151027?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/KaymeKaydex",
"html_url": "https://github.com/KaymeKaydex",
"followers_url": "https://api.github.com/... | [] | closed | false | {
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/... | [
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "htt... | null | 3 | 2024-12-16T12:12:05 | 2025-01-13T19:26:24 | 2025-01-13T19:26:24 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8117",
"html_url": "https://github.com/ollama/ollama/pull/8117",
"diff_url": "https://github.com/ollama/ollama/pull/8117.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8117.patch",
"merged_at": null
} | This commit updates the langchain dependencies to the latest stable versions and adds the UnstructuredRSTLoader for loading RST files.
- Update langchain dependencies to the latest stable versions.
- Add UnstructuredRSTLoader for loading RST files. | {
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8117/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8117/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3074 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3074/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3074/comments | https://api.github.com/repos/ollama/ollama/issues/3074/events | https://github.com/ollama/ollama/issues/3074 | 2,181,146,552 | I_kwDOJ0Z1Ps6CAae4 | 3,074 | CUDA error: an illegal memory access was encountered | {
"login": "lizhichao999",
"id": 34128722,
"node_id": "MDQ6VXNlcjM0MTI4NzIy",
"avatar_url": "https://avatars.githubusercontent.com/u/34128722?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lizhichao999",
"html_url": "https://github.com/lizhichao999",
"followers_url": "https://api.github.c... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg... | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 6 | 2024-03-12T09:35:45 | 2024-07-24T23:14:51 | 2024-07-24T23:14:51 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ollama run llama2
When I executed the command: ollama run llama2, an error occurred related to the data information.

I only have a 128G GPU. Do I need more than 2 GPUs? I can run it on my local laptop with 2 GPU... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3074/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3074/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7704 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7704/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7704/comments | https://api.github.com/repos/ollama/ollama/issues/7704/events | https://github.com/ollama/ollama/issues/7704 | 2,665,123,889 | I_kwDOJ0Z1Ps6e2pAx | 7,704 | On Windows 11 pro, it does work to right click "restart to update" | {
"login": "jack-liew",
"id": 102201538,
"node_id": "U_kgDOBhd4wg",
"avatar_url": "https://avatars.githubusercontent.com/u/102201538?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jack-liew",
"html_url": "https://github.com/jack-liew",
"followers_url": "https://api.github.com/users/jack-l... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg... | closed | false | null | [] | null | 1 | 2024-11-17T00:32:03 | 2024-11-18T22:39:53 | 2024-11-18T22:39:53 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
It does work to right click "restart to update".
Thanks fo great work.
----------------------
### OS
Windows
### GPU
Intel
### CPU
Intel
### Ollama version
0.4.1 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7704/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7704/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8642 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8642/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8642/comments | https://api.github.com/repos/ollama/ollama/issues/8642/events | https://github.com/ollama/ollama/pull/8642 | 2,816,895,956 | PR_kwDOJ0Z1Ps6JSssP | 8,642 | Update README with DeepSeek model in the examples | {
"login": "harsha-s",
"id": 29120299,
"node_id": "MDQ6VXNlcjI5MTIwMjk5",
"avatar_url": "https://avatars.githubusercontent.com/u/29120299?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/harsha-s",
"html_url": "https://github.com/harsha-s",
"followers_url": "https://api.github.com/users/har... | [] | open | false | null | [] | null | 0 | 2025-01-28T23:10:31 | 2025-01-28T23:15:57 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8642",
"html_url": "https://github.com/ollama/ollama/pull/8642",
"diff_url": "https://github.com/ollama/ollama/pull/8642.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8642.patch",
"merged_at": null
} | Updated the README with the an example of DeepSeek model | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8642/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8642/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7306 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7306/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7306/comments | https://api.github.com/repos/ollama/ollama/issues/7306/events | https://github.com/ollama/ollama/pull/7306 | 2,603,930,359 | PR_kwDOJ0Z1Ps5_Y0rC | 7,306 | integration: harden embedding test | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-10-22T00:15:14 | 2024-10-22T22:25:25 | 2024-10-22T22:25:22 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7306",
"html_url": "https://github.com/ollama/ollama/pull/7306",
"diff_url": "https://github.com/ollama/ollama/pull/7306.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7306.patch",
"merged_at": "2024-10-22T22:25:22"
} | Use cosine similarity to make the embeddings tests more robust | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7306/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7306/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1671 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1671/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1671/comments | https://api.github.com/repos/ollama/ollama/issues/1671/events | https://github.com/ollama/ollama/issues/1671 | 2,053,904,792 | I_kwDOJ0Z1Ps56bBmY | 1,671 | Error: connect ECONNREFUSED 127.0.0.1:11434 | {
"login": "LTtt456c",
"id": 53040718,
"node_id": "MDQ6VXNlcjUzMDQwNzE4",
"avatar_url": "https://avatars.githubusercontent.com/u/53040718?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/LTtt456c",
"html_url": "https://github.com/LTtt456c",
"followers_url": "https://api.github.com/users/LTt... | [] | closed | false | null | [] | null | 8 | 2023-12-22T13:12:25 | 2024-06-14T10:53:14 | 2023-12-29T13:49:09 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hello everyone!
My ollama in My docker
docker Start ollama command is docker run -e OLLAMA_HOST=0.0.0.0:11434 -d -v ollama serve -p 11434:11434 --name ollama ollama/ollama
Then I in vscode open chatbot-ollama And then input npm run dev And then Report an error
↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓ Here is the error lo... | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1671/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1671/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/756 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/756/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/756/comments | https://api.github.com/repos/ollama/ollama/issues/756/events | https://github.com/ollama/ollama/issues/756 | 1,938,044,665 | I_kwDOJ0Z1Ps5zhDb5 | 756 | Mistral - Failed To Load Model | {
"login": "mattdavenport",
"id": 1127393,
"node_id": "MDQ6VXNlcjExMjczOTM=",
"avatar_url": "https://avatars.githubusercontent.com/u/1127393?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mattdavenport",
"html_url": "https://github.com/mattdavenport",
"followers_url": "https://api.github.... | [] | closed | false | null | [] | null | 2 | 2023-10-11T15:08:01 | 2023-10-11T15:40:14 | 2023-10-11T15:10:58 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I'm running macOS (Ventura 13.0.1) 16in. M1 2021. I am able to run all of the llama2 models just fine, but the following occurs when attempting to run the mistral model:
```
~ % ollama pull mistral:latest ... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/756/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/756/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2363 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2363/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2363/comments | https://api.github.com/repos/ollama/ollama/issues/2363/events | https://github.com/ollama/ollama/pull/2363 | 2,119,594,984 | PR_kwDOJ0Z1Ps5mFH5O | 2,363 | Move hub auth out to new package | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 1 | 2024-02-05T22:39:49 | 2024-02-15T17:42:21 | 2024-02-15T17:42:17 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2363",
"html_url": "https://github.com/ollama/ollama/pull/2363",
"diff_url": "https://github.com/ollama/ollama/pull/2363.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2363.patch",
"merged_at": null
} | null | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2363/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2363/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2514 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2514/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2514/comments | https://api.github.com/repos/ollama/ollama/issues/2514/events | https://github.com/ollama/ollama/issues/2514 | 2,136,760,146 | I_kwDOJ0Z1Ps5_XF9S | 2,514 | Support for safetensors | {
"login": "gracehubai",
"id": 159539863,
"node_id": "U_kgDOCYJilw",
"avatar_url": "https://avatars.githubusercontent.com/u/159539863?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gracehubai",
"html_url": "https://github.com/gracehubai",
"followers_url": "https://api.github.com/users/gra... | [] | closed | false | null | [] | null | 3 | 2024-02-15T15:02:55 | 2024-02-21T05:38:43 | 2024-02-18T06:49:05 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Do we already support inferencing safetensors? | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2514/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2514/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2214 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2214/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2214/comments | https://api.github.com/repos/ollama/ollama/issues/2214/events | https://github.com/ollama/ollama/pull/2214 | 2,102,794,214 | PR_kwDOJ0Z1Ps5lMKh3 | 2,214 | Detect lack of AVX and fallback to CPU mode | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-01-26T19:42:11 | 2024-01-26T20:06:47 | 2024-01-26T20:06:44 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2214",
"html_url": "https://github.com/ollama/ollama/pull/2214",
"diff_url": "https://github.com/ollama/ollama/pull/2214.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2214.patch",
"merged_at": "2024-01-26T20:06:44"
} | We build the GPU libraries with AVX enabled to ensure that if not all layers fit on the GPU we get better performance in a mixed mode. If the user is using a virtualization/emulation system that lacks AVX this used to result in an illegal instruction error and crash before this fix. Now we will report a warning in the... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2214/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2214/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8431 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8431/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8431/comments | https://api.github.com/repos/ollama/ollama/issues/8431/events | https://github.com/ollama/ollama/issues/8431 | 2,788,732,623 | I_kwDOJ0Z1Ps6mOK7P | 8,431 | Embedding with bge-large:latest fails with "GGML_ASSERT(i01 >= 0 && i01 < ne01) failed" | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | [] | open | false | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.g... | null | 3 | 2025-01-15T03:07:10 | 2025-01-15T06:35:56 | null | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | > @rick-github Thanks,I tried the "options":{"num_ctx":512} ,It works well! 💯
> but the logs always shows that
>
> 
_Originally posted by @Justin-12138 in [#7441](https://github.com/ollama/ollama/issues/7441#issuecomment-2591... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8431/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8431/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/1128 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1128/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1128/comments | https://api.github.com/repos/ollama/ollama/issues/1128/events | https://github.com/ollama/ollama/pull/1128 | 1,993,443,855 | PR_kwDOJ0Z1Ps5fcrpD | 1,128 | FAQ: answer a few faq questions | {
"login": "technovangelist",
"id": 633681,
"node_id": "MDQ6VXNlcjYzMzY4MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/technovangelist",
"html_url": "https://github.com/technovangelist",
"followers_url": "https://api.git... | [] | closed | false | null | [] | null | 0 | 2023-11-14T19:43:51 | 2023-11-15T23:05:14 | 2023-11-15T23:05:13 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1128",
"html_url": "https://github.com/ollama/ollama/pull/1128",
"diff_url": "https://github.com/ollama/ollama/pull/1128.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1128.patch",
"merged_at": "2023-11-15T23:05:13"
} | null | {
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/us... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1128/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1128/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4247 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4247/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4247/comments | https://api.github.com/repos/ollama/ollama/issues/4247/events | https://github.com/ollama/ollama/pull/4247 | 2,284,495,403 | PR_kwDOJ0Z1Ps5u0qMP | 4,247 | skip hidden files in list models handler | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 0 | 2024-05-08T01:34:13 | 2024-05-08T02:01:46 | 2024-05-08T02:01:45 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4247",
"html_url": "https://github.com/ollama/ollama/pull/4247",
"diff_url": "https://github.com/ollama/ollama/pull/4247.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4247.patch",
"merged_at": "2024-05-08T02:01:45"
} | Hidden files on MacOS (ex: `.DS_Store`) cause the list command to file when not skipped:
```
❯ ollama ls
Error: unqualified name:
``` | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4247/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4247/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4025 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4025/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4025/comments | https://api.github.com/repos/ollama/ollama/issues/4025/events | https://github.com/ollama/ollama/pull/4025 | 2,269,092,269 | PR_kwDOJ0Z1Ps5uBD92 | 4,025 | Update linux.md, suggest containerized install | {
"login": "ericcurtin",
"id": 1694275,
"node_id": "MDQ6VXNlcjE2OTQyNzU=",
"avatar_url": "https://avatars.githubusercontent.com/u/1694275?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ericcurtin",
"html_url": "https://github.com/ericcurtin",
"followers_url": "https://api.github.com/users... | [] | open | false | null | [] | null | 2 | 2024-04-29T13:43:19 | 2024-05-06T22:54:56 | null | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4025",
"html_url": "https://github.com/ollama/ollama/pull/4025",
"diff_url": "https://github.com/ollama/ollama/pull/4025.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4025.patch",
"merged_at": null
} | It's the simplest install process, once you run the one liner, everything just works. | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4025/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4025/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5860 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5860/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5860/comments | https://api.github.com/repos/ollama/ollama/issues/5860/events | https://github.com/ollama/ollama/pull/5860 | 2,423,755,415 | PR_kwDOJ0Z1Ps52I6Yu | 5,860 | auth: update auth | {
"login": "joshyan1",
"id": 76125168,
"node_id": "MDQ6VXNlcjc2MTI1MTY4",
"avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joshyan1",
"html_url": "https://github.com/joshyan1",
"followers_url": "https://api.github.com/users/jos... | [] | closed | false | null | [] | null | 2 | 2024-07-22T20:55:20 | 2024-09-04T13:43:34 | 2024-09-04T13:43:33 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5860",
"html_url": "https://github.com/ollama/ollama/pull/5860",
"diff_url": "https://github.com/ollama/ollama/pull/5860.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5860.patch",
"merged_at": null
} | pre cursor to setting up tests for https://github.com/ollama/ollama/pull/5441
moved initialize keypair to auth
use initialize keypair implicitly when reading pubkey | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5860/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5860/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/511 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/511/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/511/comments | https://api.github.com/repos/ollama/ollama/issues/511/events | https://github.com/ollama/ollama/issues/511 | 1,891,545,545 | I_kwDOJ0Z1Ps5wvrHJ | 511 | Support InternLM | {
"login": "vansinhu",
"id": 60632596,
"node_id": "MDQ6VXNlcjYwNjMyNTk2",
"avatar_url": "https://avatars.githubusercontent.com/u/60632596?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/vansinhu",
"html_url": "https://github.com/vansinhu",
"followers_url": "https://api.github.com/users/van... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 6 | 2023-09-12T02:09:05 | 2024-06-24T08:28:12 | 2024-03-11T19:32:53 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Dear ollama developer,
Greetings! I am vansinhu, a community developer and volunteer at InternLM. [InternLM](https://github.com/InternLM/InternLM) is a large language model similar to llama2, and we look forward to InternLM being supported in ollama. If there are any challenges or inquiries regarding support for Inter... | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/511/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/511/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8383 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8383/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8383/comments | https://api.github.com/repos/ollama/ollama/issues/8383/events | https://github.com/ollama/ollama/pull/8383 | 2,781,700,019 | PR_kwDOJ0Z1Ps6Hah72 | 8,383 | fix: run container as non-root user | {
"login": "chaturvedi-kna",
"id": 63336082,
"node_id": "MDQ6VXNlcjYzMzM2MDgy",
"avatar_url": "https://avatars.githubusercontent.com/u/63336082?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/chaturvedi-kna",
"html_url": "https://github.com/chaturvedi-kna",
"followers_url": "https://api.gi... | [] | open | false | null | [] | null | 1 | 2025-01-11T08:28:55 | 2025-01-19T02:54:33 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8383",
"html_url": "https://github.com/ollama/ollama/pull/8383",
"diff_url": "https://github.com/ollama/ollama/pull/8383.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8383.patch",
"merged_at": null
} | **Enable non-root user Container Deployment Support**
**Description**
This PR adds support for running ollama in rootless containers by properly setting up permissions for the `.ollama` directory. This change enables ollama to run in environments that enforce non-root execution such as:
* OpenShift
* Kubernet... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8383/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8383/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5306 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5306/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5306/comments | https://api.github.com/repos/ollama/ollama/issues/5306/events | https://github.com/ollama/ollama/pull/5306 | 2,375,887,192 | PR_kwDOJ0Z1Ps5zqhZW | 5,306 | Do not reinstall the CLI tools if they are already installed on macOS | {
"login": "seanchristians",
"id": 25487785,
"node_id": "MDQ6VXNlcjI1NDg3Nzg1",
"avatar_url": "https://avatars.githubusercontent.com/u/25487785?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/seanchristians",
"html_url": "https://github.com/seanchristians",
"followers_url": "https://api.gi... | [] | closed | false | null | [] | null | 2 | 2024-06-26T17:17:30 | 2024-09-06T17:49:40 | 2024-09-05T04:41:02 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5306",
"html_url": "https://github.com/ollama/ollama/pull/5306",
"diff_url": "https://github.com/ollama/ollama/pull/5306.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5306.patch",
"merged_at": null
} | Check `installed()` inside of the `install` function before running the cli setup commands. Fix #5305. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5306/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5306/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6165 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6165/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6165/comments | https://api.github.com/repos/ollama/ollama/issues/6165/events | https://github.com/ollama/ollama/issues/6165 | 2,447,319,803 | I_kwDOJ0Z1Ps6R3yL7 | 6,165 | No devices found using AMD gpus | {
"login": "cinglish",
"id": 40436106,
"node_id": "MDQ6VXNlcjQwNDM2MTA2",
"avatar_url": "https://avatars.githubusercontent.com/u/40436106?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/cinglish",
"html_url": "https://github.com/cinglish",
"followers_url": "https://api.github.com/users/cin... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5755339642,
"node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg... | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 2 | 2024-08-04T22:23:38 | 2024-09-05T19:03:53 | 2024-09-05T19:03:53 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Getting the following error when loading models with AMD gpus (Instinct MI60s):
```
rocBLAS error: Could not initialize Tensile host: No devices found
```
I have 4 devices allocated to the container and it seems to still discover them at startup:
```
time=2024-08-04T22:17:48.256Z level=I... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6165/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6165/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2989 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2989/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2989/comments | https://api.github.com/repos/ollama/ollama/issues/2989/events | https://github.com/ollama/ollama/pull/2989 | 2,174,611,866 | PR_kwDOJ0Z1Ps5pAaFv | 2,989 | CI updates | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 1 | 2024-03-07T19:26:30 | 2024-03-09T21:20:27 | 2024-03-09T21:20:24 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2989",
"html_url": "https://github.com/ollama/ollama/pull/2989",
"diff_url": "https://github.com/ollama/ollama/pull/2989.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2989.patch",
"merged_at": null
} | Not ready for review yet... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2989/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2989/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/476 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/476/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/476/comments | https://api.github.com/repos/ollama/ollama/issues/476/events | https://github.com/ollama/ollama/pull/476 | 1,884,716,024 | PR_kwDOJ0Z1Ps5ZuHR7 | 476 | tighten up the error string for `ollama show` flags | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | [] | closed | false | null | [] | null | 0 | 2023-09-06T20:37:08 | 2023-09-06T20:38:50 | 2023-09-06T20:38:49 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/476",
"html_url": "https://github.com/ollama/ollama/pull/476",
"diff_url": "https://github.com/ollama/ollama/pull/476.diff",
"patch_url": "https://github.com/ollama/ollama/pull/476.patch",
"merged_at": "2023-09-06T20:38:49"
} | null | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/476/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/476/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3018 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3018/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3018/comments | https://api.github.com/repos/ollama/ollama/issues/3018/events | https://github.com/ollama/ollama/pull/3018 | 2,177,073,503 | PR_kwDOJ0Z1Ps5pIy3- | 3,018 | Add allowed host middleware and remove `workDir` middleware | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 0 | 2024-03-09T06:14:34 | 2024-03-09T06:23:48 | 2024-03-09T06:23:47 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3018",
"html_url": "https://github.com/ollama/ollama/pull/3018",
"diff_url": "https://github.com/ollama/ollama/pull/3018.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3018.patch",
"merged_at": "2024-03-09T06:23:47"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3018/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3018/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/638 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/638/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/638/comments | https://api.github.com/repos/ollama/ollama/issues/638/events | https://github.com/ollama/ollama/pull/638 | 1,918,206,635 | PR_kwDOJ0Z1Ps5beytn | 638 | add cuda to windows generation | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 0 | 2023-09-28T20:15:47 | 2023-10-20T16:44:24 | 2023-09-29T16:53:14 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/638",
"html_url": "https://github.com/ollama/ollama/pull/638",
"diff_url": "https://github.com/ollama/ollama/pull/638.diff",
"patch_url": "https://github.com/ollama/ollama/pull/638.patch",
"merged_at": null
} | Still a bunch of stuff to do for packaging this for distribution, but I tested building and running this locally and it works along with #637 so may as well get it in. | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/638/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/638/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3616 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3616/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3616/comments | https://api.github.com/repos/ollama/ollama/issues/3616/events | https://github.com/ollama/ollama/issues/3616 | 2,240,686,081 | I_kwDOJ0Z1Ps6FjigB | 3,616 | Richer grammars | {
"login": "tezlm",
"id": 56274831,
"node_id": "MDQ6VXNlcjU2Mjc0ODMx",
"avatar_url": "https://avatars.githubusercontent.com/u/56274831?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tezlm",
"html_url": "https://github.com/tezlm",
"followers_url": "https://api.github.com/users/tezlm/follow... | [] | closed | false | null | [] | null | 9 | 2024-04-12T18:23:12 | 2024-12-05T00:37:28 | 2024-12-05T00:37:28 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What are you trying to do?
Being able to specify grammars is great, but it seems a bit underutilized at the moment. This is mostly a thought dump on how it could be improved from experimentation...
### How should we solve this?
- Using llama.cpp grammar directly would be pretty powerful and nice to have
- Speci... | {
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3616/reactions",
"total_count": 9,
"+1": 9,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3616/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8624 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8624/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8624/comments | https://api.github.com/repos/ollama/ollama/issues/8624/events | https://github.com/ollama/ollama/issues/8624 | 2,814,669,000 | I_kwDOJ0Z1Ps6nxHDI | 8,624 | Deepseek 80% size reduction | {
"login": "gileneusz",
"id": 34601970,
"node_id": "MDQ6VXNlcjM0NjAxOTcw",
"avatar_url": "https://avatars.githubusercontent.com/u/34601970?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gileneusz",
"html_url": "https://github.com/gileneusz",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | open | false | null | [] | null | 1 | 2025-01-28T05:27:54 | 2025-01-28T22:28:24 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | New quants done by unsloth.ai:
| MoE Bits | Disk Size | Type | Quality | Link | Down_proj |
|-----------|-----------|----------|---------|---------------------------------... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8624/reactions",
"total_count": 8,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 8,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8624/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/59 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/59/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/59/comments | https://api.github.com/repos/ollama/ollama/issues/59/events | https://github.com/ollama/ollama/issues/59 | 1,795,166,966 | I_kwDOJ0Z1Ps5rABL2 | 59 | generate pauses after about 50 tokens | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2023-07-08T23:42:53 | 2023-07-13T02:15:17 | 2023-07-13T02:15:17 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Generation will get paused after about 50 tokens being provided
```
% ollama run orca
>>> Write a review of the restaurant "five guys"
As an AI assistant, I cannot write a biased or subjective review, but I can provide you with some general information about the restaurant "Five Guys". Five Guys is an American f... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/59/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/59/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5278 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5278/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5278/comments | https://api.github.com/repos/ollama/ollama/issues/5278/events | https://github.com/ollama/ollama/issues/5278 | 2,373,277,484 | I_kwDOJ0Z1Ps6NdVcs | 5,278 | Is it possible to start llama server through dynamic dependency library? | {
"login": "leeyiding",
"id": 49098278,
"node_id": "MDQ6VXNlcjQ5MDk4Mjc4",
"avatar_url": "https://avatars.githubusercontent.com/u/49098278?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/leeyiding",
"html_url": "https://github.com/leeyiding",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 2 | 2024-06-25T17:43:21 | 2024-07-03T02:36:33 | 2024-07-03T02:36:33 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi, I'm trying to run Ollama in [Nanos Unikernel](https://github.com/nanovms/nanos), Unikernel a single-process operating system.
I found that in https://github.com/ollama/ollama/commit/58d95cc9bd446a8209e7388a96c70367cbafd653, the running mode of llama server changed from loading dynamic dependency libraries to run... | {
"login": "leeyiding",
"id": 49098278,
"node_id": "MDQ6VXNlcjQ5MDk4Mjc4",
"avatar_url": "https://avatars.githubusercontent.com/u/49098278?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/leeyiding",
"html_url": "https://github.com/leeyiding",
"followers_url": "https://api.github.com/users/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5278/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5278/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1579 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1579/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1579/comments | https://api.github.com/repos/ollama/ollama/issues/1579/events | https://github.com/ollama/ollama/issues/1579 | 2,046,156,422 | I_kwDOJ0Z1Ps559d6G | 1,579 | Error while running ollama locally. | {
"login": "nehalmathew1996",
"id": 82964954,
"node_id": "MDQ6VXNlcjgyOTY0OTU0",
"avatar_url": "https://avatars.githubusercontent.com/u/82964954?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nehalmathew1996",
"html_url": "https://github.com/nehalmathew1996",
"followers_url": "https://api... | [] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 11 | 2023-12-18T09:28:17 | 2024-04-05T21:14:06 | 2024-02-01T23:19:46 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ConnectionError: HTTPConnectionPool(host='localhost', port=11434): Max retries exceeded with url: /api/generate/ (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x000002298AE1EF50>: Failed to establish a new connection: [WinError 10061] No connection could be made because the target machine ... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1579/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1579/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7394 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7394/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7394/comments | https://api.github.com/repos/ollama/ollama/issues/7394/events | https://github.com/ollama/ollama/issues/7394 | 2,617,770,363 | I_kwDOJ0Z1Ps6cCAF7 | 7,394 | The Open WebUI generate unsense text, but the cli terminal can chat normally with the same hosted model | {
"login": "QiuJYWX",
"id": 68810978,
"node_id": "MDQ6VXNlcjY4ODEwOTc4",
"avatar_url": "https://avatars.githubusercontent.com/u/68810978?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/QiuJYWX",
"html_url": "https://github.com/QiuJYWX",
"followers_url": "https://api.github.com/users/QiuJYW... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q... | closed | false | null | [] | null | 9 | 2024-10-28T09:06:43 | 2024-11-13T22:12:46 | 2024-11-13T22:12:46 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
The results generated by cli terminal:

The results generated by open webui:

, not system RAM | {
"login": "mlibre",
"id": 8473036,
"node_id": "MDQ6VXNlcjg0NzMwMzY=",
"avatar_url": "https://avatars.githubusercontent.com/u/8473036?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mlibre",
"html_url": "https://github.com/mlibre",
"followers_url": "https://api.github.com/users/mlibre/foll... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 4 | 2025-01-06T08:02:41 | 2025-01-24T09:45:19 | 2025-01-24T09:45:19 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### Feature Request
**Description**:
I would like to request a feature that allows Ollama to be configured to use only GPU RAM, without utilizing CPU or system RAM. This would be beneficial for systems with limited RAM resources and high GPU RAM (VRAM) memory availability.
**Benefits**:
- Optimized performance ... | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8316/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8316/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8562 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8562/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8562/comments | https://api.github.com/repos/ollama/ollama/issues/8562/events | https://github.com/ollama/ollama/issues/8562 | 2,809,089,096 | I_kwDOJ0Z1Ps6nb0xI | 8,562 | None value being returned from with_structured_output request | {
"login": "jonmach",
"id": 7843429,
"node_id": "MDQ6VXNlcjc4NDM0Mjk=",
"avatar_url": "https://avatars.githubusercontent.com/u/7843429?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jonmach",
"html_url": "https://github.com/jonmach",
"followers_url": "https://api.github.com/users/jonmach/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 6 | 2025-01-24T10:26:05 | 2025-01-27T10:15:42 | 2025-01-24T13:27:20 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Python versions are:
langchain 0.3.15
langchain-community 0.3.15
langchain-core 0.3.31
langchain-ollama 0.2.2
ollama 0.4.7
Running ollama 0.5.7 (pip install -U ollama did not increase the version beyond 0.4.7)
---
Using **with_structured_output()** seems to work for a very simple example suc... | {
"login": "jonmach",
"id": 7843429,
"node_id": "MDQ6VXNlcjc4NDM0Mjk=",
"avatar_url": "https://avatars.githubusercontent.com/u/7843429?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jonmach",
"html_url": "https://github.com/jonmach",
"followers_url": "https://api.github.com/users/jonmach/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8562/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8562/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/130 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/130/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/130/comments | https://api.github.com/repos/ollama/ollama/issues/130/events | https://github.com/ollama/ollama/pull/130 | 1,812,847,886 | PR_kwDOJ0Z1Ps5V8NI1 | 130 | vendor in progress bar and change to bytes instead of bibytes | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | [] | closed | false | null | [] | null | 0 | 2023-07-19T22:54:28 | 2023-07-20T00:24:03 | 2023-07-20T00:24:03 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/130",
"html_url": "https://github.com/ollama/ollama/pull/130",
"diff_url": "https://github.com/ollama/ollama/pull/130.diff",
"patch_url": "https://github.com/ollama/ollama/pull/130.patch",
"merged_at": "2023-07-20T00:24:03"
} | null | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/130/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/130/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4611 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4611/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4611/comments | https://api.github.com/repos/ollama/ollama/issues/4611/events | https://github.com/ollama/ollama/issues/4611 | 2,314,861,010 | I_kwDOJ0Z1Ps6J-fnS | 4,611 | User comments on personal model page | {
"login": "razvanab",
"id": 2854730,
"node_id": "MDQ6VXNlcjI4NTQ3MzA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2854730?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/razvanab",
"html_url": "https://github.com/razvanab",
"followers_url": "https://api.github.com/users/razva... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 6573197867,
"node_id": ... | open | false | null | [] | null | 0 | 2024-05-24T09:05:07 | 2024-07-08T17:22:09 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I would like to suggest a section for user comments on the personal Ollama model page. This will help with better tuning the parameters for your version of the model. | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4611/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4611/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/4287 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4287/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4287/comments | https://api.github.com/repos/ollama/ollama/issues/4287/events | https://github.com/ollama/ollama/pull/4287 | 2,287,796,918 | PR_kwDOJ0Z1Ps5u_yAg | 4,287 | suppress ollama_llama_server.exe blank command window popup | {
"login": "ashokgelal",
"id": 401055,
"node_id": "MDQ6VXNlcjQwMTA1NQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/401055?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ashokgelal",
"html_url": "https://github.com/ashokgelal",
"followers_url": "https://api.github.com/users/a... | [] | closed | false | null | [] | null | 10 | 2024-05-09T14:19:20 | 2025-01-29T16:31:20 | 2024-11-23T21:29:08 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4287",
"html_url": "https://github.com/ollama/ollama/pull/4287",
"diff_url": "https://github.com/ollama/ollama/pull/4287.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4287.patch",
"merged_at": null
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4287/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4287/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1827 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1827/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1827/comments | https://api.github.com/repos/ollama/ollama/issues/1827/events | https://github.com/ollama/ollama/issues/1827 | 2,068,731,697 | I_kwDOJ0Z1Ps57Tlcx | 1,827 | Massive slowdown on v 0.1.18 vs 0.1.17 with same model on Intel Mac | {
"login": "pjv",
"id": 327716,
"node_id": "MDQ6VXNlcjMyNzcxNg==",
"avatar_url": "https://avatars.githubusercontent.com/u/327716?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pjv",
"html_url": "https://github.com/pjv",
"followers_url": "https://api.github.com/users/pjv/followers",
"fol... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 8 | 2024-01-06T17:18:27 | 2024-04-15T01:15:54 | 2024-01-07T05:46:19 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | <img width="255" alt="SCR-20240106-kfri" src="https://github.com/jmorganca/ollama/assets/327716/c1b5ce5b-acd5-4c81-a59a-6db39ff6a257">
I don’t have exact timings but the same model (in this case, `deepseek-coder:6.7b-instruct-q4_K_S`) generates tokens roughly 5 times faster on 0.1.17 than on 0.1.18 on my Intel Mac.
... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1827/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1827/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8108 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8108/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8108/comments | https://api.github.com/repos/ollama/ollama/issues/8108/events | https://github.com/ollama/ollama/issues/8108 | 2,740,794,394 | I_kwDOJ0Z1Ps6jXTQa | 8,108 | Error: llama runner process has terminated: error:/opt/rocm/lib/libhipblas.so.2: undefined symbol: rocblas_sgbmv_64 | {
"login": "dernikolas",
"id": 150927066,
"node_id": "U_kgDOCP722g",
"avatar_url": "https://avatars.githubusercontent.com/u/150927066?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dernikolas",
"html_url": "https://github.com/dernikolas",
"followers_url": "https://api.github.com/users/der... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5755339642,
"node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg... | open | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 2 | 2024-12-15T17:57:17 | 2025-01-06T17:54:55 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
When I start llama3.2 with `ollama run llama3.2` this error spawns:
`Error: llama runner process has terminated: error:/opt/rocm/lib/libhipblas.so.2: undefined symbol: rocblas_sgbmv_64`
I am on Cachy OS with ollama and ollama-rocm installed from extra repo and a AMD RX6800 is the GPU
### OS... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8108/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8108/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/6571 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6571/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6571/comments | https://api.github.com/repos/ollama/ollama/issues/6571/events | https://github.com/ollama/ollama/issues/6571 | 2,497,977,180 | I_kwDOJ0Z1Ps6U5Btc | 6,571 | Impossible to connect to ollama locally from another pc | {
"login": "Wilnox23",
"id": 82610998,
"node_id": "MDQ6VXNlcjgyNjEwOTk4",
"avatar_url": "https://avatars.githubusercontent.com/u/82610998?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Wilnox23",
"html_url": "https://github.com/Wilnox23",
"followers_url": "https://api.github.com/users/Wil... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 4 | 2024-08-30T18:45:36 | 2024-08-30T21:12:58 | 2024-08-30T21:11:07 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Hello, I'm using ollama on windows but I can't connect to the server from another PC on the local network.
I have already changed the following system environment variables.
`OLLAMA_HOST=0.0.0.0`
`OLLAMA_ORIGINS=*`
Thank you for your assistance.
### OS
Windows
### GPU
Nvidia
### CPU
In... | {
"login": "Wilnox23",
"id": 82610998,
"node_id": "MDQ6VXNlcjgyNjEwOTk4",
"avatar_url": "https://avatars.githubusercontent.com/u/82610998?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Wilnox23",
"html_url": "https://github.com/Wilnox23",
"followers_url": "https://api.github.com/users/Wil... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6571/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6571/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2693 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2693/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2693/comments | https://api.github.com/repos/ollama/ollama/issues/2693/events | https://github.com/ollama/ollama/issues/2693 | 2,149,835,982 | I_kwDOJ0Z1Ps6AI-TO | 2,693 | Ollama windows version not working | {
"login": "sdrshn-nmbr",
"id": 122426494,
"node_id": "U_kgDOB0wUfg",
"avatar_url": "https://avatars.githubusercontent.com/u/122426494?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sdrshn-nmbr",
"html_url": "https://github.com/sdrshn-nmbr",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 2 | 2024-02-22T19:55:55 | 2024-02-22T20:24:32 | 2024-02-22T20:24:12 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | It was working fine even yesterday, but I got an update notification and it hasn't been working since. I even tried deleting and reinstalling the installer exe, but it seems the app shows up for a few seconds and then disappears again, but powershell still recognizes the command - it just says ollama not running. | {
"login": "sdrshn-nmbr",
"id": 122426494,
"node_id": "U_kgDOB0wUfg",
"avatar_url": "https://avatars.githubusercontent.com/u/122426494?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sdrshn-nmbr",
"html_url": "https://github.com/sdrshn-nmbr",
"followers_url": "https://api.github.com/users/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2693/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2693/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5758 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5758/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5758/comments | https://api.github.com/repos/ollama/ollama/issues/5758/events | https://github.com/ollama/ollama/pull/5758 | 2,414,681,278 | PR_kwDOJ0Z1Ps51sq18 | 5,758 | marshal json automatically for some template values | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2024-07-17T22:02:46 | 2024-07-17T22:35:13 | 2024-07-17T22:35:11 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5758",
"html_url": "https://github.com/ollama/ollama/pull/5758",
"diff_url": "https://github.com/ollama/ollama/pull/5758.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5758.patch",
"merged_at": "2024-07-17T22:35:11"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5758/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5758/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3678 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3678/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3678/comments | https://api.github.com/repos/ollama/ollama/issues/3678/events | https://github.com/ollama/ollama/pull/3678 | 2,246,645,169 | PR_kwDOJ0Z1Ps5s1fe1 | 3,678 | darwin: no partial offloading if required memory greater than system | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2024-04-16T18:23:02 | 2024-04-16T19:05:57 | 2024-04-16T19:05:57 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3678",
"html_url": "https://github.com/ollama/ollama/pull/3678",
"diff_url": "https://github.com/ollama/ollama/pull/3678.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3678.patch",
"merged_at": "2024-04-16T19:05:56"
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3678/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3678/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4381 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4381/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4381/comments | https://api.github.com/repos/ollama/ollama/issues/4381/events | https://github.com/ollama/ollama/issues/4381 | 2,291,487,916 | I_kwDOJ0Z1Ps6IlVSs | 4,381 | Yi 1.5 was released | {
"login": "ddpasa",
"id": 112642920,
"node_id": "U_kgDOBrbLaA",
"avatar_url": "https://avatars.githubusercontent.com/u/112642920?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ddpasa",
"html_url": "https://github.com/ddpasa",
"followers_url": "https://api.github.com/users/ddpasa/follower... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 2 | 2024-05-12T20:23:39 | 2024-05-14T03:35:51 | 2024-05-13T21:34:04 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | https://github.com/01-ai/Yi-1.5
https://huggingface.co/collections/01-ai/yi-15-2024-05-663f3ecab5f815a3eaca7ca8
| {
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/us... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4381/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4381/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/783 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/783/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/783/comments | https://api.github.com/repos/ollama/ollama/issues/783/events | https://github.com/ollama/ollama/pull/783 | 1,942,570,643 | PR_kwDOJ0Z1Ps5cxVPh | 783 | fix: offloading on low end GPUs | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-10-13T20:08:46 | 2023-10-13T21:36:45 | 2023-10-13T21:36:44 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/783",
"html_url": "https://github.com/ollama/ollama/pull/783",
"diff_url": "https://github.com/ollama/ollama/pull/783.diff",
"patch_url": "https://github.com/ollama/ollama/pull/783.patch",
"merged_at": "2023-10-13T21:36:44"
} | Fixes two issues when using low end GPUs:
GPUs with low VRAM are disproportionately affected by overhead when offloading so any device that has less than 2GB VRAM will be exclusively CPU unless overwritten by num_gpu.
A CUDA-enabled runner will still offload to GPU even if num_gpu is 0. This is problematic when t... | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/783/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/783/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/861 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/861/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/861/comments | https://api.github.com/repos/ollama/ollama/issues/861/events | https://github.com/ollama/ollama/pull/861 | 1,954,903,919 | PR_kwDOJ0Z1Ps5dakP4 | 861 | Added /clear command to clear the buffer | {
"login": "pepperoni21",
"id": 29759371,
"node_id": "MDQ6VXNlcjI5NzU5Mzcx",
"avatar_url": "https://avatars.githubusercontent.com/u/29759371?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pepperoni21",
"html_url": "https://github.com/pepperoni21",
"followers_url": "https://api.github.com/... | [] | closed | false | null | [] | null | 3 | 2023-10-20T19:25:06 | 2023-10-20T20:10:22 | 2023-10-20T20:05:19 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/861",
"html_url": "https://github.com/ollama/ollama/pull/861",
"diff_url": "https://github.com/ollama/ollama/pull/861.diff",
"patch_url": "https://github.com/ollama/ollama/pull/861.patch",
"merged_at": null
} | Hey,
I thought a /clear command could enhance user experience.
It should work for both windows and linux/mac. | {
"login": "pepperoni21",
"id": 29759371,
"node_id": "MDQ6VXNlcjI5NzU5Mzcx",
"avatar_url": "https://avatars.githubusercontent.com/u/29759371?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pepperoni21",
"html_url": "https://github.com/pepperoni21",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/861/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/861/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4830 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4830/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4830/comments | https://api.github.com/repos/ollama/ollama/issues/4830/events | https://github.com/ollama/ollama/issues/4830 | 2,335,295,422 | I_kwDOJ0Z1Ps6LMce- | 4,830 | model import to Ollama | {
"login": "fingerfree77",
"id": 124861936,
"node_id": "U_kgDOB3E98A",
"avatar_url": "https://avatars.githubusercontent.com/u/124861936?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/fingerfree77",
"html_url": "https://github.com/fingerfree77",
"followers_url": "https://api.github.com/use... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 3 | 2024-06-05T08:54:55 | 2024-09-04T13:31:21 | 2024-09-04T13:31:21 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Is it possible to import model to Ollama from HuggingFace which is not listed in the script convert-hf-to-gguf.py?
In my case model: https://huggingface.co/ai-forever/ruGPT-3.5-13B
When I try import got error:
```
Traceback (most recent call last):
File "llm/llama.cpp/convert-hf-to-gguf.py", line 2865, in <modu... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4830/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4830/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3569 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3569/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3569/comments | https://api.github.com/repos/ollama/ollama/issues/3569/events | https://github.com/ollama/ollama/issues/3569 | 2,234,581,592 | I_kwDOJ0Z1Ps6FMQJY | 3,569 | Could anyone tell me the purpose of "ollama-linux-amd64-rocm.tgz" in the released files? | {
"login": "airforce011",
"id": 36444699,
"node_id": "MDQ6VXNlcjM2NDQ0Njk5",
"avatar_url": "https://avatars.githubusercontent.com/u/36444699?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/airforce011",
"html_url": "https://github.com/airforce011",
"followers_url": "https://api.github.com/... | [
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 1 | 2024-04-10T01:29:39 | 2024-04-23T00:02:08 | 2024-04-23T00:02:08 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What are you trying to do?
I'm trying to run Ollama on Linux (Ubuntu 22.04) with ROCm, but I'm unable to use it even after following the installation instructions in the documentation. Do I need to use the "ollama-linux-amd64-rocm.tgz" file? However, I'm not sure how to use this file, and there are no instructions... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3569/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3569/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4963 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4963/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4963/comments | https://api.github.com/repos/ollama/ollama/issues/4963/events | https://github.com/ollama/ollama/issues/4963 | 2,344,382,940 | I_kwDOJ0Z1Ps6LvHHc | 4,963 | Ollama in private EC2 | {
"login": "arpitdish",
"id": 123781157,
"node_id": "U_kgDOB2DAJQ",
"avatar_url": "https://avatars.githubusercontent.com/u/123781157?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/arpitdish",
"html_url": "https://github.com/arpitdish",
"followers_url": "https://api.github.com/users/arpitd... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2024-06-10T16:40:51 | 2024-07-10T23:01:21 | 2024-07-10T23:01:21 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
How to run ollama on private EC2
### OS
Linux
### GPU
Nvidia
### CPU
_No response_
### Ollama version
_No response_ | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4963/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4963/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4290 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4290/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4290/comments | https://api.github.com/repos/ollama/ollama/issues/4290/events | https://github.com/ollama/ollama/pull/4290 | 2,288,039,651 | PR_kwDOJ0Z1Ps5vAm1e | 4,290 | fix(server): prune files | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 1 | 2024-05-09T16:25:11 | 2024-11-23T21:28:14 | 2024-11-23T21:28:14 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4290",
"html_url": "https://github.com/ollama/ollama/pull/4290",
"diff_url": "https://github.com/ollama/ollama/pull/4290.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4290.patch",
"merged_at": null
} | follow up to #4269 | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4290/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4290/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5627 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5627/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5627/comments | https://api.github.com/repos/ollama/ollama/issues/5627/events | https://github.com/ollama/ollama/issues/5627 | 2,402,504,053 | I_kwDOJ0Z1Ps6PM011 | 5,627 | unknown model architecture: 'gemma2' | {
"login": "mouse33333",
"id": 4304174,
"node_id": "MDQ6VXNlcjQzMDQxNzQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/4304174?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mouse33333",
"html_url": "https://github.com/mouse33333",
"followers_url": "https://api.github.com/users... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2024-07-11T07:47:19 | 2024-07-11T08:02:42 | 2024-07-11T07:55:05 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I'm trying to run models on my internal Intel GPU.
Here's the software env:
1. activate a virtual env in Anaconda
2. already installed ipex-llm[cpp], ipex-llm[xpu]
3. start the Ollama serve
4. start gemma2 model
Then the log said: unknown model architecture: 'gemma2'
Here's the sre... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5627/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5627/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4011 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4011/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4011/comments | https://api.github.com/repos/ollama/ollama/issues/4011/events | https://github.com/ollama/ollama/issues/4011 | 2,267,922,429 | I_kwDOJ0Z1Ps6HLb_9 | 4,011 | When my ollama has both the gemma and llama large models, how can I enable them at the same time?Thank you. | {
"login": "joylijoy",
"id": 167687386,
"node_id": "U_kgDOCf602g",
"avatar_url": "https://avatars.githubusercontent.com/u/167687386?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joylijoy",
"html_url": "https://github.com/joylijoy",
"followers_url": "https://api.github.com/users/joylijoy/... | [] | closed | false | null | [] | null | 2 | 2024-04-29T00:54:44 | 2024-05-01T04:56:32 | 2024-05-01T04:56:32 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | When my ollama has both the gemma and llama large models, how can I enable them at the same time? Does pressing CTRL+D mean to exit the large model? Thank you. | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4011/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4011/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8457 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8457/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8457/comments | https://api.github.com/repos/ollama/ollama/issues/8457/events | https://github.com/ollama/ollama/issues/8457 | 2,792,336,253 | I_kwDOJ0Z1Ps6mb6t9 | 8,457 | model wanted in ollama please:Qwen2.5-Math-PRM-7B | {
"login": "twythebest",
"id": 89891289,
"node_id": "MDQ6VXNlcjg5ODkxMjg5",
"avatar_url": "https://avatars.githubusercontent.com/u/89891289?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/twythebest",
"html_url": "https://github.com/twythebest",
"followers_url": "https://api.github.com/use... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | open | false | null | [] | null | 0 | 2025-01-16T10:36:00 | 2025-01-16T10:36:00 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | model wanted in ollama please:Qwen2.5-Math-PRM-7B | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8457/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8457/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/6544 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6544/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6544/comments | https://api.github.com/repos/ollama/ollama/issues/6544/events | https://github.com/ollama/ollama/issues/6544 | 2,492,824,185 | I_kwDOJ0Z1Ps6UlXp5 | 6,544 | Specifying options via openai client extra_body are not handled by ollama | {
"login": "gaardhus",
"id": 46934916,
"node_id": "MDQ6VXNlcjQ2OTM0OTE2",
"avatar_url": "https://avatars.githubusercontent.com/u/46934916?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gaardhus",
"html_url": "https://github.com/gaardhus",
"followers_url": "https://api.github.com/users/gaa... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 5 | 2024-08-28T19:19:11 | 2024-09-05T21:47:15 | 2024-09-04T13:25:17 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
So I've been trying to set the num_ctx for the mistral-nemo through the openai api client, however it does not seem to have an effect.
```python
# Ollama client
client = OllamaAsyncClient(host=base_url, **kwargs)
chat_completion = await client.chat(
mes... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6544/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6544/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7106 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7106/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7106/comments | https://api.github.com/repos/ollama/ollama/issues/7106/events | https://github.com/ollama/ollama/issues/7106 | 2,568,514,399 | I_kwDOJ0Z1Ps6ZGGtf | 7,106 | Update /api/show endpoint to return a boolean describing if the model supports FIM | {
"login": "tomasz-stefaniak",
"id": 7911227,
"node_id": "MDQ6VXNlcjc5MTEyMjc=",
"avatar_url": "https://avatars.githubusercontent.com/u/7911227?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tomasz-stefaniak",
"html_url": "https://github.com/tomasz-stefaniak",
"followers_url": "https://ap... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 2 | 2024-10-06T07:32:55 | 2024-11-17T14:49:48 | 2024-11-17T14:49:48 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi,
We support Ollama as a [model provider in Continue.dev](https://docs.continue.dev/customize/model-providers/ollama) and have recently added FIM (Fill-in-the-Middle) support. Currently, there is no API we are aware of that can be used to determine if a model supports FIM, so we guess its capabilities by checking ... | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7106/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7106/timeline | null | not_planned | false |
https://api.github.com/repos/ollama/ollama/issues/6749 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6749/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6749/comments | https://api.github.com/repos/ollama/ollama/issues/6749/events | https://github.com/ollama/ollama/issues/6749 | 2,519,216,246 | I_kwDOJ0Z1Ps6WKDB2 | 6,749 | Add version when the docker container is starting | {
"login": "svaningelgem",
"id": 199434,
"node_id": "MDQ6VXNlcjE5OTQzNA==",
"avatar_url": "https://avatars.githubusercontent.com/u/199434?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/svaningelgem",
"html_url": "https://github.com/svaningelgem",
"followers_url": "https://api.github.com/u... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 2 | 2024-09-11T09:46:48 | 2024-09-12T03:50:22 | 2024-09-11T23:56:59 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | It'd be nice to know what version is running when sharing logs. So my feature request is to add the current version when you start the docker container.
| {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6749/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6749/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4271 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4271/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4271/comments | https://api.github.com/repos/ollama/ollama/issues/4271/events | https://github.com/ollama/ollama/issues/4271 | 2,286,720,880 | I_kwDOJ0Z1Ps6ITJdw | 4,271 | Partial pruning does not wrok | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 0 | 2024-05-09T01:17:56 | 2024-05-09T01:17:56 | null | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
```
time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't get file path for 'sha256:cfcf93119280c4a10c1df57335bad341e000cabbc4faff125531d941a5b0befa:partial:13': invalid digest format"
time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't g... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4271/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4271/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/8649 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8649/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8649/comments | https://api.github.com/repos/ollama/ollama/issues/8649/events | https://github.com/ollama/ollama/issues/8649 | 2,817,207,537 | I_kwDOJ0Z1Ps6n6yzx | 8,649 | Short run response duration calculations are off | {
"login": "NerdyShawn",
"id": 16088118,
"node_id": "MDQ6VXNlcjE2MDg4MTE4",
"avatar_url": "https://avatars.githubusercontent.com/u/16088118?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/NerdyShawn",
"html_url": "https://github.com/NerdyShawn",
"followers_url": "https://api.github.com/use... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 3 | 2025-01-29T04:22:55 | 2025-01-29T14:03:03 | 2025-01-29T14:03:01 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Running the smaller `deepseek-r1:1.5b` model it seems like the very short time and duration how it is being calculated is off given the duration in the response. It seems since its close to zero it throws off how the time is measured.
---
 | {
"login": "trymeouteh",
"id": 31172274,
"node_id": "MDQ6VXNlcjMxMTcyMjc0",
"avatar_url": "https://avatars.githubusercontent.com/u/31172274?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/trymeouteh",
"html_url": "https://github.com/trymeouteh",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 4 | 2024-03-01T14:12:41 | 2024-03-08T14:09:08 | 2024-03-07T18:54:15 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | To my understanding it will seem Ollama can only support text chat based models and to my understand it could support more kinds of models by adding a reading files and writing files feature...
- A way to have a model read one or more file on your system and be able to send it to the AI model to process.
- A way fo... | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2864/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2864/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/502 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/502/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/502/comments | https://api.github.com/repos/ollama/ollama/issues/502/events | https://github.com/ollama/ollama/issues/502 | 1,889,051,436 | I_kwDOJ0Z1Ps5wmKMs | 502 | Better Document /api/embeddings | {
"login": "FairyTail2000",
"id": 22645621,
"node_id": "MDQ6VXNlcjIyNjQ1NjIx",
"avatar_url": "https://avatars.githubusercontent.com/u/22645621?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/FairyTail2000",
"html_url": "https://github.com/FairyTail2000",
"followers_url": "https://api.githu... | [] | closed | false | null | [] | null | 5 | 2023-09-10T08:37:43 | 2023-10-26T00:12:03 | 2023-10-26T00:12:03 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | It's currently not very clear on how to use is and how it's relation is to the EMBED Modelfile instruction.
Can you enlighten me and/or update the wiki and examples? | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/502/reactions",
"total_count": 4,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 2
} | https://api.github.com/repos/ollama/ollama/issues/502/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7634 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7634/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7634/comments | https://api.github.com/repos/ollama/ollama/issues/7634/events | https://github.com/ollama/ollama/issues/7634 | 2,653,045,312 | I_kwDOJ0Z1Ps6eIkJA | 7,634 | Custom tool are not getting called in tool call | {
"login": "Vikneshkumarmohan",
"id": 71374469,
"node_id": "MDQ6VXNlcjcxMzc0NDY5",
"avatar_url": "https://avatars.githubusercontent.com/u/71374469?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Vikneshkumarmohan",
"html_url": "https://github.com/Vikneshkumarmohan",
"followers_url": "https... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2024-11-12T18:51:44 | 2024-12-03T03:55:02 | 2024-12-02T08:01:16 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
when using ChatOllama in Langgraph tool call is not happening for the custom tools , but it is working as expected with tools like tavily search ,
https://langchain-ai.github.io/langgraph/tutorials/multi_agent/multi-agent-collaboration/#invoke
above is the example for reference code , when ... | {
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7634/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/ollama/ollama/issues/7634/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6186 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6186/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6186/comments | https://api.github.com/repos/ollama/ollama/issues/6186/events | https://github.com/ollama/ollama/pull/6186 | 2,449,377,760 | PR_kwDOJ0Z1Ps53fc2u | 6,186 | Implement linux NUMA detection | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-08-05T20:01:08 | 2024-08-05T22:20:11 | 2024-08-05T22:20:07 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/6186",
"html_url": "https://github.com/ollama/ollama/pull/6186",
"diff_url": "https://github.com/ollama/ollama/pull/6186.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6186.patch",
"merged_at": "2024-08-05T22:20:07"
} | If the system has multiple numa nodes, enable numa support in llama.cpp If we detect `numactl` in the path, use that, else use the basic "distribute" mode.
This also removes the `use_numa` setting as the bool type is no longer useful given this parameter now requires a mode.
Fixes #6093
Fixes #2496 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6186/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6186/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5715 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5715/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5715/comments | https://api.github.com/repos/ollama/ollama/issues/5715/events | https://github.com/ollama/ollama/issues/5715 | 2,410,125,004 | I_kwDOJ0Z1Ps6Pp5bM | 5,715 | Allow using `"""` in TEMPLATE Modelfile command | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2024-07-16T03:33:31 | 2024-07-16T03:33:31 | null | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | null | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5715/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5715/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/7121 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7121/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7121/comments | https://api.github.com/repos/ollama/ollama/issues/7121/events | https://github.com/ollama/ollama/pull/7121 | 2,571,261,705 | PR_kwDOJ0Z1Ps5925z4 | 7,121 | llama: refine developer docs | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-10-07T19:18:10 | 2024-10-07T19:43:49 | 2024-10-07T19:43:46 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7121",
"html_url": "https://github.com/ollama/ollama/pull/7121",
"diff_url": "https://github.com/ollama/ollama/pull/7121.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7121.patch",
"merged_at": "2024-10-07T19:43:46"
} | null | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7121/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7121/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8183 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8183/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8183/comments | https://api.github.com/repos/ollama/ollama/issues/8183/events | https://github.com/ollama/ollama/issues/8183 | 2,752,749,059 | I_kwDOJ0Z1Ps6kE54D | 8,183 | How do I specify specific GPUs when running a model? | {
"login": "any35",
"id": 2082210,
"node_id": "MDQ6VXNlcjIwODIyMTA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2082210?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/any35",
"html_url": "https://github.com/any35",
"followers_url": "https://api.github.com/users/any35/follower... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 2 | 2024-12-20T13:30:03 | 2025-01-13T01:43:05 | 2025-01-13T01:43:04 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I have a server with 16 RTX4090 graphics cards installed.
Now I need to run different models at the same time and specify the corresponding GPUs. How can I do this?
For example: run llama 3.3 on GPU0,1; llama 3.2 on gpu 2;... | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8183/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8183/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1306 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1306/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1306/comments | https://api.github.com/repos/ollama/ollama/issues/1306/events | https://github.com/ollama/ollama/pull/1306 | 2,014,904,973 | PR_kwDOJ0Z1Ps5glFGZ | 1,306 | Fix Docker image not using GPU | {
"login": "bwest2397",
"id": 73724178,
"node_id": "MDQ6VXNlcjczNzI0MTc4",
"avatar_url": "https://avatars.githubusercontent.com/u/73724178?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bwest2397",
"html_url": "https://github.com/bwest2397",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 11 | 2023-11-28T17:06:07 | 2023-12-12T19:34:56 | 2023-11-28T21:25:23 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1306",
"html_url": "https://github.com/ollama/ollama/pull/1306",
"diff_url": "https://github.com/ollama/ollama/pull/1306.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1306.patch",
"merged_at": null
} | As it currently stands, the Docker image that gets built is seemingly unable to use the GPU despite the initial "build" stage of the Dockerfile ostensibly being built with CUDA support (i.e. built using the `nvidia/cuda` base image). As reported in https://github.com/jmorganca/ollama/issues/797, it seems that simply se... | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1306/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1306/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5667 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5667/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5667/comments | https://api.github.com/repos/ollama/ollama/issues/5667/events | https://github.com/ollama/ollama/pull/5667 | 2,406,750,057 | PR_kwDOJ0Z1Ps51SJUt | 5,667 | Enhance list command | {
"login": "kaichen",
"id": 4048,
"node_id": "MDQ6VXNlcjQwNDg=",
"avatar_url": "https://avatars.githubusercontent.com/u/4048?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kaichen",
"html_url": "https://github.com/kaichen",
"followers_url": "https://api.github.com/users/kaichen/followers"... | [] | closed | false | null | [] | null | 2 | 2024-07-13T07:21:13 | 2024-11-22T00:49:46 | 2024-11-22T00:49:46 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5667",
"html_url": "https://github.com/ollama/ollama/pull/5667",
"diff_url": "https://github.com/ollama/ollama/pull/5667.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5667.patch",
"merged_at": null
} | changes
- add missing filter description for list command
- add sort option to list command | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5667/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5667/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/24 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/24/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/24/comments | https://api.github.com/repos/ollama/ollama/issues/24/events | https://github.com/ollama/ollama/pull/24 | 1,781,626,501 | PR_kwDOJ0Z1Ps5USMRS | 24 | remove add cmd | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-06-29T23:20:02 | 2023-06-29T23:29:30 | 2023-06-29T23:29:23 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/24",
"html_url": "https://github.com/ollama/ollama/pull/24",
"diff_url": "https://github.com/ollama/ollama/pull/24.diff",
"patch_url": "https://github.com/ollama/ollama/pull/24.patch",
"merged_at": "2023-06-29T23:29:23"
} | 🧹 | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/24/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/24/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2417 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2417/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2417/comments | https://api.github.com/repos/ollama/ollama/issues/2417/events | https://github.com/ollama/ollama/issues/2417 | 2,126,159,957 | I_kwDOJ0Z1Ps5-uqBV | 2,417 | Tools support in ChatCompletion endpoints | {
"login": "pamelafox",
"id": 297042,
"node_id": "MDQ6VXNlcjI5NzA0Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/297042?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pamelafox",
"html_url": "https://github.com/pamelafox",
"followers_url": "https://api.github.com/users/pame... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 6657611864,
"node_id": ... | closed | false | {
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers"... | [
{
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/... | null | 10 | 2024-02-08T22:17:27 | 2024-06-25T04:30:24 | 2024-06-25T04:30:24 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | We'd love tools support so we can use ollama with our existing OpenAI-using apps. Not sure if that's possible across the board with all models. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2417/reactions",
"total_count": 40,
"+1": 40,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2417/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3265 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3265/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3265/comments | https://api.github.com/repos/ollama/ollama/issues/3265/events | https://github.com/ollama/ollama/issues/3265 | 2,197,028,474 | I_kwDOJ0Z1Ps6C8_56 | 3,265 | Does ollama also plan to support the sound models? | {
"login": "insooneelife",
"id": 8437769,
"node_id": "MDQ6VXNlcjg0Mzc3Njk=",
"avatar_url": "https://avatars.githubusercontent.com/u/8437769?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/insooneelife",
"html_url": "https://github.com/insooneelife",
"followers_url": "https://api.github.com... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 2 | 2024-03-20T09:18:59 | 2024-12-25T15:57:35 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What are you trying to do?
Currently, ollama supports most llms and I know that it also supports vision model. I thought it would be nice if the sound model could also be used through ollama. I wonder if there are any plans like this.
### How should we solve this?
I know that there are currently sound models rel... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3265/reactions",
"total_count": 7,
"+1": 7,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3265/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/6098 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6098/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6098/comments | https://api.github.com/repos/ollama/ollama/issues/6098/events | https://github.com/ollama/ollama/issues/6098 | 2,440,088,309 | I_kwDOJ0Z1Ps6RcMr1 | 6,098 | Why is the llama3 model missing after I restart Ollama? When I run “ollama run llama3”, it re-pulls the manifest. | {
"login": "fanjikang",
"id": 146172937,
"node_id": "U_kgDOCLZsCQ",
"avatar_url": "https://avatars.githubusercontent.com/u/146172937?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/fanjikang",
"html_url": "https://github.com/fanjikang",
"followers_url": "https://api.github.com/users/fanjik... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 3 | 2024-07-31T13:38:24 | 2024-08-27T10:25:25 | 2024-08-27T10:25:25 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Why is the llama3 model missing after I restart Ollama? When I run “ollama run llama3”, it re-pulls the manifest.
### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
v0.2.8 | {
"login": "fanjikang",
"id": 146172937,
"node_id": "U_kgDOCLZsCQ",
"avatar_url": "https://avatars.githubusercontent.com/u/146172937?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/fanjikang",
"html_url": "https://github.com/fanjikang",
"followers_url": "https://api.github.com/users/fanjik... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6098/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6098/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7150 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7150/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7150/comments | https://api.github.com/repos/ollama/ollama/issues/7150/events | https://github.com/ollama/ollama/issues/7150 | 2,576,056,905 | I_kwDOJ0Z1Ps6Zi4JJ | 7,150 | CORS (Cross-Origin Resource Sharing) | {
"login": "Apex-Predator-786",
"id": 94287474,
"node_id": "U_kgDOBZ62cg",
"avatar_url": "https://avatars.githubusercontent.com/u/94287474?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Apex-Predator-786",
"html_url": "https://github.com/Apex-Predator-786",
"followers_url": "https://api.g... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 7706482389,
"node_id": "LA_kwDOJ0Z1Ps8AAAABy1eW1Q... | open | false | null | [] | null | 0 | 2024-10-09T14:08:10 | 2024-11-06T00:01:14 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
please enable CORS (Cross-Origin Resource Sharing) in rest api
| null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7150/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7150/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/2771 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2771/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2771/comments | https://api.github.com/repos/ollama/ollama/issues/2771/events | https://github.com/ollama/ollama/pull/2771 | 2,155,434,878 | PR_kwDOJ0Z1Ps5n-7UH | 2,771 | Bump llama.cpp to b2276 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-02-27T00:51:50 | 2024-02-27T19:29:56 | 2024-02-27T19:29:53 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2771",
"html_url": "https://github.com/ollama/ollama/pull/2771",
"diff_url": "https://github.com/ollama/ollama/pull/2771.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2771.patch",
"merged_at": "2024-02-27T19:29:53"
} | Fixes #2758 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2771/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2771/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3271 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3271/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3271/comments | https://api.github.com/repos/ollama/ollama/issues/3271/events | https://github.com/ollama/ollama/issues/3271 | 2,197,650,730 | I_kwDOJ0Z1Ps6C_X0q | 3,271 | Cloudflare Tunnels Not Work After Update 1.2.9 | {
"login": "projetosfsi",
"id": 117665999,
"node_id": "U_kgDOBwNwzw",
"avatar_url": "https://avatars.githubusercontent.com/u/117665999?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/projetosfsi",
"html_url": "https://github.com/projetosfsi",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 9 | 2024-03-20T14:15:13 | 2024-11-07T18:46:03 | 2024-03-20T16:49:15 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I'm using Cloudfalre Tunnels to connect Ollama to a web domain.
Everything was working fine until the update to version 1.2.9
I started getting the 403 Access Denied error.
Regardless of how I run it, with administrative access or as a regular user, nothing changes.
After downgrading... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3271/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3271/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1730 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1730/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1730/comments | https://api.github.com/repos/ollama/ollama/issues/1730/events | https://github.com/ollama/ollama/issues/1730 | 2,057,739,752 | I_kwDOJ0Z1Ps56pp3o | 1,730 | MLX backend | {
"login": "ageorgios",
"id": 3126351,
"node_id": "MDQ6VXNlcjMxMjYzNTE=",
"avatar_url": "https://avatars.githubusercontent.com/u/3126351?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ageorgios",
"html_url": "https://github.com/ageorgios",
"followers_url": "https://api.github.com/users/ag... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 57 | 2023-12-27T20:10:50 | 2025-01-24T14:12:55 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Can ollama be converted to use MLX from Apple as backend for the models ? | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1730/reactions",
"total_count": 204,
"+1": 164,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 40,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1730/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/4407 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4407/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4407/comments | https://api.github.com/repos/ollama/ollama/issues/4407/events | https://github.com/ollama/ollama/issues/4407 | 2,293,538,684 | I_kwDOJ0Z1Ps6ItJ98 | 4,407 | Discrepancy between FAQ and program | {
"login": "FrederikSchack",
"id": 47929714,
"node_id": "MDQ6VXNlcjQ3OTI5NzE0",
"avatar_url": "https://avatars.githubusercontent.com/u/47929714?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/FrederikSchack",
"html_url": "https://github.com/FrederikSchack",
"followers_url": "https://api.gi... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2024-05-13T18:34:20 | 2024-05-14T01:02:27 | 2024-05-14T01:02:27 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
According to your FAQ, there should be a control panel in the Windows version for setting the environment variables:
https://github.com/ollama/ollama/blob/main/docs/faq.md
There isn't any control panel when you start the program, no apparent control panel or configuration file in the Ollama ... | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4407/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4407/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/852 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/852/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/852/comments | https://api.github.com/repos/ollama/ollama/issues/852/events | https://github.com/ollama/ollama/issues/852 | 1,954,491,700 | I_kwDOJ0Z1Ps50fy00 | 852 | Compress with xz for macOS and Linux GitHub Releases | {
"login": "coolaj86",
"id": 122831,
"node_id": "MDQ6VXNlcjEyMjgzMQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/122831?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/coolaj86",
"html_url": "https://github.com/coolaj86",
"followers_url": "https://api.github.com/users/coolaj8... | [] | closed | false | null | [] | null | 3 | 2023-10-20T15:08:55 | 2023-10-26T05:14:35 | 2023-10-26T00:49:39 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I am willing to help with this.
## Why `xz`?
`xz` is a standard compression tool for binary files (just like `gzip` is a standard tool for compressing text files).
It comes preinstalled on macOS and _most_ Linuxes (at least the ones used for desktop and server use, though not always in containers).
Double c... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/852/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/852/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/572 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/572/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/572/comments | https://api.github.com/repos/ollama/ollama/issues/572/events | https://github.com/ollama/ollama/issues/572 | 1,908,543,284 | I_kwDOJ0Z1Ps5xwg80 | 572 | How to prevent model answer runaway? | {
"login": "FairyTail2000",
"id": 22645621,
"node_id": "MDQ6VXNlcjIyNjQ1NjIx",
"avatar_url": "https://avatars.githubusercontent.com/u/22645621?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/FairyTail2000",
"html_url": "https://github.com/FairyTail2000",
"followers_url": "https://api.githu... | [] | closed | false | null | [] | null | 9 | 2023-09-22T09:30:55 | 2023-11-09T21:16:28 | 2023-11-09T21:16:27 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I've encountered the following issue with some models:
The model just keeps answering, providing question in the training format and answering them. This just keeps on going or it stops after one or two extra answers.
Example:
medllama2:7b-q5_0 docker gpu
Question from me (fictional):
> A male 23 year old pers... | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/572/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/572/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2658 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2658/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2658/comments | https://api.github.com/repos/ollama/ollama/issues/2658/events | https://github.com/ollama/ollama/issues/2658 | 2,147,927,324 | I_kwDOJ0Z1Ps6ABsUc | 2,658 | Ollama should clear temp files on exit. | {
"login": "amnweb",
"id": 16545063,
"node_id": "MDQ6VXNlcjE2NTQ1MDYz",
"avatar_url": "https://avatars.githubusercontent.com/u/16545063?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/amnweb",
"html_url": "https://github.com/amnweb",
"followers_url": "https://api.github.com/users/amnweb/fo... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 2 | 2024-02-22T00:12:32 | 2024-03-20T15:28:06 | 2024-03-20T15:28:06 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Found that upon exiting, Ollama does not delete temporary files, but upon starting, Ollama creates new identical files again. in temp folder ''..AppData\Local\Temp"
OS: Windows 11

| {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2658/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2658/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8304 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8304/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8304/comments | https://api.github.com/repos/ollama/ollama/issues/8304/events | https://github.com/ollama/ollama/issues/8304 | 2,768,646,500 | I_kwDOJ0Z1Ps6lBjFk | 8,304 | Unable to bind 0.0.0.0 port to all interfaces, except localhost | {
"login": "drigio",
"id": 22986282,
"node_id": "MDQ6VXNlcjIyOTg2Mjgy",
"avatar_url": "https://avatars.githubusercontent.com/u/22986282?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/drigio",
"html_url": "https://github.com/drigio",
"followers_url": "https://api.github.com/users/drigio/fo... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 6 | 2025-01-04T08:36:06 | 2025-01-24T09:46:59 | 2025-01-24T09:46:59 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I'm trying to `ollama serve` by setting env variable `OLLAMA_HOST=0.0.0.0` on windows. However I am only able to access it via localhost:11434. Unable to access it via my host ip, nor via 127.0.0.1.
There is no proxy enabled, no vpn.
I have disabled firewall in my windows machine.
ollam... | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8304/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8304/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7425 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7425/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7425/comments | https://api.github.com/repos/ollama/ollama/issues/7425/events | https://github.com/ollama/ollama/issues/7425 | 2,624,721,955 | I_kwDOJ0Z1Ps6cchQj | 7,425 | Model loaded to RAM and not VRAM | {
"login": "johanesalxd",
"id": 17249308,
"node_id": "MDQ6VXNlcjE3MjQ5MzA4",
"avatar_url": "https://avatars.githubusercontent.com/u/17249308?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/johanesalxd",
"html_url": "https://github.com/johanesalxd",
"followers_url": "https://api.github.com/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg... | closed | false | null | [] | null | 1 | 2024-10-30T16:52:20 | 2024-10-30T16:55:15 | 2024-10-30T16:54:58 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?

[server.log](https://github.com/user-attachments/files/17576314/server.log)
[lmstudio-issues](https://discord.com/channels/1110598183144399058/1294935199892639834)
[complete-specs](https://pcpartpicker... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7425/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7425/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7136 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7136/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7136/comments | https://api.github.com/repos/ollama/ollama/issues/7136/events | https://github.com/ollama/ollama/issues/7136 | 2,573,570,199 | I_kwDOJ0Z1Ps6ZZZCX | 7,136 | Tool binding prevent streaming | {
"login": "pyautoml",
"id": 147545771,
"node_id": "U_kgDOCMteqw",
"avatar_url": "https://avatars.githubusercontent.com/u/147545771?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pyautoml",
"html_url": "https://github.com/pyautoml",
"followers_url": "https://api.github.com/users/pyautoml/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 3 | 2024-10-08T15:39:32 | 2024-10-16T01:41:11 | 2024-10-16T01:41:11 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Hi there. I discovered recently that onces a model is delivered tool via bind_tools is stops recognizing streeaming, and each time it waits for the data to be fully loaded.
This snippet below works perfectly fine until tools are added:
print("AI:", end=" ", flush=True)
for chunk in self.... | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7136/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/ollama/ollama/issues/7136/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8118 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8118/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8118/comments | https://api.github.com/repos/ollama/ollama/issues/8118/events | https://github.com/ollama/ollama/pull/8118 | 2,742,578,632 | PR_kwDOJ0Z1Ps6FXfka | 8,118 | Change the update policy for OLLAMA_ORIGINS to Override | {
"login": "yaoxi-std",
"id": 57950207,
"node_id": "MDQ6VXNlcjU3OTUwMjA3",
"avatar_url": "https://avatars.githubusercontent.com/u/57950207?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yaoxi-std",
"html_url": "https://github.com/yaoxi-std",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 1 | 2024-12-16T14:43:16 | 2024-12-23T15:41:52 | 2024-12-23T15:41:51 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8118",
"html_url": "https://github.com/ollama/ollama/pull/8118",
"diff_url": "https://github.com/ollama/ollama/pull/8118.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8118.patch",
"merged_at": null
} | So that I will be able to remove some of the default origins. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8118/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8118/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7557 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7557/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7557/comments | https://api.github.com/repos/ollama/ollama/issues/7557/events | https://github.com/ollama/ollama/issues/7557 | 2,641,000,580 | I_kwDOJ0Z1Ps6daniE | 7,557 | 我的文件明明只有12GB,为什么加载的时候,却显示需要22GB显存? | {
"login": "fg2501",
"id": 164639270,
"node_id": "U_kgDOCdAyJg",
"avatar_url": "https://avatars.githubusercontent.com/u/164639270?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/fg2501",
"html_url": "https://github.com/fg2501",
"followers_url": "https://api.github.com/users/fg2501/follower... | [
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] | closed | false | null | [] | null | 4 | 2024-11-07T13:13:24 | 2024-11-08T03:21:17 | 2024-11-07T21:50:01 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?


我用的qwen2.5:14b模型,为什么我加载的时候,显示却是需要22GB显存呢?
### OS
Windows
### GPU
Nvidia
### CPU... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7557/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7557/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4825 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4825/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4825/comments | https://api.github.com/repos/ollama/ollama/issues/4825/events | https://github.com/ollama/ollama/issues/4825 | 2,334,841,826 | I_kwDOJ0Z1Ps6LKtvi | 4,825 | Adapter doesnt work | {
"login": "namogg",
"id": 108812102,
"node_id": "U_kgDOBnxXRg",
"avatar_url": "https://avatars.githubusercontent.com/u/108812102?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/namogg",
"html_url": "https://github.com/namogg",
"followers_url": "https://api.github.com/users/namogg/follower... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | [
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/us... | null | 6 | 2024-06-05T03:54:33 | 2024-07-10T19:38:32 | 2024-07-10T18:30:41 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I have converted my adapter to GGML format and create ollama model sucessfully with this comman
ADAPTER downloads/ggml-adapter-model.bin
However when testing the model, the adapter have no affect on the model.
### OS
Linux
### GPU
Nvidia
### CPU
Intel, AMD
### Ollama version
0.1.29 | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4825/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4825/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1641 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1641/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1641/comments | https://api.github.com/repos/ollama/ollama/issues/1641/events | https://github.com/ollama/ollama/issues/1641 | 2,051,188,544 | I_kwDOJ0Z1Ps56QqdA | 1,641 | Ollama hangs sometimes if it runs out of VRAM | {
"login": "nick-tonjum",
"id": 81722223,
"node_id": "MDQ6VXNlcjgxNzIyMjIz",
"avatar_url": "https://avatars.githubusercontent.com/u/81722223?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nick-tonjum",
"html_url": "https://github.com/nick-tonjum",
"followers_url": "https://api.github.com/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 3 | 2023-12-20T20:04:18 | 2024-01-21T01:54:47 | 2024-01-21T01:54:47 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi! I just have been having an issue with models that cause the system to run out of VRAM. It usually does the following:
1. (attempt to run a model via api, for example Llama2 70b)
1. ollama-runner tries to load the model into VRAM
2. ollama-runner runs out of VRAM and the process kills
3. the API hangs indefini... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1641/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1641/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5384 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5384/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5384/comments | https://api.github.com/repos/ollama/ollama/issues/5384/events | https://github.com/ollama/ollama/issues/5384 | 2,381,814,914 | I_kwDOJ0Z1Ps6N95yC | 5,384 | dolphin-phi3 and dolphin-qwen2 | {
"login": "olumolu",
"id": 162728301,
"node_id": "U_kgDOCbMJbQ",
"avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/olumolu",
"html_url": "https://github.com/olumolu",
"followers_url": "https://api.github.com/users/olumolu/foll... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 4 | 2024-06-29T15:25:22 | 2024-07-03T17:42:14 | 2024-07-03T17:42:13 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | dolphin-phi3 and qwen2 dolphin https://huggingface.co/cognitivecomputations
can we have this models so people can us them like dolphin-llama3 | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5384/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5384/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/44 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/44/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/44/comments | https://api.github.com/repos/ollama/ollama/issues/44/events | https://github.com/ollama/ollama/pull/44 | 1,792,041,442 | PR_kwDOJ0Z1Ps5U1i1u | 44 | add ollama system tray | {
"login": "hoyyeva",
"id": 63033505,
"node_id": "MDQ6VXNlcjYzMDMzNTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoyyeva",
"html_url": "https://github.com/hoyyeva",
"followers_url": "https://api.github.com/users/hoyyev... | [] | closed | false | null | [] | null | 1 | 2023-07-06T18:33:01 | 2023-07-07T17:51:57 | 2023-07-06T19:22:55 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/44",
"html_url": "https://github.com/ollama/ollama/pull/44",
"diff_url": "https://github.com/ollama/ollama/pull/44.diff",
"patch_url": "https://github.com/ollama/ollama/pull/44.patch",
"merged_at": null
} | null | {
"login": "hoyyeva",
"id": 63033505,
"node_id": "MDQ6VXNlcjYzMDMzNTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoyyeva",
"html_url": "https://github.com/hoyyeva",
"followers_url": "https://api.github.com/users/hoyyev... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/44/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/44/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5741 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5741/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5741/comments | https://api.github.com/repos/ollama/ollama/issues/5741/events | https://github.com/ollama/ollama/issues/5741 | 2,412,692,108 | I_kwDOJ0Z1Ps6PzsKM | 5,741 | ROCm Memory Issues with Long Contexts | {
"login": "ProjectMoon",
"id": 183856,
"node_id": "MDQ6VXNlcjE4Mzg1Ng==",
"avatar_url": "https://avatars.githubusercontent.com/u/183856?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ProjectMoon",
"html_url": "https://github.com/ProjectMoon",
"followers_url": "https://api.github.com/user... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 2 | 2024-07-17T05:37:09 | 2024-12-19T11:53:09 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Similar (?) to #1952. I've been noticing that ollama will crash when using long context lengths on ROCm. In particular, the most noticeable thing is that I can continue large conversations with ollama from start, while the model remains loaded in memory. But after coming back later and the model... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5741/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5741/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/8354 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8354/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8354/comments | https://api.github.com/repos/ollama/ollama/issues/8354/events | https://github.com/ollama/ollama/issues/8354 | 2,776,575,597 | I_kwDOJ0Z1Ps6lfy5t | 8,354 | Dynamic context size in OpenAI API compatibility. | {
"login": "x0wllaar",
"id": 10964379,
"node_id": "MDQ6VXNlcjEwOTY0Mzc5",
"avatar_url": "https://avatars.githubusercontent.com/u/10964379?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/x0wllaar",
"html_url": "https://github.com/x0wllaar",
"followers_url": "https://api.github.com/users/x0w... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 1 | 2025-01-09T00:08:41 | 2025-01-16T00:01:18 | 2025-01-16T00:01:18 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I noticed that the issue (#5356 ) regarding dynamically setting the context size (num_ctx) in OpenAI API was closed with a note saying it wasn't possible due to limitations of the API. However, I'd like to reopen this discussion as there seems to be a using the extra_body parameter available in the OpenAI API clients. ... | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8354/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8354/timeline | null | duplicate | false |
https://api.github.com/repos/ollama/ollama/issues/4529 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4529/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4529/comments | https://api.github.com/repos/ollama/ollama/issues/4529/events | https://github.com/ollama/ollama/issues/4529 | 2,304,933,652 | I_kwDOJ0Z1Ps6JYn8U | 4,529 | error loading model: error loading model vocabulary: unknown pre-tokenizer type: 'qwen2' | {
"login": "Anorid",
"id": 139095718,
"node_id": "U_kgDOCEpupg",
"avatar_url": "https://avatars.githubusercontent.com/u/139095718?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Anorid",
"html_url": "https://github.com/Anorid",
"followers_url": "https://api.github.com/users/Anorid/follower... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 22 | 2024-05-20T02:14:35 | 2024-07-15T09:32:18 | 2024-06-24T16:48:15 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I carefully read the contents of the readme's documentation to try and found that something went wrong
time=2024-05-20T10:06:02.688+08:00 level=INFO source=server.go:320 msg="starting llama server" cmd="/tmp/ollama2132883000/runners/cuda_v11/ollama_llama_server --model /root/autodl-tmp/models... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4529/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4529/timeline | null | completed | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.