url
stringlengths
51
54
repository_url
stringclasses
1 value
labels_url
stringlengths
65
68
comments_url
stringlengths
60
63
events_url
stringlengths
58
61
html_url
stringlengths
39
44
id
int64
1.78B
2.82B
node_id
stringlengths
18
19
number
int64
1
8.69k
title
stringlengths
1
382
user
dict
labels
listlengths
0
5
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
2
milestone
null
comments
int64
0
323
created_at
timestamp[s]
updated_at
timestamp[s]
closed_at
timestamp[s]
author_association
stringclasses
4 values
sub_issues_summary
dict
active_lock_reason
null
draft
bool
2 classes
pull_request
dict
body
stringlengths
2
118k
closed_by
dict
reactions
dict
timeline_url
stringlengths
60
63
performed_via_github_app
null
state_reason
stringclasses
4 values
is_pull_request
bool
2 classes
https://api.github.com/repos/ollama/ollama/issues/5915
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5915/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5915/comments
https://api.github.com/repos/ollama/ollama/issues/5915/events
https://github.com/ollama/ollama/pull/5915
2,427,724,333
PR_kwDOJ0Z1Ps52WWxZ
5,915
added tools when `stream=true`
{ "login": "vertrue", "id": 30557724, "node_id": "MDQ6VXNlcjMwNTU3NzI0", "avatar_url": "https://avatars.githubusercontent.com/u/30557724?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vertrue", "html_url": "https://github.com/vertrue", "followers_url": "https://api.github.com/users/vertru...
[]
closed
false
null
[]
null
4
2024-07-24T14:18:44
2024-07-30T18:17:25
2024-07-30T18:17:25
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5915", "html_url": "https://github.com/ollama/ollama/pull/5915", "diff_url": "https://github.com/ollama/ollama/pull/5915.diff", "patch_url": "https://github.com/ollama/ollama/pull/5915.patch", "merged_at": null }
fixes #5796 ``` curl --location '127.0.0.1:11434/v1/chat/completions' \ --header 'Content-Type: application/json' \ --data '{ "messages": [ { "content": "You are a helpful AI assistant that can use tools.", "role": "system" }, { "content": "What is function_1(10, 11)? use pr...
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjha...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5915/reactions", "total_count": 18, "+1": 9, "-1": 0, "laugh": 0, "hooray": 5, "confused": 0, "heart": 0, "rocket": 3, "eyes": 1 }
https://api.github.com/repos/ollama/ollama/issues/5915/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7435
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7435/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7435/comments
https://api.github.com/repos/ollama/ollama/issues/7435/events
https://github.com/ollama/ollama/issues/7435
2,625,700,359
I_kwDOJ0Z1Ps6cgQIH
7,435
加载训练生成的gguf,/v1/chat/completions和/api/generate接口返回的结果不一样
{ "login": "czhcc", "id": 4754730, "node_id": "MDQ6VXNlcjQ3NTQ3MzA=", "avatar_url": "https://avatars.githubusercontent.com/u/4754730?v=4", "gravatar_id": "", "url": "https://api.github.com/users/czhcc", "html_url": "https://github.com/czhcc", "followers_url": "https://api.github.com/users/czhcc/follower...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-10-31T01:08:25
2024-11-17T14:17:43
2024-11-17T14:17:43
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? 训练的基础模型是qwen2.5-7b。 对于同样的内容, /v1/chat/completions和/api/generate接口返回不同的结果。 /api/generate返回是正确的训练要求的结果。 ### OS Linux, Docker ### GPU Nvidia ### CPU Intel ### Ollama version 0.3.14
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7435/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7435/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8260
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8260/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8260/comments
https://api.github.com/repos/ollama/ollama/issues/8260/events
https://github.com/ollama/ollama/issues/8260
2,761,465,899
I_kwDOJ0Z1Ps6kmKAr
8,260
For CPU systems, getting files which are being used for GPU
{ "login": "Abubakkar13", "id": 45032674, "node_id": "MDQ6VXNlcjQ1MDMyNjc0", "avatar_url": "https://avatars.githubusercontent.com/u/45032674?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Abubakkar13", "html_url": "https://github.com/Abubakkar13", "followers_url": "https://api.github.com/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-12-28T04:56:52
2024-12-29T03:14:47
2024-12-29T03:14:46
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Hey, Happy holidays 😍!!! I just came across with Ollama version update, then saw some files are being downloaded, I noticed that few dll libraries are being download, So i just cross checked the folder there are 7 files(Specifically starting with **cuda*** or **cublas*** ) and in runners ...
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8260/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8260/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/382
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/382/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/382/comments
https://api.github.com/repos/ollama/ollama/issues/382/events
https://github.com/ollama/ollama/pull/382
1,857,140,673
PR_kwDOJ0Z1Ps5YRY9L
382
Closes #371
{ "login": "jesjess243", "id": 123990927, "node_id": "U_kgDOB2Pzjw", "avatar_url": "https://avatars.githubusercontent.com/u/123990927?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jesjess243", "html_url": "https://github.com/jesjess243", "followers_url": "https://api.github.com/users/jes...
[]
closed
false
null
[]
null
1
2023-08-18T18:43:29
2023-08-22T19:21:00
2023-08-22T19:21:00
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/382", "html_url": "https://github.com/ollama/ollama/pull/382", "diff_url": "https://github.com/ollama/ollama/pull/382.diff", "patch_url": "https://github.com/ollama/ollama/pull/382.patch", "merged_at": null }
Hello, ModelPath.ParseModelPath(name) now trims https:// from name if it is present. Keeping in the scope of the issue, no alternative protocols are accepted and a path with no protocol defaults to https as before. My first approach would've allowed garbage text in the protocol, but this lacks that side-effect.
{ "login": "jesjess243", "id": 123990927, "node_id": "U_kgDOB2Pzjw", "avatar_url": "https://avatars.githubusercontent.com/u/123990927?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jesjess243", "html_url": "https://github.com/jesjess243", "followers_url": "https://api.github.com/users/jes...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/382/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/382/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6473
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6473/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6473/comments
https://api.github.com/repos/ollama/ollama/issues/6473/events
https://github.com/ollama/ollama/issues/6473
2,482,749,930
I_kwDOJ0Z1Ps6T-8Hq
6,473
OpenAI Structured Output Compatability
{ "login": "jd-solanki", "id": 47495003, "node_id": "MDQ6VXNlcjQ3NDk1MDAz", "avatar_url": "https://avatars.githubusercontent.com/u/47495003?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jd-solanki", "html_url": "https://github.com/jd-solanki", "followers_url": "https://api.github.com/use...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
9
2024-08-23T09:42:33
2024-12-05T00:45:49
2024-12-05T00:45:49
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi 👋🏻 Loving ollama always ❤️ I'm eager to use newly released structured output using ollama but it looks like ollama doesn't have compatibility yet so I can just put base_url and I'll get response from my local LLM. Also, I would like will it support [streaming](https://python.useinstructor.com/concepts/p...
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6473/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6473/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4292
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4292/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4292/comments
https://api.github.com/repos/ollama/ollama/issues/4292/events
https://github.com/ollama/ollama/pull/4292
2,288,040,474
PR_kwDOJ0Z1Ps5vAnA9
4,292
remove last bits of ParseModelPath
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
open
false
null
[]
null
1
2024-05-09T16:25:46
2024-08-29T23:26:19
null
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4292", "html_url": "https://github.com/ollama/ollama/pull/4292", "diff_url": "https://github.com/ollama/ollama/pull/4292.diff", "patch_url": "https://github.com/ollama/ollama/pull/4292.patch", "merged_at": null }
follow up to #4291
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4292/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4292/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5455
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5455/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5455/comments
https://api.github.com/repos/ollama/ollama/issues/5455/events
https://github.com/ollama/ollama/issues/5455
2,388,109,167
I_kwDOJ0Z1Ps6OV6dv
5,455
ollama does not work on ALL GPU automatically
{ "login": "HeroSong666", "id": 142960235, "node_id": "U_kgDOCIVmaw", "avatar_url": "https://avatars.githubusercontent.com/u/142960235?v=4", "gravatar_id": "", "url": "https://api.github.com/users/HeroSong666", "html_url": "https://github.com/HeroSong666", "followers_url": "https://api.github.com/users/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
5
2024-07-03T09:12:14
2024-11-08T08:27:58
2024-08-01T23:51:38
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? when I use the ollama:0.1.38, I use the following command to start: ` docker run -d --gpus=all -v /root/ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama ` ollama will automatically use all 4 gpu cards for inference. when I upgrate to ollama:0.1.48, I use the same comm...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5455/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5455/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6939
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6939/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6939/comments
https://api.github.com/repos/ollama/ollama/issues/6939/events
https://github.com/ollama/ollama/pull/6939
2,546,109,699
PR_kwDOJ0Z1Ps58jqoL
6,939
docs: update llamaindex links
{ "login": "himself65", "id": 14026360, "node_id": "MDQ6VXNlcjE0MDI2MzYw", "avatar_url": "https://avatars.githubusercontent.com/u/14026360?v=4", "gravatar_id": "", "url": "https://api.github.com/users/himself65", "html_url": "https://github.com/himself65", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
0
2024-09-24T18:44:21
2024-09-24T19:15:59
2024-09-24T19:15:43
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6939", "html_url": "https://github.com/ollama/ollama/pull/6939", "diff_url": "https://github.com/ollama/ollama/pull/6939.diff", "patch_url": "https://github.com/ollama/ollama/pull/6939.patch", "merged_at": "2024-09-24T19:15:43" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6939/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6939/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1756
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1756/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1756/comments
https://api.github.com/repos/ollama/ollama/issues/1756/events
https://github.com/ollama/ollama/issues/1756
2,061,766,722
I_kwDOJ0Z1Ps565BBC
1,756
Older CUDA compute capability 3.5 and 3.7 support
{ "login": "orlyandico", "id": 1325420, "node_id": "MDQ6VXNlcjEzMjU0MjA=", "avatar_url": "https://avatars.githubusercontent.com/u/1325420?v=4", "gravatar_id": "", "url": "https://api.github.com/users/orlyandico", "html_url": "https://github.com/orlyandico", "followers_url": "https://api.github.com/users...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 6430601766, "node_id": ...
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
64
2024-01-01T20:49:39
2025-01-27T15:35:54
2024-12-10T17:47:23
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I recently put together an (old) physical machine with an Nvidia K80, which is only supported up to CUDA 11.4 and Nvidia driver 470. All my previous experiments with Ollama were with more modern GPU's. I found that Ollama doesn't use the GPU at all. I cannot find any documentation on the minimum required CUDA vers...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1756/reactions", "total_count": 3, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 3, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1756/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8543
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8543/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8543/comments
https://api.github.com/repos/ollama/ollama/issues/8543/events
https://github.com/ollama/ollama/issues/8543
2,805,677,144
I_kwDOJ0Z1Ps6nOzxY
8,543
Ollama does not perform structured output correctly.
{ "login": "aJarOfCode", "id": 103819351, "node_id": "U_kgDOBjAoVw", "avatar_url": "https://avatars.githubusercontent.com/u/103819351?v=4", "gravatar_id": "", "url": "https://api.github.com/users/aJarOfCode", "html_url": "https://github.com/aJarOfCode", "followers_url": "https://api.github.com/users/aJa...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
1
2025-01-23T01:19:21
2025-01-27T09:10:56
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? request: `{ "model":"llama3.2", "messages":datas+[ { "role":"user", "content":input_data, "images":[screenshot_base64] } ], "stream":False, "for...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8543/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8543/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/7615
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7615/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7615/comments
https://api.github.com/repos/ollama/ollama/issues/7615/events
https://github.com/ollama/ollama/pull/7615
2,648,429,620
PR_kwDOJ0Z1Ps6BeCEy
7,615
Allow Compile on older GPUs - still on CUDA 11.3
{ "login": "langstonmeister", "id": 65471211, "node_id": "MDQ6VXNlcjY1NDcxMjEx", "avatar_url": "https://avatars.githubusercontent.com/u/65471211?v=4", "gravatar_id": "", "url": "https://api.github.com/users/langstonmeister", "html_url": "https://github.com/langstonmeister", "followers_url": "https://api...
[]
open
false
null
[]
null
1
2024-11-11T07:31:03
2024-11-22T17:01:26
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7615", "html_url": "https://github.com/ollama/ollama/pull/7615", "diff_url": "https://github.com/ollama/ollama/pull/7615.diff", "patch_url": "https://github.com/ollama/ollama/pull/7615.patch", "merged_at": null }
My Tesla k40s are still able to run this version of CUDA, but are technically on Compute 3.5. These commits will allow them to compile and run the software, and are working well for me.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7615/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7615/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2331
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2331/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2331/comments
https://api.github.com/repos/ollama/ollama/issues/2331/events
https://github.com/ollama/ollama/pull/2331
2,115,267,354
PR_kwDOJ0Z1Ps5l2Y_S
2,331
Update README.md
{ "login": "AshD", "id": 233016, "node_id": "MDQ6VXNlcjIzMzAxNg==", "avatar_url": "https://avatars.githubusercontent.com/u/233016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AshD", "html_url": "https://github.com/AshD", "followers_url": "https://api.github.com/users/AshD/followers", ...
[]
closed
false
null
[]
null
4
2024-02-02T16:12:29
2024-11-21T09:13:03
2024-11-21T09:13:02
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2331", "html_url": "https://github.com/ollama/ollama/pull/2331", "diff_url": "https://github.com/ollama/ollama/pull/2331.diff", "patch_url": "https://github.com/ollama/ollama/pull/2331.patch", "merged_at": null }
Adding info on Fusion Quill. Setup info is here https://fusionquill.ai/help-setup-ollama/ Fusion Quill Personal Edition is a Windows app on the Microsoft App Store that connects to multiple AI models with workflows and UX like an Integrated Word processor with AI Chat in a split-pane UI that enables creating documen...
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/us...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2331/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2331/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4268
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4268/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4268/comments
https://api.github.com/repos/ollama/ollama/issues/4268/events
https://github.com/ollama/ollama/pull/4268
2,286,626,904
PR_kwDOJ0Z1Ps5u72dc
4,268
Convert directly from llama3
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
[]
closed
false
null
[]
null
1
2024-05-08T23:14:20
2024-05-21T21:43:38
2024-05-21T21:43:37
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4268", "html_url": "https://github.com/ollama/ollama/pull/4268", "diff_url": "https://github.com/ollama/ollama/pull/4268.diff", "patch_url": "https://github.com/ollama/ollama/pull/4268.patch", "merged_at": "2024-05-21T21:43:37" }
This change allows you to convert directly from a llama3 derived safetensors model into Ollama. It is currently *missing*: * pytorch *almost* works however the embeddings layer size is off by the eos/bos tokens This *will* work with most llama3 derivatives if they are using safetensors including `dolphin-2.9-lla...
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4268/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4268/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1651
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1651/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1651/comments
https://api.github.com/repos/ollama/ollama/issues/1651/events
https://github.com/ollama/ollama/issues/1651
2,051,690,293
I_kwDOJ0Z1Ps56Sk81
1,651
GPU not in use?
{ "login": "Y2K350", "id": 140640993, "node_id": "U_kgDOCGIC4Q", "avatar_url": "https://avatars.githubusercontent.com/u/140640993?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Y2K350", "html_url": "https://github.com/Y2K350", "followers_url": "https://api.github.com/users/Y2K350/follower...
[]
closed
false
null
[]
null
25
2023-12-21T04:51:03
2025-01-03T21:54:51
2024-01-10T15:08:43
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I am running Ollama which was installed on an arch linux system using "sudo pacman -S ollama" I am using a RTX 4090 with Nvidia's latest drivers. I also installed cuda using "sudo pacman -S cuda" I run the LLM using the command "ollama run dolphin-mixtral:latest" it does not appear to use the GPU based on GPU usa...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1651/reactions", "total_count": 8, "+1": 8, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1651/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1930
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1930/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1930/comments
https://api.github.com/repos/ollama/ollama/issues/1930/events
https://github.com/ollama/ollama/issues/1930
2,077,220,925
I_kwDOJ0Z1Ps57z-A9
1,930
Support for CogVLM wanted. CogVLM is an alternative for LLaVA
{ "login": "henryclw", "id": 97804910, "node_id": "U_kgDOBdRibg", "avatar_url": "https://avatars.githubusercontent.com/u/97804910?v=4", "gravatar_id": "", "url": "https://api.github.com/users/henryclw", "html_url": "https://github.com/henryclw", "followers_url": "https://api.github.com/users/henryclw/fo...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
7
2024-01-11T17:23:22
2024-06-24T18:14:00
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Currently ollama is supporting LLaVA, which is super great. I wonder is there a chance to load other similar models like CogVLM? https://github.com/THUDM/CogVLM
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1930/reactions", "total_count": 6, "+1": 6, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1930/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/5525
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5525/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5525/comments
https://api.github.com/repos/ollama/ollama/issues/5525/events
https://github.com/ollama/ollama/issues/5525
2,393,903,563
I_kwDOJ0Z1Ps6OsBHL
5,525
Suggestions
{ "login": "EchoOfMedivhCheats", "id": 174923936, "node_id": "U_kgDOCm0goA", "avatar_url": "https://avatars.githubusercontent.com/u/174923936?v=4", "gravatar_id": "", "url": "https://api.github.com/users/EchoOfMedivhCheats", "html_url": "https://github.com/EchoOfMedivhCheats", "followers_url": "https://...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-07-07T05:43:01
2024-07-07T05:43:01
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Terminal color/bold codes (needs 3 and 3.%) Custom terminal app (Vulkan for stuff like svgs? Maybe for videos?)) Selinux for English (nouns verbs pronouns) Voice command and output Don’t know that much assembly but I think you can daisy chain stuff I know this is written in go, but a scripting language (if x = che...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5525/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5525/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/8690
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8690/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8690/comments
https://api.github.com/repos/ollama/ollama/issues/8690/events
https://github.com/ollama/ollama/issues/8690
2,820,660,880
I_kwDOJ0Z1Ps6oH96Q
8,690
Deepseek-671B: Error: timed out waiting for llama runner to start - progress 0.00 on 8x L40S
{ "login": "orlyandico", "id": 1325420, "node_id": "MDQ6VXNlcjEzMjU0MjA=", "avatar_url": "https://avatars.githubusercontent.com/u/1325420?v=4", "gravatar_id": "", "url": "https://api.github.com/users/orlyandico", "html_url": "https://github.com/orlyandico", "followers_url": "https://api.github.com/users...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
1
2025-01-30T12:08:46
2025-01-30T12:12:22
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Ollama (0.5.7) appears to be correctly calculating how many layers to offload to the GPU with default settings. This is on a g6e.48xlarge which has 1.5TB of RAM. ``` Jan 30 11:56:19 ip-172-31-21-180 ollama[3237]: time=2025-01-30T11:56:19.283Z level=INFO source=memory.go:356 msg="offload to cuda...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8690/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8690/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/106
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/106/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/106/comments
https://api.github.com/repos/ollama/ollama/issues/106/events
https://github.com/ollama/ollama/pull/106
1,810,840,013
PR_kwDOJ0Z1Ps5V1VZH
106
README typo fix
{ "login": "isaac-mcfadyen", "id": 6243993, "node_id": "MDQ6VXNlcjYyNDM5OTM=", "avatar_url": "https://avatars.githubusercontent.com/u/6243993?v=4", "gravatar_id": "", "url": "https://api.github.com/users/isaac-mcfadyen", "html_url": "https://github.com/isaac-mcfadyen", "followers_url": "https://api.gith...
[]
closed
false
null
[]
null
1
2023-07-18T23:00:59
2023-07-18T23:24:58
2023-07-18T23:24:58
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/106", "html_url": "https://github.com/ollama/ollama/pull/106", "diff_url": "https://github.com/ollama/ollama/pull/106.diff", "patch_url": "https://github.com/ollama/ollama/pull/106.patch", "merged_at": "2023-07-18T23:24:58" }
- Simple typo fix in README.md - `13` changed to `13B` (the `B` was missing) on Nous-Hermes model in models table - Also edited `hous-hermes` to `nous-hermes` (should be n instead of h)
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/106/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/106/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3936
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3936/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3936/comments
https://api.github.com/repos/ollama/ollama/issues/3936/events
https://github.com/ollama/ollama/pull/3936
2,265,316,091
PR_kwDOJ0Z1Ps5t0VM1
3,936
pulled model not found fix - added to FAQ
{ "login": "lambdaofgod", "id": 3647577, "node_id": "MDQ6VXNlcjM2NDc1Nzc=", "avatar_url": "https://avatars.githubusercontent.com/u/3647577?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lambdaofgod", "html_url": "https://github.com/lambdaofgod", "followers_url": "https://api.github.com/us...
[]
closed
false
null
[]
null
5
2024-04-26T09:04:22
2024-07-03T17:00:00
2024-07-03T16:59:59
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3936", "html_url": "https://github.com/ollama/ollama/pull/3936", "diff_url": "https://github.com/ollama/ollama/pull/3936.diff", "patch_url": "https://github.com/ollama/ollama/pull/3936.patch", "merged_at": null }
I've added a fix by [sridvijay](https://github.com/sridvijay) from #3876 to the documentation.
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3936/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3936/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7864
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7864/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7864/comments
https://api.github.com/repos/ollama/ollama/issues/7864/events
https://github.com/ollama/ollama/issues/7864
2,698,937,631
I_kwDOJ0Z1Ps6g3oUf
7,864
Please sync with llama.cpp for the update of bert_base like models.
{ "login": "BeNhNp", "id": 33339730, "node_id": "MDQ6VXNlcjMzMzM5NzMw", "avatar_url": "https://avatars.githubusercontent.com/u/33339730?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BeNhNp", "html_url": "https://github.com/BeNhNp", "followers_url": "https://api.github.com/users/BeNhNp/fo...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
0
2024-11-27T15:34:28
2024-11-27T15:34:28
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? currently [ollama_llama_server](https://github.com/ollama/ollama/blob/main/llm/server.go#L894) can't return properly with google bert_base models, while llama.cpp have already supported, I've varified with `transformers`. ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ##...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7864/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7864/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/3601
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3601/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3601/comments
https://api.github.com/repos/ollama/ollama/issues/3601/events
https://github.com/ollama/ollama/issues/3601
2,238,409,901
I_kwDOJ0Z1Ps6Fa2yt
3,601
Docker 0.1.31 the 2nd Ollama cannot use its designated GPU
{ "login": "ww2283", "id": 21956104, "node_id": "MDQ6VXNlcjIxOTU2MTA0", "avatar_url": "https://avatars.githubusercontent.com/u/21956104?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ww2283", "html_url": "https://github.com/ww2283", "followers_url": "https://api.github.com/users/ww2283/fo...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-04-11T19:26:22
2024-04-11T19:36:45
2024-04-11T19:36:45
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I'm on Ubuntu 20.04 with two Ada6000 cards. I use docker compose to host two instance of Ollama, each has its own model because I want to use them for AutoGen. For each Ollama container I want to assign an individual GPU. In real use, I noticed that ollama1 is working perfectly fine, while oll...
{ "login": "ww2283", "id": 21956104, "node_id": "MDQ6VXNlcjIxOTU2MTA0", "avatar_url": "https://avatars.githubusercontent.com/u/21956104?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ww2283", "html_url": "https://github.com/ww2283", "followers_url": "https://api.github.com/users/ww2283/fo...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3601/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3601/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6696
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6696/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6696/comments
https://api.github.com/repos/ollama/ollama/issues/6696/events
https://github.com/ollama/ollama/pull/6696
2,512,217,526
PR_kwDOJ0Z1Ps56wsWl
6,696
Update README.md
{ "login": "rapidarchitect", "id": 126218667, "node_id": "U_kgDOB4Xxqw", "avatar_url": "https://avatars.githubusercontent.com/u/126218667?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rapidarchitect", "html_url": "https://github.com/rapidarchitect", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
0
2024-09-08T06:08:35
2024-09-08T07:36:00
2024-09-08T07:36:00
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6696", "html_url": "https://github.com/ollama/ollama/pull/6696", "diff_url": "https://github.com/ollama/ollama/pull/6696.diff", "patch_url": "https://github.com/ollama/ollama/pull/6696.patch", "merged_at": "2024-09-08T07:36:00" }
added crewai with mesop anew since there was a conflict previously
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6696/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6696/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6029
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6029/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6029/comments
https://api.github.com/repos/ollama/ollama/issues/6029/events
https://github.com/ollama/ollama/issues/6029
2,434,070,538
I_kwDOJ0Z1Ps6RFPgK
6,029
Prompt evaluation progress indicator
{ "login": "drazdra", "id": 133811709, "node_id": "U_kgDOB_nN_Q", "avatar_url": "https://avatars.githubusercontent.com/u/133811709?v=4", "gravatar_id": "", "url": "https://api.github.com/users/drazdra", "html_url": "https://github.com/drazdra", "followers_url": "https://api.github.com/users/drazdra/foll...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
1
2024-07-28T19:05:02
2024-08-14T05:23:01
null
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Prompt evaluation can take huge time, especially with long context. it can literary be hours. Right now Ollama just hangs in that phase. Is it possible to have some messages updating on the progress? It's tough when after an hour of 100% load you still don't know if it's at 15% or at 99%..
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6029/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6029/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/823
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/823/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/823/comments
https://api.github.com/repos/ollama/ollama/issues/823/events
https://github.com/ollama/ollama/issues/823
1,948,000,329
I_kwDOJ0Z1Ps50HCBJ
823
failed to verify certificate: x509: certificate signed by unknown authority
{ "login": "sureshpatel66", "id": 87146534, "node_id": "MDQ6VXNlcjg3MTQ2NTM0", "avatar_url": "https://avatars.githubusercontent.com/u/87146534?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sureshpatel66", "html_url": "https://github.com/sureshpatel66", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
18
2023-10-17T18:21:33
2024-11-21T10:18:36
2023-10-25T19:10:43
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
~$ docker exec -it ollama ollama run llama2 pulling manifest Error: pull model manifest: Get "https://registry.ollama.ai/v2/library/llama2/manifests/latest": tls: failed to verify certificate: x509: certificate signed by unknown authority please guide to solve this issue
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/823/reactions", "total_count": 7, "+1": 7, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/823/timeline
null
not_planned
false
https://api.github.com/repos/ollama/ollama/issues/7281
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7281/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7281/comments
https://api.github.com/repos/ollama/ollama/issues/7281/events
https://github.com/ollama/ollama/pull/7281
2,600,936,244
PR_kwDOJ0Z1Ps5_PgOY
7,281
added formatting to ollama ls
{ "login": "robbiemu", "id": 248927, "node_id": "MDQ6VXNlcjI0ODkyNw==", "avatar_url": "https://avatars.githubusercontent.com/u/248927?v=4", "gravatar_id": "", "url": "https://api.github.com/users/robbiemu", "html_url": "https://github.com/robbiemu", "followers_url": "https://api.github.com/users/robbiem...
[]
open
false
null
[]
null
1
2024-10-20T22:44:50
2024-11-06T22:54:30
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7281", "html_url": "https://github.com/ollama/ollama/pull/7281", "diff_url": "https://github.com/ollama/ollama/pull/7281.diff", "patch_url": "https://github.com/ollama/ollama/pull/7281.patch", "merged_at": null }
This groups output by model on disk, and does not pollute the filesize in the listing when the manifests point to the same model. ``` ./ollama ls NAME ID SIZE MODIFIED qwen2.5:32b-instruct-q4_K_M 9f13ba1299af 19 GB 9 hours ago bge-m3...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7281/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7281/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1685
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1685/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1685/comments
https://api.github.com/repos/ollama/ollama/issues/1685/events
https://github.com/ollama/ollama/issues/1685
2,054,605,354
I_kwDOJ0Z1Ps56dsoq
1,685
Ollama mobile app appeared on iOS app store
{ "login": "netsap", "id": 41242395, "node_id": "MDQ6VXNlcjQxMjQyMzk1", "avatar_url": "https://avatars.githubusercontent.com/u/41242395?v=4", "gravatar_id": "", "url": "https://api.github.com/users/netsap", "html_url": "https://github.com/netsap", "followers_url": "https://api.github.com/users/netsap/fo...
[]
closed
false
null
[]
null
5
2023-12-23T01:05:23
2024-05-10T00:27:44
2024-05-10T00:27:44
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Accessible here: https://apps.apple.com/us/app/ollama/id6471840114 I'm assuming this is not legitimate?
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1685/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 2, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1685/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2043
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2043/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2043/comments
https://api.github.com/repos/ollama/ollama/issues/2043/events
https://github.com/ollama/ollama/pull/2043
2,087,600,434
PR_kwDOJ0Z1Ps5kY2C0
2,043
Dockerfile: use variables for package version
{ "login": "stevenbecht", "id": 9442836, "node_id": "MDQ6VXNlcjk0NDI4MzY=", "avatar_url": "https://avatars.githubusercontent.com/u/9442836?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stevenbecht", "html_url": "https://github.com/stevenbecht", "followers_url": "https://api.github.com/us...
[]
closed
false
null
[]
null
2
2024-01-18T06:23:39
2024-02-21T00:26:49
2024-02-21T00:26:49
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2043", "html_url": "https://github.com/ollama/ollama/pull/2043", "diff_url": "https://github.com/ollama/ollama/pull/2043.diff", "patch_url": "https://github.com/ollama/ollama/pull/2043.patch", "merged_at": null }
Update Dockerfile to use variables instead of hardcoded values
{ "login": "stevenbecht", "id": 9442836, "node_id": "MDQ6VXNlcjk0NDI4MzY=", "avatar_url": "https://avatars.githubusercontent.com/u/9442836?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stevenbecht", "html_url": "https://github.com/stevenbecht", "followers_url": "https://api.github.com/us...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2043/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2043/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5285
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5285/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5285/comments
https://api.github.com/repos/ollama/ollama/issues/5285/events
https://github.com/ollama/ollama/pull/5285
2,373,797,637
PR_kwDOJ0Z1Ps5zjnWI
5,285
OpenAI: /v1/embeddings compatibility
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjha...
[]
closed
false
null
[]
null
1
2024-06-25T22:47:36
2024-07-16T20:36:11
2024-07-16T20:36:09
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5285", "html_url": "https://github.com/ollama/ollama/pull/5285", "diff_url": "https://github.com/ollama/ollama/pull/5285.diff", "patch_url": "https://github.com/ollama/ollama/pull/5285.patch", "merged_at": "2024-07-16T20:36:08" }
In anticipation of #5127 Resolves #2416 ``` curl http://localhost:11434/v1/embeddings \ -H "Content-Type: application/json" \ -d '{ "input": ["hello", "hi"], "model": "all-minilm" }' ``` ``` { "object": "list", "data": [ { "object": "embedding", "embedding": [ ...
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjha...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5285/reactions", "total_count": 3, "+1": 0, "-1": 0, "laugh": 0, "hooray": 2, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/ollama/ollama/issues/5285/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7969
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7969/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7969/comments
https://api.github.com/repos/ollama/ollama/issues/7969/events
https://github.com/ollama/ollama/issues/7969
2,723,362,146
I_kwDOJ0Z1Ps6iUzVi
7,969
Administrative / silent install is borked
{ "login": "Lukas-UAUX", "id": 43141094, "node_id": "MDQ6VXNlcjQzMTQxMDk0", "avatar_url": "https://avatars.githubusercontent.com/u/43141094?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Lukas-UAUX", "html_url": "https://github.com/Lukas-UAUX", "followers_url": "https://api.github.com/use...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 5860134234, "node_id": ...
open
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
1
2024-12-06T15:50:06
2024-12-10T17:30:35
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? For deployment scenarios like for classrooms or to regular managed devices the setup needs to perform an administrative / unattended install. For InnoSetup built installers this can normally be done via the command-line switches /SILENT (or /VERYSILENT) or via a response file via /LOADINF. (See:...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7969/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7969/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2811
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2811/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2811/comments
https://api.github.com/repos/ollama/ollama/issues/2811/events
https://github.com/ollama/ollama/issues/2811
2,159,235,157
I_kwDOJ0Z1Ps6As1BV
2,811
Mistral Instruct models prompt does not use <s> or </s>
{ "login": "louisabraham", "id": 13174805, "node_id": "MDQ6VXNlcjEzMTc0ODA1", "avatar_url": "https://avatars.githubusercontent.com/u/13174805?v=4", "gravatar_id": "", "url": "https://api.github.com/users/louisabraham", "html_url": "https://github.com/louisabraham", "followers_url": "https://api.github.c...
[]
closed
false
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
[ { "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api...
null
5
2024-02-28T15:29:34
2024-07-18T22:45:02
2024-07-18T22:45:02
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
from https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1#instruction-format > \<s> [INST] Instruction [/INST] Model answer\</s> [INST] Follow-up instruction [/INST] I didn't see `<s>` as part of the prompt when using `OLLAMA_DEBUG=1 ollama start`. Is that expected?
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2811/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2811/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6611
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6611/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6611/comments
https://api.github.com/repos/ollama/ollama/issues/6611/events
https://github.com/ollama/ollama/pull/6611
2,503,463,117
PR_kwDOJ0Z1Ps56S5Sr
6,611
Make stall duration timeout configurable
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
1
2024-09-03T18:04:25
2024-09-05T21:00:08
2024-09-05T21:00:08
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6611", "html_url": "https://github.com/ollama/ollama/pull/6611", "diff_url": "https://github.com/ollama/ollama/pull/6611.diff", "patch_url": "https://github.com/ollama/ollama/pull/6611.patch", "merged_at": "2024-09-05T21:00:08" }
With the new very large parameter models, some users are willing to wait for a very long time for models to load. Fixes #6031
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6611/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 1, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6611/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/329
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/329/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/329/comments
https://api.github.com/repos/ollama/ollama/issues/329/events
https://github.com/ollama/ollama/pull/329
1,846,179,798
PR_kwDOJ0Z1Ps5XsTL-
329
Add tutorials for using Langchain with ollama
{ "login": "technovangelist", "id": 633681, "node_id": "MDQ6VXNlcjYzMzY4MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4", "gravatar_id": "", "url": "https://api.github.com/users/technovangelist", "html_url": "https://github.com/technovangelist", "followers_url": "https://api.git...
[]
closed
false
null
[]
null
1
2023-08-11T04:30:50
2024-01-05T18:40:54
2023-08-11T22:19:39
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/329", "html_url": "https://github.com/ollama/ollama/pull/329", "diff_url": "https://github.com/ollama/ollama/pull/329.diff", "patch_url": "https://github.com/ollama/ollama/pull/329.patch", "merged_at": "2023-08-11T22:19:39" }
null
{ "login": "technovangelist", "id": 633681, "node_id": "MDQ6VXNlcjYzMzY4MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4", "gravatar_id": "", "url": "https://api.github.com/users/technovangelist", "html_url": "https://github.com/technovangelist", "followers_url": "https://api.git...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/329/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/329/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7241
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7241/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7241/comments
https://api.github.com/repos/ollama/ollama/issues/7241/events
https://github.com/ollama/ollama/issues/7241
2,595,035,489
I_kwDOJ0Z1Ps6arRlh
7,241
add module/
{ "login": "malv-c", "id": 19170213, "node_id": "MDQ6VXNlcjE5MTcwMjEz", "avatar_url": "https://avatars.githubusercontent.com/u/19170213?v=4", "gravatar_id": "", "url": "https://api.github.com/users/malv-c", "html_url": "https://github.com/malv-c", "followers_url": "https://api.github.com/users/malv-c/fo...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2024-10-17T15:12:19
2024-10-20T06:06:27
2024-10-20T06:06:27
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
null
{ "login": "malv-c", "id": 19170213, "node_id": "MDQ6VXNlcjE5MTcwMjEz", "avatar_url": "https://avatars.githubusercontent.com/u/19170213?v=4", "gravatar_id": "", "url": "https://api.github.com/users/malv-c", "html_url": "https://github.com/malv-c", "followers_url": "https://api.github.com/users/malv-c/fo...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7241/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7241/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2602
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2602/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2602/comments
https://api.github.com/repos/ollama/ollama/issues/2602/events
https://github.com/ollama/ollama/issues/2602
2,143,418,344
I_kwDOJ0Z1Ps5_wffo
2,602
Windows preview - please let us set the location where models are stored
{ "login": "axefrog", "id": 298883, "node_id": "MDQ6VXNlcjI5ODg4Mw==", "avatar_url": "https://avatars.githubusercontent.com/u/298883?v=4", "gravatar_id": "", "url": "https://api.github.com/users/axefrog", "html_url": "https://github.com/axefrog", "followers_url": "https://api.github.com/users/axefrog/fo...
[]
closed
false
null
[]
null
2
2024-02-20T02:36:27
2024-03-06T18:06:53
2024-02-20T03:38:25
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
My drive C is a bit ancient right now. It's an old 250GB SSD and at any given time seems to have about 5-10gb free, so I'm forever doing cleanups to stop it running out of space. In contrast, I have about 2-3TB of free space on my other drives. I would _much_ prefer it if ollama would store models on one of those drive...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2602/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2602/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7647
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7647/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7647/comments
https://api.github.com/repos/ollama/ollama/issues/7647/events
https://github.com/ollama/ollama/issues/7647
2,654,957,622
I_kwDOJ0Z1Ps6eP3A2
7,647
Sending "hidden" messages for attachments
{ "login": "robotom", "id": 45123215, "node_id": "MDQ6VXNlcjQ1MTIzMjE1", "avatar_url": "https://avatars.githubusercontent.com/u/45123215?v=4", "gravatar_id": "", "url": "https://api.github.com/users/robotom", "html_url": "https://github.com/robotom", "followers_url": "https://api.github.com/users/roboto...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2024-11-13T10:20:06
2024-11-13T19:53:25
2024-11-13T19:53:24
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I'm building a frontend and trying to send attachment contents and an accompanying prompt (e.g. "analyze this and wait for user prompt before responding") as a "hidden" message so that it's not rendered to the user on the frontend. I'm trying to avoid seeing 10 pages of a doc rendered in the chat window. Perhaps th...
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7647/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7647/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8349
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8349/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8349/comments
https://api.github.com/repos/ollama/ollama/issues/8349/events
https://github.com/ollama/ollama/issues/8349
2,775,993,379
I_kwDOJ0Z1Ps6ldkwj
8,349
can't use gpu after update
{ "login": "Heart-eartH", "id": 75364983, "node_id": "MDQ6VXNlcjc1MzY0OTgz", "avatar_url": "https://avatars.githubusercontent.com/u/75364983?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Heart-eartH", "html_url": "https://github.com/Heart-eartH", "followers_url": "https://api.github.com/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg...
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
7
2025-01-08T17:57:28
2025-01-28T20:46:29
2025-01-28T20:46:29
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Previously, it run on the GPU by default, but after the update, it run on the CPU. What can I do to change it ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version 0.5.4
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8349/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8349/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1552
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1552/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1552/comments
https://api.github.com/repos/ollama/ollama/issues/1552/events
https://github.com/ollama/ollama/pull/1552
2,044,274,427
PR_kwDOJ0Z1Ps5iI-Kk
1,552
add lint and test on pull_request
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
0
2023-12-15T19:34:07
2024-01-11T17:37:47
2024-01-11T17:37:46
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1552", "html_url": "https://github.com/ollama/ollama/pull/1552", "diff_url": "https://github.com/ollama/ollama/pull/1552.diff", "patch_url": "https://github.com/ollama/ollama/pull/1552.patch", "merged_at": "2024-01-11T17:37:46" }
fixes a bug with generate where `get_flags` errors on ubuntu (and likely windows) when building cuda on a cuda-less system fixes a bug in windows where `/api/list` does not return models correctly both lint and test requires go generate results so do it once then propagate the artifacts to the rest of the pipelin...
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1552/reactions", "total_count": 2, "+1": 1, "-1": 0, "laugh": 0, "hooray": 1, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1552/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6272
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6272/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6272/comments
https://api.github.com/repos/ollama/ollama/issues/6272/events
https://github.com/ollama/ollama/issues/6272
2,457,035,391
I_kwDOJ0Z1Ps6Sc2J_
6,272
Ollama Creat 手动部署 报错 Error: invalid file magic
{ "login": "JaminYan", "id": 83393121, "node_id": "MDQ6VXNlcjgzMzkzMTIx", "avatar_url": "https://avatars.githubusercontent.com/u/83393121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/JaminYan", "html_url": "https://github.com/JaminYan", "followers_url": "https://api.github.com/users/Jam...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
22
2024-08-09T03:47:19
2024-11-11T09:41:36
2024-08-25T20:15:11
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? 按飞书文档找的modelfile 文件,ollama creat 后输出报错 Error: invalid file magic ,因此无法部署在ollama ### OS Windows ### GPU Nvidia ### CPU AMD ### Ollama version 0.34
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6272/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6272/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5633
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5633/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5633/comments
https://api.github.com/repos/ollama/ollama/issues/5633/events
https://github.com/ollama/ollama/pull/5633
2,403,487,442
PR_kwDOJ0Z1Ps51HNIc
5,633
Update README.md >> Extension & Plugins :Terraform AWS Ollama & Open WebUI
{ "login": "xuyangbocn", "id": 11894567, "node_id": "MDQ6VXNlcjExODk0NTY3", "avatar_url": "https://avatars.githubusercontent.com/u/11894567?v=4", "gravatar_id": "", "url": "https://api.github.com/users/xuyangbocn", "html_url": "https://github.com/xuyangbocn", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
0
2024-07-11T15:29:25
2024-11-21T10:28:58
2024-11-21T10:28:58
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5633", "html_url": "https://github.com/ollama/ollama/pull/5633", "diff_url": "https://github.com/ollama/ollama/pull/5633.diff", "patch_url": "https://github.com/ollama/ollama/pull/5633.patch", "merged_at": "2024-11-21T10:28:58" }
A Terraform module to deploy on AWS a ready-to-use Ollama service, together with its front end Open WebUI service.
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/us...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5633/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5633/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3733
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3733/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3733/comments
https://api.github.com/repos/ollama/ollama/issues/3733/events
https://github.com/ollama/ollama/issues/3733
2,250,701,058
I_kwDOJ0Z1Ps6GJvkC
3,733
I wrote an uninstallation script.
{ "login": "TikaFlow", "id": 33306473, "node_id": "MDQ6VXNlcjMzMzA2NDcz", "avatar_url": "https://avatars.githubusercontent.com/u/33306473?v=4", "gravatar_id": "", "url": "https://api.github.com/users/TikaFlow", "html_url": "https://github.com/TikaFlow", "followers_url": "https://api.github.com/users/Tik...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-04-18T13:26:28
2024-04-19T01:59:32
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi, I wrote an uninstallation script and [this](https://github.com/TikaFlow/tika_grocery/tree/main/delOllama) is my script. If you think it's not bad, you might consider referencing my script in your [documentation](https://github.com/ollama/ollama/blob/main/docs/linux.md#uninstall). Thanks!
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3733/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3733/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/4409
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4409/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4409/comments
https://api.github.com/repos/ollama/ollama/issues/4409/events
https://github.com/ollama/ollama/issues/4409
2,293,631,935
I_kwDOJ0Z1Ps6Itgu_
4,409
Windows security: ALERT: Trojan:Win32/Wacatac.H!ml detected
{ "login": "boessu", "id": 2807976, "node_id": "MDQ6VXNlcjI4MDc5NzY=", "avatar_url": "https://avatars.githubusercontent.com/u/2807976?v=4", "gravatar_id": "", "url": "https://api.github.com/users/boessu", "html_url": "https://github.com/boessu", "followers_url": "https://api.github.com/users/boessu/foll...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-05-13T19:17:50
2024-05-13T20:03:24
2024-05-13T20:03:24
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? There is an alarm in Windows Defender which sets the following file under quarantine: file: %USERPROFILE%\AppData\Local\Programs\Ollama\ollama app.exe Windows Version: Windows 11 Pro, 23H2 You'll get the warning with update and also if you install it new. ### OS Windows ### GPU AMD #...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4409/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4409/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5743
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5743/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5743/comments
https://api.github.com/repos/ollama/ollama/issues/5743/events
https://github.com/ollama/ollama/issues/5743
2,413,191,986
I_kwDOJ0Z1Ps6P1mMy
5,743
process cant't kill on windows 10
{ "login": "wszgrcy", "id": 9607121, "node_id": "MDQ6VXNlcjk2MDcxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/9607121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wszgrcy", "html_url": "https://github.com/wszgrcy", "followers_url": "https://api.github.com/users/wszgrcy/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg...
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
5
2024-07-17T10:04:20
2024-07-19T03:29:24
2024-07-19T03:29:24
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I use child process in node to start ollama,and when I kill process,The process will be partially retained(`ollama.exe` is killed ollama_llama_server retained) ![image](https://github.com/user-attachments/assets/48d9f7d5-92dc-4357-ac1b-f293b994e788) Ubuntu will correctly kill all processes ...
{ "login": "wszgrcy", "id": 9607121, "node_id": "MDQ6VXNlcjk2MDcxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/9607121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wszgrcy", "html_url": "https://github.com/wszgrcy", "followers_url": "https://api.github.com/users/wszgrcy/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5743/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5743/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7662
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7662/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7662/comments
https://api.github.com/repos/ollama/ollama/issues/7662/events
https://github.com/ollama/ollama/issues/7662
2,657,523,102
I_kwDOJ0Z1Ps6eZpWe
7,662
How can I remove specific PARAMETER from model?
{ "login": "wwjCMP", "id": 32979859, "node_id": "MDQ6VXNlcjMyOTc5ODU5", "avatar_url": "https://avatars.githubusercontent.com/u/32979859?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wwjCMP", "html_url": "https://github.com/wwjCMP", "followers_url": "https://api.github.com/users/wwjCMP/fo...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-11-14T04:27:14
2024-11-14T14:17:11
2024-11-14T14:17:11
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
How can I remove specific PARAMETER from model For example, delete PARAMETER num_gpu 30 from qwen2.5-coder:32b-instruct-q8_0 ollama show qwen2.5-coder:32b-instruct-q8_0 Model architecture qwen2 parameters 32.8B context length 32768 embedding length 5120 quantiza...
{ "login": "wwjCMP", "id": 32979859, "node_id": "MDQ6VXNlcjMyOTc5ODU5", "avatar_url": "https://avatars.githubusercontent.com/u/32979859?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wwjCMP", "html_url": "https://github.com/wwjCMP", "followers_url": "https://api.github.com/users/wwjCMP/fo...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7662/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7662/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3330
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3330/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3330/comments
https://api.github.com/repos/ollama/ollama/issues/3330/events
https://github.com/ollama/ollama/issues/3330
2,204,580,969
I_kwDOJ0Z1Ps6DZzxp
3,330
Failed to open the file /tmp/tmp.RbBP0lFvPD/ollama:
{ "login": "hansaskov", "id": 70580312, "node_id": "MDQ6VXNlcjcwNTgwMzEy", "avatar_url": "https://avatars.githubusercontent.com/u/70580312?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hansaskov", "html_url": "https://github.com/hansaskov", "followers_url": "https://api.github.com/users/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-03-24T22:08:56
2024-03-24T22:17:04
2024-03-24T22:17:04
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I am unable to download and install ollama on ubuntu 22.04. I used the following command ```bash sudo curl -fsSL https://ollama.com/install.sh | sh ``` And i get the following output ``` >>> Downloading ollama... Warning: Failed to open the file /tmp/tmp.uSBa9PbYEo/ollama: No such ...
{ "login": "hansaskov", "id": 70580312, "node_id": "MDQ6VXNlcjcwNTgwMzEy", "avatar_url": "https://avatars.githubusercontent.com/u/70580312?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hansaskov", "html_url": "https://github.com/hansaskov", "followers_url": "https://api.github.com/users/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3330/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3330/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8271
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8271/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8271/comments
https://api.github.com/repos/ollama/ollama/issues/8271/events
https://github.com/ollama/ollama/issues/8271
2,763,974,531
I_kwDOJ0Z1Ps6kvueD
8,271
llama runner process terminated: CUDA error
{ "login": "iplayfast", "id": 751306, "node_id": "MDQ6VXNlcjc1MTMwNg==", "avatar_url": "https://avatars.githubusercontent.com/u/751306?v=4", "gravatar_id": "", "url": "https://api.github.com/users/iplayfast", "html_url": "https://github.com/iplayfast", "followers_url": "https://api.github.com/users/ipla...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q...
closed
false
null
[]
null
14
2024-12-31T05:34:54
2025-01-01T07:33:02
2025-01-01T07:33:02
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I'm making a little embedding example. The text is chunked fine, and embedded. But on retrieval, the I get this weird message. Then if I go to the cli and try running the same model, I get the same message. I'm pretty sure it has to do with having the embedded model in memory at the same ...
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8271/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8271/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7839
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7839/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7839/comments
https://api.github.com/repos/ollama/ollama/issues/7839/events
https://github.com/ollama/ollama/pull/7839
2,693,432,609
PR_kwDOJ0Z1Ps6DI6lo
7,839
fix(docs): Replace hard coded num of make threads w/ CPU cores
{ "login": "sammcj", "id": 862951, "node_id": "MDQ6VXNlcjg2Mjk1MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sammcj", "html_url": "https://github.com/sammcj", "followers_url": "https://api.github.com/users/sammcj/follow...
[]
closed
false
null
[]
null
0
2024-11-26T06:54:28
2024-12-06T22:51:09
2024-12-06T22:51:04
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7839", "html_url": "https://github.com/ollama/ollama/pull/7839", "diff_url": "https://github.com/ollama/ollama/pull/7839.diff", "patch_url": "https://github.com/ollama/ollama/pull/7839.patch", "merged_at": null }
Replace the seemingly random number of make threads with the number CPU cores / 2.
{ "login": "sammcj", "id": 862951, "node_id": "MDQ6VXNlcjg2Mjk1MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sammcj", "html_url": "https://github.com/sammcj", "followers_url": "https://api.github.com/users/sammcj/follow...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7839/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7839/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3914
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3914/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3914/comments
https://api.github.com/repos/ollama/ollama/issues/3914/events
https://github.com/ollama/ollama/pull/3914
2,264,038,325
PR_kwDOJ0Z1Ps5twARO
3,914
Improve mac parallel performance
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
0
2024-04-25T16:40:14
2024-04-25T23:28:34
2024-04-25T23:28:31
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3914", "html_url": "https://github.com/ollama/ollama/pull/3914", "diff_url": "https://github.com/ollama/ollama/pull/3914.diff", "patch_url": "https://github.com/ollama/ollama/pull/3914.patch", "merged_at": "2024-04-25T23:28:31" }
Carries #3900 Testing underway...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3914/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3914/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4048
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4048/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4048/comments
https://api.github.com/repos/ollama/ollama/issues/4048/events
https://github.com/ollama/ollama/issues/4048
2,271,291,702
I_kwDOJ0Z1Ps6HYSk2
4,048
Windows installation is behaving like malware
{ "login": "eabase", "id": 52289379, "node_id": "MDQ6VXNlcjUyMjg5Mzc5", "avatar_url": "https://avatars.githubusercontent.com/u/52289379?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eabase", "html_url": "https://github.com/eabase", "followers_url": "https://api.github.com/users/eabase/fo...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
8
2024-04-30T12:06:32
2024-05-08T12:43:37
2024-05-01T16:34:27
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? This is beyond bad. 1. It installs stuff all over the place, without informing the user. ``` C:\Users\xxxx\AppData\Roaming\Microsoft\Windows\Start Menu\Programs C:\Users\xxxx\AppData\Local\Temp\ollamaNNNNNN C:\Users\xxxx\.ollama\id_ed25519 C:\Users\xxxx\AppData\Local\Programs\Oll...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4048/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4048/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3025
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3025/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3025/comments
https://api.github.com/repos/ollama/ollama/issues/3025/events
https://github.com/ollama/ollama/pull/3025
2,177,367,153
PR_kwDOJ0Z1Ps5pJu-X
3,025
Rename ROCm deps file to avoid confusion
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
0
2024-03-09T19:22:33
2024-03-10T03:50:45
2024-03-10T01:48:39
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3025", "html_url": "https://github.com/ollama/ollama/pull/3025", "diff_url": "https://github.com/ollama/ollama/pull/3025.diff", "patch_url": "https://github.com/ollama/ollama/pull/3025.patch", "merged_at": "2024-03-10T01:48:39" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3025/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3025/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7111
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7111/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7111/comments
https://api.github.com/repos/ollama/ollama/issues/7111/events
https://github.com/ollama/ollama/issues/7111
2,569,542,877
I_kwDOJ0Z1Ps6ZKBzd
7,111
Any plans to add nvidia/NVLM-D-72B new model to ollama?
{ "login": "SilverLaius", "id": 26542231, "node_id": "MDQ6VXNlcjI2NTQyMjMx", "avatar_url": "https://avatars.githubusercontent.com/u/26542231?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SilverLaius", "html_url": "https://github.com/SilverLaius", "followers_url": "https://api.github.com/...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
1
2024-10-07T07:08:00
2024-10-07T07:09:13
2024-10-07T07:08:51
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Nvidia just released a state of the art model that would be a good addition to ollama. Any plans to add it? [https://huggingface.co/nvidia/NVLM-D-72B/tree/main](https://huggingface.co/nvidia/NVLM-D-72B/tree/main)
{ "login": "SilverLaius", "id": 26542231, "node_id": "MDQ6VXNlcjI2NTQyMjMx", "avatar_url": "https://avatars.githubusercontent.com/u/26542231?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SilverLaius", "html_url": "https://github.com/SilverLaius", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7111/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7111/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7619
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7619/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7619/comments
https://api.github.com/repos/ollama/ollama/issues/7619/events
https://github.com/ollama/ollama/issues/7619
2,648,701,609
I_kwDOJ0Z1Ps6d3_qp
7,619
llama3.2-vision on multi gpu error
{ "login": "18600709862", "id": 46145698, "node_id": "MDQ6VXNlcjQ2MTQ1Njk4", "avatar_url": "https://avatars.githubusercontent.com/u/46145698?v=4", "gravatar_id": "", "url": "https://api.github.com/users/18600709862", "html_url": "https://github.com/18600709862", "followers_url": "https://api.github.com/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg...
closed
false
null
[]
null
7
2024-11-11T09:26:43
2024-11-12T20:59:09
2024-11-12T20:58:58
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? multi gpu ollama run llama3.2-vision >>> The image is a book cover. Output should be in this format - <Name of the Book>: <Name of the Author>. Do not output anything else /media/root/ssd2t/data/pro/tmp/o ... l/new/FastChat/image.png Added image '/media/root/ssd2t/data/pro/tmp/ol/new/FastCha...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7619/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7619/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/748
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/748/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/748/comments
https://api.github.com/repos/ollama/ollama/issues/748/events
https://github.com/ollama/ollama/issues/748
1,935,466,908
I_kwDOJ0Z1Ps5zXOGc
748
Enabling langchain
{ "login": "suoko", "id": 3659980, "node_id": "MDQ6VXNlcjM2NTk5ODA=", "avatar_url": "https://avatars.githubusercontent.com/u/3659980?v=4", "gravatar_id": "", "url": "https://api.github.com/users/suoko", "html_url": "https://github.com/suoko", "followers_url": "https://api.github.com/users/suoko/follower...
[]
closed
false
null
[]
null
2
2023-10-10T14:13:06
2023-10-11T15:45:32
2023-10-11T15:45:32
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I found this video about implementing langchain with ollama. If available it should be easy to use it via ollama-ui https://www.youtube.com/watch?v=CPgp8MhmGVY
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/748/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/748/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2844
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2844/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2844/comments
https://api.github.com/repos/ollama/ollama/issues/2844/events
https://github.com/ollama/ollama/issues/2844
2,162,130,461
I_kwDOJ0Z1Ps6A334d
2,844
OpenAI package compatibility
{ "login": "eliranwong", "id": 25262722, "node_id": "MDQ6VXNlcjI1MjYyNzIy", "avatar_url": "https://avatars.githubusercontent.com/u/25262722?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eliranwong", "html_url": "https://github.com/eliranwong", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
6
2024-02-29T21:26:43
2024-05-02T01:23:23
2024-03-01T01:08:58
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I read the example in https://ollama.com/blog/openai-compatibility The example works, but it doesn't when I add "response_format={ "type": "json_object" }," https://platform.openai.com/docs/guides/text-generation/json-mode
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2844/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2844/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/285
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/285/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/285/comments
https://api.github.com/repos/ollama/ollama/issues/285/events
https://github.com/ollama/ollama/pull/285
1,837,254,131
PR_kwDOJ0Z1Ps5XOXrT
285
Update README.md
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
[]
closed
false
null
[]
null
0
2023-08-04T19:43:08
2023-08-04T19:45:56
2023-08-04T19:45:55
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/285", "html_url": "https://github.com/ollama/ollama/pull/285", "diff_url": "https://github.com/ollama/ollama/pull/285.diff", "patch_url": "https://github.com/ollama/ollama/pull/285.patch", "merged_at": "2023-08-04T19:45:55" }
Ollama now supports Intel Macs
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/us...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/285/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/285/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2392
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2392/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2392/comments
https://api.github.com/repos/ollama/ollama/issues/2392/events
https://github.com/ollama/ollama/issues/2392
2,123,591,240
I_kwDOJ0Z1Ps5-k25I
2,392
unable to initialize llm library Radeon card detected
{ "login": "sigmaya", "id": 159318222, "node_id": "U_kgDOCX8Azg", "avatar_url": "https://avatars.githubusercontent.com/u/159318222?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sigmaya", "html_url": "https://github.com/sigmaya", "followers_url": "https://api.github.com/users/sigmaya/foll...
[]
closed
false
null
[]
null
6
2024-02-07T18:04:01
2024-02-29T00:50:07
2024-02-07T22:47:19
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hello, I am trying to run as user and manually, i get this error: time=2024-02-07T19:00:18.967+01:00 level=INFO source=payload_common.go:106 msg="Extracting dynamic libraries..." Error: unable to initialize llm library Radeon card detected, but permissions not set up properly. Either run ollama as root, or add you ...
{ "login": "sigmaya", "id": 159318222, "node_id": "U_kgDOCX8Azg", "avatar_url": "https://avatars.githubusercontent.com/u/159318222?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sigmaya", "html_url": "https://github.com/sigmaya", "followers_url": "https://api.github.com/users/sigmaya/foll...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2392/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2392/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5488
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5488/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5488/comments
https://api.github.com/repos/ollama/ollama/issues/5488/events
https://github.com/ollama/ollama/issues/5488
2,391,347,345
I_kwDOJ0Z1Ps6OiRCR
5,488
falcon2 fails to genarate code in almalinux9
{ "login": "olumolu", "id": 162728301, "node_id": "U_kgDOCbMJbQ", "avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4", "gravatar_id": "", "url": "https://api.github.com/users/olumolu", "html_url": "https://github.com/olumolu", "followers_url": "https://api.github.com/users/olumolu/foll...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
0
2024-07-04T18:55:46
2024-09-26T12:39:47
2024-09-26T12:39:47
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I got a error that it could not able to genarate any code but it failed and just skipped and jumped 5lines of free space and ask for input again. ### OS Linux ### GPU _No response_ ### CPU Intel ### Ollama version 1.48
{ "login": "olumolu", "id": 162728301, "node_id": "U_kgDOCbMJbQ", "avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4", "gravatar_id": "", "url": "https://api.github.com/users/olumolu", "html_url": "https://github.com/olumolu", "followers_url": "https://api.github.com/users/olumolu/foll...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5488/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5488/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3287
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3287/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3287/comments
https://api.github.com/repos/ollama/ollama/issues/3287/events
https://github.com/ollama/ollama/issues/3287
2,200,586,023
I_kwDOJ0Z1Ps6DKkcn
3,287
Allow system message to be on its own
{ "login": "jackielii", "id": 360983, "node_id": "MDQ6VXNlcjM2MDk4Mw==", "avatar_url": "https://avatars.githubusercontent.com/u/360983?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jackielii", "html_url": "https://github.com/jackielii", "followers_url": "https://api.github.com/users/jack...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
[ { "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api...
null
3
2024-03-21T16:05:18
2024-09-04T04:42:37
2024-09-04T04:42:37
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What are you trying to do? Current prompt rendering assumes there is only one system message. With the [Command-R model from Cohere AI](https://huggingface.co/CohereForAI/c4ai-command-r-v01), there can be multiple system message: https://docs.cohere.com/docs/prompting-command-r#structured-prompt-templates. There i...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3287/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3287/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3711
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3711/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3711/comments
https://api.github.com/repos/ollama/ollama/issues/3711/events
https://github.com/ollama/ollama/issues/3711
2,249,202,461
I_kwDOJ0Z1Ps6GEBsd
3,711
CUDA malloc fails on newly supported models in 0.1.32 (dual-GPU setup with 72GB VRAM and 128GB RAM)
{ "login": "mz2", "id": 71363, "node_id": "MDQ6VXNlcjcxMzYz", "avatar_url": "https://avatars.githubusercontent.com/u/71363?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mz2", "html_url": "https://github.com/mz2", "followers_url": "https://api.github.com/users/mz2/followers", "following...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
16
2024-04-17T20:18:35
2024-06-22T00:10:31
2024-06-22T00:10:31
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I am getting cuda malloc errors with v0.1.32 (as well as with the current head of main branch) when trying any of the new big models: wizardlm2, mixtral:8x22b, dbrx (command-r+ does work) with my dual GPU setup (A6000 + RTX 3090, i.e. combined 72GB VRAM) with a 24-core 13th gen Intel CPU (128GB ...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3711/reactions", "total_count": 8, "+1": 8, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3711/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3779
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3779/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3779/comments
https://api.github.com/repos/ollama/ollama/issues/3779/events
https://github.com/ollama/ollama/issues/3779
2,254,549,576
I_kwDOJ0Z1Ps6GYbJI
3,779
How to check memory utilization rate
{ "login": "taozhiyuai", "id": 146583103, "node_id": "U_kgDOCLyuPw", "avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4", "gravatar_id": "", "url": "https://api.github.com/users/taozhiyuai", "html_url": "https://github.com/taozhiyuai", "followers_url": "https://api.github.com/users/tao...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-04-20T13:19:09
2024-05-02T10:55:11
2024-05-02T10:55:11
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
after loading a model,How to check v-memory utilization rate
{ "login": "taozhiyuai", "id": 146583103, "node_id": "U_kgDOCLyuPw", "avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4", "gravatar_id": "", "url": "https://api.github.com/users/taozhiyuai", "html_url": "https://github.com/taozhiyuai", "followers_url": "https://api.github.com/users/tao...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3779/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3779/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/431
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/431/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/431/comments
https://api.github.com/repos/ollama/ollama/issues/431/events
https://github.com/ollama/ollama/issues/431
1,868,486,927
I_kwDOJ0Z1Ps5vXtkP
431
UTF-8 characters
{ "login": "arpecop", "id": 168974, "node_id": "MDQ6VXNlcjE2ODk3NA==", "avatar_url": "https://avatars.githubusercontent.com/u/168974?v=4", "gravatar_id": "", "url": "https://api.github.com/users/arpecop", "html_url": "https://github.com/arpecop", "followers_url": "https://api.github.com/users/arpecop/fo...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA...
closed
false
null
[]
null
4
2023-08-27T12:56:34
2023-08-30T21:11:50
2023-08-30T21:11:50
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
How to render properly such characters , is this per-model issue ... or font related ? ![Screenshot 2023-08-27 at 15 54 34](https://github.com/jmorganca/ollama/assets/168974/6cc84d4a-f438-467f-bf7c-207ec45ee6fe)
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/us...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/431/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/431/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1546
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1546/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1546/comments
https://api.github.com/repos/ollama/ollama/issues/1546/events
https://github.com/ollama/ollama/issues/1546
2,044,086,055
I_kwDOJ0Z1Ps551kcn
1,546
Is running behind a proxy fully supported?
{ "login": "pardeike", "id": 853584, "node_id": "MDQ6VXNlcjg1MzU4NA==", "avatar_url": "https://avatars.githubusercontent.com/u/853584?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pardeike", "html_url": "https://github.com/pardeike", "followers_url": "https://api.github.com/users/pardeik...
[]
closed
false
null
[]
null
4
2023-12-15T17:16:41
2023-12-28T12:17:29
2023-12-15T17:24:09
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi, when I run with a configured proxy, I get the error `lookup registry.ollama.ai: no such host`: ``` > HTTPS_PROXY=http://REDACTED.XXX:10000 ollama run llama2 pulling manifest Error: pull model manifest: Get "https://registry.ollama.ai/v2/library/llama2/manifests/latest": dial tcp: lookup registry.ollama.ai: ...
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1546/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1546/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1151
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1151/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1151/comments
https://api.github.com/repos/ollama/ollama/issues/1151/events
https://github.com/ollama/ollama/pull/1151
1,996,326,933
PR_kwDOJ0Z1Ps5fmfF0
1,151
install: fix enable contrib on debian 12
{ "login": "yanndegat", "id": 17246629, "node_id": "MDQ6VXNlcjE3MjQ2NjI5", "avatar_url": "https://avatars.githubusercontent.com/u/17246629?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yanndegat", "html_url": "https://github.com/yanndegat", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
1
2023-11-16T08:23:44
2023-11-16T20:53:07
2023-11-16T20:53:07
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1151", "html_url": "https://github.com/ollama/ollama/pull/1151", "diff_url": "https://github.com/ollama/ollama/pull/1151.diff", "patch_url": "https://github.com/ollama/ollama/pull/1151.patch", "merged_at": "2023-11-16T20:53:07" }
On debian 12, sources definitions have moved from /etc/apt/sources.list to /etc/apt/sources.list.d/debian.sources
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1151/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1151/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7571
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7571/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7571/comments
https://api.github.com/repos/ollama/ollama/issues/7571/events
https://github.com/ollama/ollama/issues/7571
2,643,262,341
I_kwDOJ0Z1Ps6djPuF
7,571
Please add Qwen-VL Qwen2.5-VL Qwen2-VL multimodal large model
{ "login": "smileyboy2019", "id": 59221294, "node_id": "MDQ6VXNlcjU5MjIxMjk0", "avatar_url": "https://avatars.githubusercontent.com/u/59221294?v=4", "gravatar_id": "", "url": "https://api.github.com/users/smileyboy2019", "html_url": "https://github.com/smileyboy2019", "followers_url": "https://api.githu...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
2
2024-11-08T08:19:59
2024-11-12T00:40:52
2024-11-12T00:40:52
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7571/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7571/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4353
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4353/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4353/comments
https://api.github.com/repos/ollama/ollama/issues/4353/events
https://github.com/ollama/ollama/issues/4353
2,290,802,195
I_kwDOJ0Z1Ps6Iit4T
4,353
Unable to determine if Ollama can be run on GPU after running Linux commands
{ "login": "jerry123cao", "id": 124961434, "node_id": "U_kgDOB3LCmg", "avatar_url": "https://avatars.githubusercontent.com/u/124961434?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jerry123cao", "html_url": "https://github.com/jerry123cao", "followers_url": "https://api.github.com/users/...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
1
2024-05-11T09:18:56
2024-05-21T23:21:51
2024-05-21T23:21:48
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hello, I have a question. If I use the command curl - fsSL https://ollama.com/install.sh |Can Ollama run on GPU after installing the Nvidia driver and cuda toolkit in SH?
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4353/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4353/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5292
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5292/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5292/comments
https://api.github.com/repos/ollama/ollama/issues/5292/events
https://github.com/ollama/ollama/issues/5292
2,374,357,161
I_kwDOJ0Z1Ps6NhdCp
5,292
请上架InternVL
{ "login": "enryteam", "id": 20081090, "node_id": "MDQ6VXNlcjIwMDgxMDkw", "avatar_url": "https://avatars.githubusercontent.com/u/20081090?v=4", "gravatar_id": "", "url": "https://api.github.com/users/enryteam", "html_url": "https://github.com/enryteam", "followers_url": "https://api.github.com/users/enr...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
1
2024-06-26T05:58:42
2025-01-28T13:33:00
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
ERROR: type should be string, got "\r\nhttps://github.com/OpenGVLab/InternVL\r\n\r\nthanks 谢谢\r\n\r\n\r\n\r\nollama0.1.43\r\nerror format not yet support!\r\n错误格式尚不支持!\r\n错误格式尚不支持!\r\n搞了多次 均报错"
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5292/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5292/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/367
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/367/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/367/comments
https://api.github.com/repos/ollama/ollama/issues/367/events
https://github.com/ollama/ollama/pull/367
1,854,193,882
PR_kwDOJ0Z1Ps5YHWrd
367
fix nous-hermes model file size listing in readme
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/us...
[]
closed
false
null
[]
null
0
2023-08-17T03:41:42
2023-08-17T03:42:01
2023-08-17T03:42:00
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/367", "html_url": "https://github.com/ollama/ollama/pull/367", "diff_url": "https://github.com/ollama/ollama/pull/367.diff", "patch_url": "https://github.com/ollama/ollama/pull/367.patch", "merged_at": "2023-08-17T03:42:00" }
fix nous-hermes model file size listing in readme
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/us...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/367/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/367/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6661
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6661/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6661/comments
https://api.github.com/repos/ollama/ollama/issues/6661/events
https://github.com/ollama/ollama/pull/6661
2,508,691,339
PR_kwDOJ0Z1Ps56kwXL
6,661
commit
{ "login": "rpreslar4765", "id": 89657947, "node_id": "MDQ6VXNlcjg5NjU3OTQ3", "avatar_url": "https://avatars.githubusercontent.com/u/89657947?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rpreslar4765", "html_url": "https://github.com/rpreslar4765", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
1
2024-09-05T20:53:42
2024-09-05T21:16:01
2024-09-05T21:16:01
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6661", "html_url": "https://github.com/ollama/ollama/pull/6661", "diff_url": "https://github.com/ollama/ollama/pull/6661.diff", "patch_url": "https://github.com/ollama/ollama/pull/6661.patch", "merged_at": null }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6661/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6661/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8345
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8345/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8345/comments
https://api.github.com/repos/ollama/ollama/issues/8345/events
https://github.com/ollama/ollama/issues/8345
2,775,164,105
I_kwDOJ0Z1Ps6laaTJ
8,345
llama3.2-vision does not support tools
{ "login": "kwaa", "id": 50108258, "node_id": "MDQ6VXNlcjUwMTA4MjU4", "avatar_url": "https://avatars.githubusercontent.com/u/50108258?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kwaa", "html_url": "https://github.com/kwaa", "followers_url": "https://api.github.com/users/kwaa/followers"...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/...
[ { "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "htt...
null
17
2025-01-08T11:34:22
2025-01-18T09:30:37
2025-01-15T21:25:43
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I get above error when I use `llama3.2-vision:11b` with tool calling. I think `llama3.2-vision:11b` should support tool calls: https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/vision_prompt_format.md#builtin-and-zero-shot-tool-calling - [llama3.2:3b template](https://ol...
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8345/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8345/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1421
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1421/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1421/comments
https://api.github.com/repos/ollama/ollama/issues/1421/events
https://github.com/ollama/ollama/pull/1421
2,031,577,402
PR_kwDOJ0Z1Ps5hd4Rw
1,421
fix redundant newline
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
0
2023-12-07T21:44:54
2023-12-07T21:47:24
2023-12-07T21:47:23
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1421", "html_url": "https://github.com/ollama/ollama/pull/1421", "diff_url": "https://github.com/ollama/ollama/pull/1421.diff", "patch_url": "https://github.com/ollama/ollama/pull/1421.patch", "merged_at": "2023-12-07T21:47:23" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1421/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1421/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8020
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8020/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8020/comments
https://api.github.com/repos/ollama/ollama/issues/8020/events
https://github.com/ollama/ollama/issues/8020
2,728,619,855
I_kwDOJ0Z1Ps6io29P
8,020
Nancy finds security vulnerabilities
{ "login": "mitar", "id": 585279, "node_id": "MDQ6VXNlcjU4NTI3OQ==", "avatar_url": "https://avatars.githubusercontent.com/u/585279?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mitar", "html_url": "https://github.com/mitar", "followers_url": "https://api.github.com/users/mitar/followers"...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
0
2024-12-10T00:29:13
2024-12-10T00:29:13
null
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? It finds: * [CVE-2024-8063](https://ossindex.sonatype.org/vulnerability/CVE-2024-8063?component-type=golang&component-name=github.com%2Follama%2Follama&utm_source=nancy-client&utm_medium=integration&utm_content=1.0.46) CWE-369: Divide By Zero * [CVE-2024-39719](https://ossindex.sonatype.org/vu...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8020/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8020/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2467
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2467/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2467/comments
https://api.github.com/repos/ollama/ollama/issues/2467/events
https://github.com/ollama/ollama/pull/2467
2,130,595,963
PR_kwDOJ0Z1Ps5mqQ3_
2,467
Check image filetype in api handlers
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[]
closed
false
null
[]
null
0
2024-02-12T17:17:19
2024-02-12T19:16:21
2024-02-12T19:16:20
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2467", "html_url": "https://github.com/ollama/ollama/pull/2467", "diff_url": "https://github.com/ollama/ollama/pull/2467.diff", "patch_url": "https://github.com/ollama/ollama/pull/2467.patch", "merged_at": "2024-02-12T19:16:20" }
Fixes: https://github.com/ollama/ollama/issues/2456
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2467/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2467/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1067
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1067/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1067/comments
https://api.github.com/repos/ollama/ollama/issues/1067/events
https://github.com/ollama/ollama/pull/1067
1,986,694,006
PR_kwDOJ0Z1Ps5fF3bO
1,067
add custom ollama-runner
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
[]
closed
false
null
[]
null
1
2023-11-10T01:45:17
2023-12-06T23:54:38
2023-11-21T20:14:50
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1067", "html_url": "https://github.com/ollama/ollama/pull/1067", "diff_url": "https://github.com/ollama/ollama/pull/1067.diff", "patch_url": "https://github.com/ollama/ollama/pull/1067.patch", "merged_at": null }
- update llama.cpp examples with custom ollama-runner - update llama.cpp gguf version to latest This change adds a custom inference server to llama.cpp based on the server we use in the current version, but with excess features removed. This allows us to have a more stable interface to build on when llama.cpp updat...
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1067/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1067/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1420
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1420/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1420/comments
https://api.github.com/repos/ollama/ollama/issues/1420/events
https://github.com/ollama/ollama/pull/1420
2,031,437,051
PR_kwDOJ0Z1Ps5hdZct
1,420
os specific ctrl-z
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
[]
closed
false
null
[]
null
3
2023-12-07T20:03:08
2023-12-14T06:46:26
2023-12-11T15:48:15
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1420", "html_url": "https://github.com/ollama/ollama/pull/1420", "diff_url": "https://github.com/ollama/ollama/pull/1420.diff", "patch_url": "https://github.com/ollama/ollama/pull/1420.patch", "merged_at": "2023-12-11T15:48:15" }
Add OS specific readline functions. Windows does not support these suspend system calls, so make ctrl-z a no-op on windows. This fixes development windows native builds. resolves #1414
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1420/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1420/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7824
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7824/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7824/comments
https://api.github.com/repos/ollama/ollama/issues/7824/events
https://github.com/ollama/ollama/issues/7824
2,688,581,053
I_kwDOJ0Z1Ps6gQH29
7,824
Tool calling parsing for llama3.2
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/...
[ { "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "htt...
null
0
2024-11-24T23:12:47
2024-11-24T23:12:47
null
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Llama 3.2 tool call outputs [are not in JSON](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/text_prompt_format.md) and so Ollama's tool parsing needs to be updated ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7824/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7824/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2586
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2586/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2586/comments
https://api.github.com/repos/ollama/ollama/issues/2586/events
https://github.com/ollama/ollama/issues/2586
2,141,379,273
I_kwDOJ0Z1Ps5_otrJ
2,586
Question on model updating
{ "login": "owenzhao", "id": 2182896, "node_id": "MDQ6VXNlcjIxODI4OTY=", "avatar_url": "https://avatars.githubusercontent.com/u/2182896?v=4", "gravatar_id": "", "url": "https://api.github.com/users/owenzhao", "html_url": "https://github.com/owenzhao", "followers_url": "https://api.github.com/users/owenz...
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
null
[]
null
2
2024-02-19T02:42:14
2024-03-11T20:45:20
2024-03-11T20:45:11
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I know I could download the latest model when a model is not installed. Say: ```swift ollama run llama2 ``` After some days, say one month, the model will be updated. Can ollama updates to the latest model automatically? I didn't find a way to do that. Please help.
{ "login": "hoyyeva", "id": 63033505, "node_id": "MDQ6VXNlcjYzMDMzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoyyeva", "html_url": "https://github.com/hoyyeva", "followers_url": "https://api.github.com/users/hoyyev...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2586/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2586/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2247
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2247/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2247/comments
https://api.github.com/repos/ollama/ollama/issues/2247/events
https://github.com/ollama/ollama/pull/2247
2,104,479,078
PR_kwDOJ0Z1Ps5lRYk5
2,247
Update import instructions to use convert and quantize tooling from llama.cpp submodule
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[]
closed
false
null
[]
null
0
2024-01-29T01:45:12
2024-02-05T05:50:45
2024-02-05T05:50:44
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2247", "html_url": "https://github.com/ollama/ollama/pull/2247", "diff_url": "https://github.com/ollama/ollama/pull/2247.diff", "patch_url": "https://github.com/ollama/ollama/pull/2247.patch", "merged_at": "2024-02-05T05:50:44" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2247/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2247/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5994
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5994/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5994/comments
https://api.github.com/repos/ollama/ollama/issues/5994/events
https://github.com/ollama/ollama/pull/5994
2,432,953,146
PR_kwDOJ0Z1Ps52nmhq
5,994
server: fix race conditions during download
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers"...
[]
closed
false
null
[]
null
0
2024-07-26T20:44:18
2024-07-26T21:24:25
2024-07-26T21:24:24
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5994", "html_url": "https://github.com/ollama/ollama/pull/5994", "diff_url": "https://github.com/ollama/ollama/pull/5994.diff", "patch_url": "https://github.com/ollama/ollama/pull/5994.patch", "merged_at": "2024-07-26T21:24:24" }
This fixes various data races scattered throughout the download/pull client where the client was accessing the download state concurrently. This commit is mostly a hot-fix and will be replaced by a new client one day soon. Also, remove the unnecessary opts argument from downloadChunk.
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers"...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5994/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 1, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5994/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3839
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3839/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3839/comments
https://api.github.com/repos/ollama/ollama/issues/3839/events
https://github.com/ollama/ollama/issues/3839
2,258,206,018
I_kwDOJ0Z1Ps6GmX1C
3,839
Feature Request: Detect Truncation Due to Exceeding Context Size
{ "login": "guoxf", "id": 3947566, "node_id": "MDQ6VXNlcjM5NDc1NjY=", "avatar_url": "https://avatars.githubusercontent.com/u/3947566?v=4", "gravatar_id": "", "url": "https://api.github.com/users/guoxf", "html_url": "https://github.com/guoxf", "followers_url": "https://api.github.com/users/guoxf/follower...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
2
2024-04-23T07:54:37
2024-07-01T17:15:46
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Understanding whether model output has been truncated due to exceeding context size is crucial for trusting the model to provide complete and accurate information. Here are some specific examples that illustrate why it's necessary to know if the output has been truncated: 1. **Question-Answering Systems**: When buildi...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3839/reactions", "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3839/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6044
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6044/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6044/comments
https://api.github.com/repos/ollama/ollama/issues/6044/events
https://github.com/ollama/ollama/issues/6044
2,435,056,414
I_kwDOJ0Z1Ps6RJAMe
6,044
Illegal istruction in ollama_llama_server runner
{ "login": "SnowyCoder", "id": 10945206, "node_id": "MDQ6VXNlcjEwOTQ1MjA2", "avatar_url": "https://avatars.githubusercontent.com/u/10945206?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SnowyCoder", "html_url": "https://github.com/SnowyCoder", "followers_url": "https://api.github.com/use...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-07-29T10:34:52
2024-07-29T10:46:24
2024-07-29T10:46:23
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I tried to run llama3 model with ollama. Reproduction (with my CPU: `AMD Ryzen 7 7735HS`) 1. Start server: `ollama serve` 2. Try to run llama3: `ollama run llama3 (The same error occurs with llama3.1) The server will execute a new runner with the following arguments: ```/tmp/ollama996131...
{ "login": "SnowyCoder", "id": 10945206, "node_id": "MDQ6VXNlcjEwOTQ1MjA2", "avatar_url": "https://avatars.githubusercontent.com/u/10945206?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SnowyCoder", "html_url": "https://github.com/SnowyCoder", "followers_url": "https://api.github.com/use...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6044/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6044/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2368
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2368/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2368/comments
https://api.github.com/repos/ollama/ollama/issues/2368/events
https://github.com/ollama/ollama/pull/2368
2,119,989,649
PR_kwDOJ0Z1Ps5mGd7d
2,368
Replace `reflect` usage in option parsing
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[]
closed
false
null
[]
null
3
2024-02-06T05:25:58
2024-08-11T16:52:11
2024-08-11T16:52:11
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2368", "html_url": "https://github.com/ollama/ollama/pull/2368", "diff_url": "https://github.com/ollama/ollama/pull/2368.diff", "patch_url": "https://github.com/ollama/ollama/pull/2368.patch", "merged_at": null }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2368/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2368/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6929
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6929/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6929/comments
https://api.github.com/repos/ollama/ollama/issues/6929/events
https://github.com/ollama/ollama/issues/6929
2,544,478,139
I_kwDOJ0Z1Ps6Xqae7
6,929
support qwen2.5:72b-instruct function call
{ "login": "chenfolingithub", "id": 9414696, "node_id": "MDQ6VXNlcjk0MTQ2OTY=", "avatar_url": "https://avatars.githubusercontent.com/u/9414696?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chenfolingithub", "html_url": "https://github.com/chenfolingithub", "followers_url": "https://api.g...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-09-24T06:36:16
2024-11-06T00:18:09
2024-11-06T00:18:09
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
qwen2.5:72b-instruct now support function call Returned its own unique data format, but currently ollama 0.3.11 does not correctly return formatted data for function calls
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6929/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6929/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4588
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4588/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4588/comments
https://api.github.com/repos/ollama/ollama/issues/4588/events
https://github.com/ollama/ollama/issues/4588
2,312,622,323
I_kwDOJ0Z1Ps6J19Dz
4,588
Tool calling feature support
{ "login": "Goekdeniz-Guelmez", "id": 60228478, "node_id": "MDQ6VXNlcjYwMjI4NDc4", "avatar_url": "https://avatars.githubusercontent.com/u/60228478?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Goekdeniz-Guelmez", "html_url": "https://github.com/Goekdeniz-Guelmez", "followers_url": "https...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-05-23T11:06:39
2024-05-23T21:39:55
2024-05-23T21:39:47
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
since there are more and more models that support function/tool calling. I think it would be a great feature for that support. It can be hard because the models could need Special tokens. My thought is that you can be able to set these spetial tokens or customizations in the Modelfile, something like: ```text PARAM...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4588/reactions", "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4588/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2424
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2424/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2424/comments
https://api.github.com/repos/ollama/ollama/issues/2424/events
https://github.com/ollama/ollama/issues/2424
2,126,622,691
I_kwDOJ0Z1Ps5-wa_j
2,424
Always getting a timeout error while querying using mistral using Ollama
{ "login": "Chakit22", "id": 118890138, "node_id": "U_kgDOBxYemg", "avatar_url": "https://avatars.githubusercontent.com/u/118890138?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Chakit22", "html_url": "https://github.com/Chakit22", "followers_url": "https://api.github.com/users/Chakit22/...
[]
closed
false
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers"...
[ { "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/...
null
19
2024-02-09T07:36:05
2025-01-16T09:16:45
2024-05-07T23:53:24
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
``` Traceback (most recent call last): File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpcore/_exceptions.py", line 10, in map_exceptions yield File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpcore/_backends/sync.py", line 126, in read return self._sock....
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2424/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2424/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1025
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1025/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1025/comments
https://api.github.com/repos/ollama/ollama/issues/1025/events
https://github.com/ollama/ollama/issues/1025
1,980,404,467
I_kwDOJ0Z1Ps52CpLz
1,025
How to stop a model aftering it's served?
{ "login": "paulwongx", "id": 20178251, "node_id": "MDQ6VXNlcjIwMTc4MjUx", "avatar_url": "https://avatars.githubusercontent.com/u/20178251?v=4", "gravatar_id": "", "url": "https://api.github.com/users/paulwongx", "html_url": "https://github.com/paulwongx", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
1
2023-11-07T02:12:01
2023-11-07T14:11:09
2023-11-07T14:11:09
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
When the modelling is already being served, what is the command to stop it so that I can serve it again with different variables? I want to run `OLLAMA_ORIGINS=https://XXXXX ollama serve` but I'm getting an error that the address is already in use. Running `ollama --help` doesn't show a command to unserve...? tld...
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1025/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1025/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4544
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4544/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4544/comments
https://api.github.com/repos/ollama/ollama/issues/4544/events
https://github.com/ollama/ollama/pull/4544
2,306,532,844
PR_kwDOJ0Z1Ps5v_RLM
4,544
update langchain-python-rag-document example to use latest langchain apis
{ "login": "pixelsoccupied", "id": 19938874, "node_id": "MDQ6VXNlcjE5OTM4ODc0", "avatar_url": "https://avatars.githubusercontent.com/u/19938874?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pixelsoccupied", "html_url": "https://github.com/pixelsoccupied", "followers_url": "https://api.gi...
[]
open
false
null
[]
null
0
2024-05-20T18:27:53
2024-10-29T07:15:48
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4544", "html_url": "https://github.com/ollama/ollama/pull/4544", "diff_url": "https://github.com/ollama/ollama/pull/4544.diff", "patch_url": "https://github.com/ollama/ollama/pull/4544.patch", "merged_at": null }
Hello! Langchain library went through a lot changes since this helpful example was first added and users will find that this will not work out of box anymore. I have updated this to work with all the latest changes and also removed any deprecated functions. Hope others find this useful!
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4544/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4544/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3174
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3174/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3174/comments
https://api.github.com/repos/ollama/ollama/issues/3174/events
https://github.com/ollama/ollama/pull/3174
2,189,661,685
PR_kwDOJ0Z1Ps5pzx1y
3,174
llama: remove server static assets
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[]
closed
false
null
[]
null
0
2024-03-16T00:42:03
2024-03-16T02:24:13
2024-03-16T02:24:12
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3174", "html_url": "https://github.com/ollama/ollama/pull/3174", "diff_url": "https://github.com/ollama/ollama/pull/3174.diff", "patch_url": "https://github.com/ollama/ollama/pull/3174.patch", "merged_at": "2024-03-16T02:24:12" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3174/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3174/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1439
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1439/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1439/comments
https://api.github.com/repos/ollama/ollama/issues/1439/events
https://github.com/ollama/ollama/issues/1439
2,033,191,401
I_kwDOJ0Z1Ps55MAnp
1,439
Serious bug
{ "login": "m-c-frank", "id": 61345033, "node_id": "MDQ6VXNlcjYxMzQ1MDMz", "avatar_url": "https://avatars.githubusercontent.com/u/61345033?v=4", "gravatar_id": "", "url": "https://api.github.com/users/m-c-frank", "html_url": "https://github.com/m-c-frank", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
0
2023-12-08T19:20:22
2023-12-08T19:23:58
2023-12-08T19:23:58
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
(base) mcfrank@interface:~$ orca show The following are not valid: show Warning: Could not load keyboard geometry for :0 BadName (named color or font does not exist) Resulting keymap file will not describe geometry ^C(base) mcfrank@interface:~$
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1439/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1439/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7411
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7411/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7411/comments
https://api.github.com/repos/ollama/ollama/issues/7411/events
https://github.com/ollama/ollama/pull/7411
2,622,487,493
PR_kwDOJ0Z1Ps6ATg6p
7,411
add more tests for getting the optimal tiled canvas
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
[]
closed
false
null
[]
null
0
2024-10-29T23:02:00
2024-10-29T23:28:04
2024-10-29T23:28:02
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7411", "html_url": "https://github.com/ollama/ollama/pull/7411", "diff_url": "https://github.com/ollama/ollama/pull/7411.diff", "patch_url": "https://github.com/ollama/ollama/pull/7411.patch", "merged_at": "2024-10-29T23:28:02" }
null
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7411/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7411/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/291
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/291/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/291/comments
https://api.github.com/repos/ollama/ollama/issues/291/events
https://github.com/ollama/ollama/issues/291
1,837,538,185
I_kwDOJ0Z1Ps5thpuJ
291
add `ollama serve` to the quickstart doc
{ "login": "marafa-sugarcrm", "id": 69474017, "node_id": "MDQ6VXNlcjY5NDc0MDE3", "avatar_url": "https://avatars.githubusercontent.com/u/69474017?v=4", "gravatar_id": "", "url": "https://api.github.com/users/marafa-sugarcrm", "html_url": "https://github.com/marafa-sugarcrm", "followers_url": "https://api...
[]
closed
false
null
[]
null
0
2023-08-05T02:51:59
2023-08-05T02:53:17
2023-08-05T02:53:17
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
i downloaded using `brew` and attempted to follow the quick start doc it is missing `ollama serve`
{ "login": "marafa-sugarcrm", "id": 69474017, "node_id": "MDQ6VXNlcjY5NDc0MDE3", "avatar_url": "https://avatars.githubusercontent.com/u/69474017?v=4", "gravatar_id": "", "url": "https://api.github.com/users/marafa-sugarcrm", "html_url": "https://github.com/marafa-sugarcrm", "followers_url": "https://api...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/291/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/291/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5881
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5881/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5881/comments
https://api.github.com/repos/ollama/ollama/issues/5881/events
https://github.com/ollama/ollama/issues/5881
2,425,609,012
I_kwDOJ0Z1Ps6Qk9s0
5,881
Is llama 3.1 already supported (on 2.8) or should we wait another update ?
{ "login": "Qualzz", "id": 35169816, "node_id": "MDQ6VXNlcjM1MTY5ODE2", "avatar_url": "https://avatars.githubusercontent.com/u/35169816?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Qualzz", "html_url": "https://github.com/Qualzz", "followers_url": "https://api.github.com/users/Qualzz/fo...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
20
2024-07-23T16:22:57
2024-09-02T00:15:39
2024-09-02T00:15:38
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? The model page seems to already exists in ollama website, but the model is clearly behaving erratically, which makes me wonder if we should wait for an update before using llama 3.1. ### OS Windows ### GPU Nvidia ### CPU AMD ### Ollama version 2.8
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5881/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5881/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1445
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1445/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1445/comments
https://api.github.com/repos/ollama/ollama/issues/1445/events
https://github.com/ollama/ollama/pull/1445
2,033,535,254
PR_kwDOJ0Z1Ps5hkiJ6
1,445
fix: parallel queueing race condition caused silent failure
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
[]
closed
false
null
[]
null
1
2023-12-09T01:58:05
2023-12-09T19:14:03
2023-12-09T19:14:02
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1445", "html_url": "https://github.com/ollama/ollama/pull/1445", "diff_url": "https://github.com/ollama/ollama/pull/1445.diff", "patch_url": "https://github.com/ollama/ollama/pull/1445.patch", "merged_at": "2023-12-09T19:14:02" }
As of the most recent llama.cpp update concurrent requests had a race condition that would result in an empty response. This was not easy to observe since the response from the llm runner subprocess was a 200 with the error {"content":"slot unavailable"} in the response stream, which just silently closed the channel...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1445/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1445/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3483
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3483/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3483/comments
https://api.github.com/repos/ollama/ollama/issues/3483/events
https://github.com/ollama/ollama/issues/3483
2,224,525,531
I_kwDOJ0Z1Ps6El5Db
3,483
Ollama hangs on CUDA devices when running multi-modal models
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-04-04T05:24:22
2024-04-04T06:30:58
2024-04-04T06:30:58
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? ``` Apr 04 05:15:04 gpu.us-central1-a.c.ollama.internal ollama[5042]: {"function":"launch_slot_with_data","level":"INFO","line":804,"msg":"slot is processing task","slot_id":0,"task_id":29930,"tid":"140079034640064","timestamp":1712207704} Apr 04 05:15:04 gpu.us-central1-a.c.ollama.internal ol...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3483/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3483/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/167
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/167/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/167/comments
https://api.github.com/repos/ollama/ollama/issues/167/events
https://github.com/ollama/ollama/pull/167
1,816,432,518
PR_kwDOJ0Z1Ps5WIaj5
167
partial decode ggml bin for more info
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
1
2023-07-21T22:10:33
2023-08-11T00:22:41
2023-08-11T00:22:40
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/167", "html_url": "https://github.com/ollama/ollama/pull/167", "diff_url": "https://github.com/ollama/ollama/pull/167.diff", "patch_url": "https://github.com/ollama/ollama/pull/167.patch", "merged_at": "2023-08-11T00:22:40" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/167/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/167/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/27
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/27/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/27/comments
https://api.github.com/repos/ollama/ollama/issues/27/events
https://github.com/ollama/ollama/issues/27
1,782,711,141
I_kwDOJ0Z1Ps5qQgNl
27
show a loading bar for model loading
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 5667396205, "node_id": ...
closed
false
null
[]
null
0
2023-06-30T15:17:59
2023-09-07T11:11:10
2023-09-07T11:11:10
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
currently we show a spinner – even better would be to show a loader similar to the download one
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/27/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/27/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4749
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4749/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4749/comments
https://api.github.com/repos/ollama/ollama/issues/4749/events
https://github.com/ollama/ollama/issues/4749
2,327,508,143
I_kwDOJ0Z1Ps6KuvSv
4,749
OLLAMA_MODELS not applied on initial start or on restart after upgrade on macOS
{ "login": "vernonstinebaker", "id": 8304954, "node_id": "MDQ6VXNlcjgzMDQ5NTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/8304954?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vernonstinebaker", "html_url": "https://github.com/vernonstinebaker", "followers_url": "https://ap...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
4
2024-05-31T10:32:39
2024-07-26T03:15:28
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
The addition of OLLAMA_MODELS is much appreciated, since it allows specifying a different location, such as an external disk, where more space might be available. One issue, however, is if we put specify OLLAMA_MODELS in our .zshrc, for example, the .zshrc file isn't read when Ollama starts initially or when Ollama ...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4749/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4749/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6280
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6280/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6280/comments
https://api.github.com/repos/ollama/ollama/issues/6280/events
https://github.com/ollama/ollama/issues/6280
2,457,405,573
I_kwDOJ0Z1Ps6SeQiF
6,280
Need qwen2:math !!
{ "login": "jsrdcht", "id": 22300065, "node_id": "MDQ6VXNlcjIyMzAwMDY1", "avatar_url": "https://avatars.githubusercontent.com/u/22300065?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jsrdcht", "html_url": "https://github.com/jsrdcht", "followers_url": "https://api.github.com/users/jsrdch...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-08-09T08:32:20
2024-09-02T23:12:26
2024-09-02T23:12:26
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
https://github.com/QwenLM/Qwen2-Math
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6280/reactions", "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6280/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8303
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8303/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8303/comments
https://api.github.com/repos/ollama/ollama/issues/8303/events
https://github.com/ollama/ollama/issues/8303
2,768,632,208
I_kwDOJ0Z1Ps6lBfmQ
8,303
Warning: could not connect to a running Ollama instance
{ "login": "rcjdut", "id": 140966542, "node_id": "U_kgDOCGb6jg", "avatar_url": "https://avatars.githubusercontent.com/u/140966542?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rcjdut", "html_url": "https://github.com/rcjdut", "followers_url": "https://api.github.com/users/rcjdut/follower...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2025-01-04T08:22:23
2025-01-04T08:40:22
2025-01-04T08:40:22
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I manually installed Ollama and successfully ran the command 'ollama serve' in a terminal,but when i try to ran the command 'ollama -v',it showed 'Warning: could not connect to a running Ollama instance Warning: client version is 0.5.4' ![屏幕截图 2025-01-04 162123](https://github.com/user-attachm...
{ "login": "rcjdut", "id": 140966542, "node_id": "U_kgDOCGb6jg", "avatar_url": "https://avatars.githubusercontent.com/u/140966542?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rcjdut", "html_url": "https://github.com/rcjdut", "followers_url": "https://api.github.com/users/rcjdut/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8303/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8303/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/692
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/692/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/692/comments
https://api.github.com/repos/ollama/ollama/issues/692/events
https://github.com/ollama/ollama/pull/692
1,924,891,665
PR_kwDOJ0Z1Ps5b1LE3
692
Increase client/server streaming buffer size to prevent `token too long` error
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
[]
closed
false
null
[]
null
0
2023-10-03T20:28:42
2023-10-04T18:09:01
2023-10-04T18:09:00
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/692", "html_url": "https://github.com/ollama/ollama/pull/692", "diff_url": "https://github.com/ollama/ollama/pull/692.diff", "patch_url": "https://github.com/ollama/ollama/pull/692.patch", "merged_at": "2023-10-04T18:09:00" }
In the case of a large input the response from `/generate` would be very long due to the encoded context length. Increase the buffer size to prevent this error. resolves #687
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/692/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/692/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1224
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1224/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1224/comments
https://api.github.com/repos/ollama/ollama/issues/1224/events
https://github.com/ollama/ollama/pull/1224
2,004,834,250
PR_kwDOJ0Z1Ps5gDPzg
1,224
update llama.cpp
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
1
2023-11-21T17:50:16
2023-11-21T20:22:00
2023-11-21T20:21:59
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1224", "html_url": "https://github.com/ollama/ollama/pull/1224", "diff_url": "https://github.com/ollama/ollama/pull/1224.diff", "patch_url": "https://github.com/ollama/ollama/pull/1224.patch", "merged_at": "2023-11-21T20:21:59" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1224/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1224/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5129
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5129/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5129/comments
https://api.github.com/repos/ollama/ollama/issues/5129/events
https://github.com/ollama/ollama/issues/5129
2,361,138,517
I_kwDOJ0Z1Ps6MvB1V
5,129
gpu destroys TMPDIR
{ "login": "vt-alt", "id": 36664211, "node_id": "MDQ6VXNlcjM2NjY0MjEx", "avatar_url": "https://avatars.githubusercontent.com/u/36664211?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vt-alt", "html_url": "https://github.com/vt-alt", "followers_url": "https://api.github.com/users/vt-alt/fo...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-06-19T02:53:24
2024-06-25T07:16:32
2024-06-20T17:41:00
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? When building ollama package for ALT Linux I noticed that %buildroot (directory where new binaries are installed) is disappeared after `go test github.com/ollama/ollama/gpu`. Our `%buildroot` (`/usr/src/tmp/ollama-buildroot`) is inside of `TMPDIR` (`/usr/src/tmp`). I am not completely investi...
{ "login": "joshyan1", "id": 76125168, "node_id": "MDQ6VXNlcjc2MTI1MTY4", "avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joshyan1", "html_url": "https://github.com/joshyan1", "followers_url": "https://api.github.com/users/jos...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5129/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5129/timeline
null
completed
false