url stringlengths 51 54 | repository_url stringclasses 1
value | labels_url stringlengths 65 68 | comments_url stringlengths 60 63 | events_url stringlengths 58 61 | html_url stringlengths 39 44 | id int64 1.78B 2.82B | node_id stringlengths 18 19 | number int64 1 8.69k | title stringlengths 1 382 | user dict | labels listlengths 0 5 | state stringclasses 2
values | locked bool 1
class | assignee dict | assignees listlengths 0 2 | milestone null | comments int64 0 323 | created_at timestamp[s] | updated_at timestamp[s] | closed_at timestamp[s] | author_association stringclasses 4
values | sub_issues_summary dict | active_lock_reason null | draft bool 2
classes | pull_request dict | body stringlengths 2 118k ⌀ | closed_by dict | reactions dict | timeline_url stringlengths 60 63 | performed_via_github_app null | state_reason stringclasses 4
values | is_pull_request bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/ollama/ollama/issues/5915 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5915/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5915/comments | https://api.github.com/repos/ollama/ollama/issues/5915/events | https://github.com/ollama/ollama/pull/5915 | 2,427,724,333 | PR_kwDOJ0Z1Ps52WWxZ | 5,915 | added tools when `stream=true` | {
"login": "vertrue",
"id": 30557724,
"node_id": "MDQ6VXNlcjMwNTU3NzI0",
"avatar_url": "https://avatars.githubusercontent.com/u/30557724?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/vertrue",
"html_url": "https://github.com/vertrue",
"followers_url": "https://api.github.com/users/vertru... | [] | closed | false | null | [] | null | 4 | 2024-07-24T14:18:44 | 2024-07-30T18:17:25 | 2024-07-30T18:17:25 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5915",
"html_url": "https://github.com/ollama/ollama/pull/5915",
"diff_url": "https://github.com/ollama/ollama/pull/5915.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5915.patch",
"merged_at": null
} | fixes #5796
```
curl --location '127.0.0.1:11434/v1/chat/completions' \
--header 'Content-Type: application/json' \
--data '{
"messages": [
{
"content": "You are a helpful AI assistant that can use tools.",
"role": "system"
},
{
"content": "What is function_1(10, 11)? use pr... | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjha... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5915/reactions",
"total_count": 18,
"+1": 9,
"-1": 0,
"laugh": 0,
"hooray": 5,
"confused": 0,
"heart": 0,
"rocket": 3,
"eyes": 1
} | https://api.github.com/repos/ollama/ollama/issues/5915/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7435 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7435/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7435/comments | https://api.github.com/repos/ollama/ollama/issues/7435/events | https://github.com/ollama/ollama/issues/7435 | 2,625,700,359 | I_kwDOJ0Z1Ps6cgQIH | 7,435 | 加载训练生成的gguf,/v1/chat/completions和/api/generate接口返回的结果不一样 | {
"login": "czhcc",
"id": 4754730,
"node_id": "MDQ6VXNlcjQ3NTQ3MzA=",
"avatar_url": "https://avatars.githubusercontent.com/u/4754730?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/czhcc",
"html_url": "https://github.com/czhcc",
"followers_url": "https://api.github.com/users/czhcc/follower... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2024-10-31T01:08:25 | 2024-11-17T14:17:43 | 2024-11-17T14:17:43 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
训练的基础模型是qwen2.5-7b。
对于同样的内容,
/v1/chat/completions和/api/generate接口返回不同的结果。
/api/generate返回是正确的训练要求的结果。
### OS
Linux, Docker
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.3.14 | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7435/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7435/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8260 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8260/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8260/comments | https://api.github.com/repos/ollama/ollama/issues/8260/events | https://github.com/ollama/ollama/issues/8260 | 2,761,465,899 | I_kwDOJ0Z1Ps6kmKAr | 8,260 | For CPU systems, getting files which are being used for GPU | {
"login": "Abubakkar13",
"id": 45032674,
"node_id": "MDQ6VXNlcjQ1MDMyNjc0",
"avatar_url": "https://avatars.githubusercontent.com/u/45032674?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Abubakkar13",
"html_url": "https://github.com/Abubakkar13",
"followers_url": "https://api.github.com/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2024-12-28T04:56:52 | 2024-12-29T03:14:47 | 2024-12-29T03:14:46 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Hey,
Happy holidays 😍!!!
I just came across with Ollama version update, then saw some files are being downloaded, I noticed that few dll libraries are being download, So i just cross checked the folder there are 7 files(Specifically starting with **cuda*** or **cublas*** ) and in runners ... | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8260/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8260/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/382 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/382/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/382/comments | https://api.github.com/repos/ollama/ollama/issues/382/events | https://github.com/ollama/ollama/pull/382 | 1,857,140,673 | PR_kwDOJ0Z1Ps5YRY9L | 382 | Closes #371 | {
"login": "jesjess243",
"id": 123990927,
"node_id": "U_kgDOB2Pzjw",
"avatar_url": "https://avatars.githubusercontent.com/u/123990927?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jesjess243",
"html_url": "https://github.com/jesjess243",
"followers_url": "https://api.github.com/users/jes... | [] | closed | false | null | [] | null | 1 | 2023-08-18T18:43:29 | 2023-08-22T19:21:00 | 2023-08-22T19:21:00 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/382",
"html_url": "https://github.com/ollama/ollama/pull/382",
"diff_url": "https://github.com/ollama/ollama/pull/382.diff",
"patch_url": "https://github.com/ollama/ollama/pull/382.patch",
"merged_at": null
} | Hello,
ModelPath.ParseModelPath(name) now trims https:// from name if it is present. Keeping in the scope of the issue, no alternative protocols are accepted and a path with no protocol defaults to https as before. My first approach would've allowed garbage text in the protocol, but this lacks that side-effect. | {
"login": "jesjess243",
"id": 123990927,
"node_id": "U_kgDOB2Pzjw",
"avatar_url": "https://avatars.githubusercontent.com/u/123990927?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jesjess243",
"html_url": "https://github.com/jesjess243",
"followers_url": "https://api.github.com/users/jes... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/382/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/382/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6473 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6473/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6473/comments | https://api.github.com/repos/ollama/ollama/issues/6473/events | https://github.com/ollama/ollama/issues/6473 | 2,482,749,930 | I_kwDOJ0Z1Ps6T-8Hq | 6,473 | OpenAI Structured Output Compatability | {
"login": "jd-solanki",
"id": 47495003,
"node_id": "MDQ6VXNlcjQ3NDk1MDAz",
"avatar_url": "https://avatars.githubusercontent.com/u/47495003?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jd-solanki",
"html_url": "https://github.com/jd-solanki",
"followers_url": "https://api.github.com/use... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 9 | 2024-08-23T09:42:33 | 2024-12-05T00:45:49 | 2024-12-05T00:45:49 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi 👋🏻
Loving ollama always ❤️
I'm eager to use newly released structured output using ollama but it looks like ollama doesn't have compatibility yet so I can just put base_url and I'll get response from my local LLM.
Also, I would like will it support [streaming](https://python.useinstructor.com/concepts/p... | {
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6473/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6473/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4292 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4292/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4292/comments | https://api.github.com/repos/ollama/ollama/issues/4292/events | https://github.com/ollama/ollama/pull/4292 | 2,288,040,474 | PR_kwDOJ0Z1Ps5vAnA9 | 4,292 | remove last bits of ParseModelPath | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | open | false | null | [] | null | 1 | 2024-05-09T16:25:46 | 2024-08-29T23:26:19 | null | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4292",
"html_url": "https://github.com/ollama/ollama/pull/4292",
"diff_url": "https://github.com/ollama/ollama/pull/4292.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4292.patch",
"merged_at": null
} | follow up to #4291 | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4292/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4292/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5455 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5455/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5455/comments | https://api.github.com/repos/ollama/ollama/issues/5455/events | https://github.com/ollama/ollama/issues/5455 | 2,388,109,167 | I_kwDOJ0Z1Ps6OV6dv | 5,455 | ollama does not work on ALL GPU automatically | {
"login": "HeroSong666",
"id": 142960235,
"node_id": "U_kgDOCIVmaw",
"avatar_url": "https://avatars.githubusercontent.com/u/142960235?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/HeroSong666",
"html_url": "https://github.com/HeroSong666",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 5 | 2024-07-03T09:12:14 | 2024-11-08T08:27:58 | 2024-08-01T23:51:38 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
when I use the ollama:0.1.38, I use the following command to start:
`
docker run -d --gpus=all -v /root/ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama
`
ollama will automatically use all 4 gpu cards for inference.
when I upgrate to ollama:0.1.48, I use the same comm... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5455/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5455/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6939 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6939/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6939/comments | https://api.github.com/repos/ollama/ollama/issues/6939/events | https://github.com/ollama/ollama/pull/6939 | 2,546,109,699 | PR_kwDOJ0Z1Ps58jqoL | 6,939 | docs: update llamaindex links | {
"login": "himself65",
"id": 14026360,
"node_id": "MDQ6VXNlcjE0MDI2MzYw",
"avatar_url": "https://avatars.githubusercontent.com/u/14026360?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/himself65",
"html_url": "https://github.com/himself65",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 0 | 2024-09-24T18:44:21 | 2024-09-24T19:15:59 | 2024-09-24T19:15:43 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/6939",
"html_url": "https://github.com/ollama/ollama/pull/6939",
"diff_url": "https://github.com/ollama/ollama/pull/6939.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6939.patch",
"merged_at": "2024-09-24T19:15:43"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6939/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6939/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1756 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1756/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1756/comments | https://api.github.com/repos/ollama/ollama/issues/1756/events | https://github.com/ollama/ollama/issues/1756 | 2,061,766,722 | I_kwDOJ0Z1Ps565BBC | 1,756 | Older CUDA compute capability 3.5 and 3.7 support | {
"login": "orlyandico",
"id": 1325420,
"node_id": "MDQ6VXNlcjEzMjU0MjA=",
"avatar_url": "https://avatars.githubusercontent.com/u/1325420?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/orlyandico",
"html_url": "https://github.com/orlyandico",
"followers_url": "https://api.github.com/users... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 6430601766,
"node_id": ... | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 64 | 2024-01-01T20:49:39 | 2025-01-27T15:35:54 | 2024-12-10T17:47:23 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I recently put together an (old) physical machine with an Nvidia K80, which is only supported up to CUDA 11.4 and Nvidia driver 470. All my previous experiments with Ollama were with more modern GPU's.
I found that Ollama doesn't use the GPU at all. I cannot find any documentation on the minimum required CUDA vers... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1756/reactions",
"total_count": 3,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 3,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1756/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8543 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8543/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8543/comments | https://api.github.com/repos/ollama/ollama/issues/8543/events | https://github.com/ollama/ollama/issues/8543 | 2,805,677,144 | I_kwDOJ0Z1Ps6nOzxY | 8,543 | Ollama does not perform structured output correctly. | {
"login": "aJarOfCode",
"id": 103819351,
"node_id": "U_kgDOBjAoVw",
"avatar_url": "https://avatars.githubusercontent.com/u/103819351?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/aJarOfCode",
"html_url": "https://github.com/aJarOfCode",
"followers_url": "https://api.github.com/users/aJa... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 1 | 2025-01-23T01:19:21 | 2025-01-27T09:10:56 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
request:
`{
"model":"llama3.2",
"messages":datas+[
{
"role":"user",
"content":input_data,
"images":[screenshot_base64]
}
],
"stream":False,
"for... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8543/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8543/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/7615 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7615/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7615/comments | https://api.github.com/repos/ollama/ollama/issues/7615/events | https://github.com/ollama/ollama/pull/7615 | 2,648,429,620 | PR_kwDOJ0Z1Ps6BeCEy | 7,615 | Allow Compile on older GPUs - still on CUDA 11.3 | {
"login": "langstonmeister",
"id": 65471211,
"node_id": "MDQ6VXNlcjY1NDcxMjEx",
"avatar_url": "https://avatars.githubusercontent.com/u/65471211?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/langstonmeister",
"html_url": "https://github.com/langstonmeister",
"followers_url": "https://api... | [] | open | false | null | [] | null | 1 | 2024-11-11T07:31:03 | 2024-11-22T17:01:26 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7615",
"html_url": "https://github.com/ollama/ollama/pull/7615",
"diff_url": "https://github.com/ollama/ollama/pull/7615.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7615.patch",
"merged_at": null
} | My Tesla k40s are still able to run this version of CUDA, but are technically on Compute 3.5. These commits will allow them to compile and run the software, and are working well for me. | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7615/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7615/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2331 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2331/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2331/comments | https://api.github.com/repos/ollama/ollama/issues/2331/events | https://github.com/ollama/ollama/pull/2331 | 2,115,267,354 | PR_kwDOJ0Z1Ps5l2Y_S | 2,331 | Update README.md | {
"login": "AshD",
"id": 233016,
"node_id": "MDQ6VXNlcjIzMzAxNg==",
"avatar_url": "https://avatars.githubusercontent.com/u/233016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/AshD",
"html_url": "https://github.com/AshD",
"followers_url": "https://api.github.com/users/AshD/followers",
... | [] | closed | false | null | [] | null | 4 | 2024-02-02T16:12:29 | 2024-11-21T09:13:03 | 2024-11-21T09:13:02 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2331",
"html_url": "https://github.com/ollama/ollama/pull/2331",
"diff_url": "https://github.com/ollama/ollama/pull/2331.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2331.patch",
"merged_at": null
} | Adding info on Fusion Quill. Setup info is here https://fusionquill.ai/help-setup-ollama/
Fusion Quill Personal Edition is a Windows app on the Microsoft App Store that connects to multiple AI models with workflows and UX like an Integrated Word processor with AI Chat in a split-pane UI that enables creating documen... | {
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/us... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2331/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2331/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4268 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4268/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4268/comments | https://api.github.com/repos/ollama/ollama/issues/4268/events | https://github.com/ollama/ollama/pull/4268 | 2,286,626,904 | PR_kwDOJ0Z1Ps5u72dc | 4,268 | Convert directly from llama3 | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | [] | closed | false | null | [] | null | 1 | 2024-05-08T23:14:20 | 2024-05-21T21:43:38 | 2024-05-21T21:43:37 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4268",
"html_url": "https://github.com/ollama/ollama/pull/4268",
"diff_url": "https://github.com/ollama/ollama/pull/4268.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4268.patch",
"merged_at": "2024-05-21T21:43:37"
} | This change allows you to convert directly from a llama3 derived safetensors model into Ollama.
It is currently *missing*:
* pytorch *almost* works however the embeddings layer size is off by the eos/bos tokens
This *will* work with most llama3 derivatives if they are using safetensors including `dolphin-2.9-lla... | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4268/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4268/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1651 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1651/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1651/comments | https://api.github.com/repos/ollama/ollama/issues/1651/events | https://github.com/ollama/ollama/issues/1651 | 2,051,690,293 | I_kwDOJ0Z1Ps56Sk81 | 1,651 | GPU not in use? | {
"login": "Y2K350",
"id": 140640993,
"node_id": "U_kgDOCGIC4Q",
"avatar_url": "https://avatars.githubusercontent.com/u/140640993?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Y2K350",
"html_url": "https://github.com/Y2K350",
"followers_url": "https://api.github.com/users/Y2K350/follower... | [] | closed | false | null | [] | null | 25 | 2023-12-21T04:51:03 | 2025-01-03T21:54:51 | 2024-01-10T15:08:43 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I am running Ollama which was installed on an arch linux system using "sudo pacman -S ollama" I am using a RTX 4090 with Nvidia's latest drivers. I also installed cuda using "sudo pacman -S cuda"
I run the LLM using the command "ollama run dolphin-mixtral:latest"
it does not appear to use the GPU based on GPU usa... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1651/reactions",
"total_count": 8,
"+1": 8,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1651/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1930 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1930/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1930/comments | https://api.github.com/repos/ollama/ollama/issues/1930/events | https://github.com/ollama/ollama/issues/1930 | 2,077,220,925 | I_kwDOJ0Z1Ps57z-A9 | 1,930 | Support for CogVLM wanted. CogVLM is an alternative for LLaVA | {
"login": "henryclw",
"id": 97804910,
"node_id": "U_kgDOBdRibg",
"avatar_url": "https://avatars.githubusercontent.com/u/97804910?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/henryclw",
"html_url": "https://github.com/henryclw",
"followers_url": "https://api.github.com/users/henryclw/fo... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | open | false | null | [] | null | 7 | 2024-01-11T17:23:22 | 2024-06-24T18:14:00 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Currently ollama is supporting LLaVA, which is super great.
I wonder is there a chance to load other similar models like CogVLM?
https://github.com/THUDM/CogVLM | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1930/reactions",
"total_count": 6,
"+1": 6,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1930/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/5525 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5525/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5525/comments | https://api.github.com/repos/ollama/ollama/issues/5525/events | https://github.com/ollama/ollama/issues/5525 | 2,393,903,563 | I_kwDOJ0Z1Ps6OsBHL | 5,525 | Suggestions | {
"login": "EchoOfMedivhCheats",
"id": 174923936,
"node_id": "U_kgDOCm0goA",
"avatar_url": "https://avatars.githubusercontent.com/u/174923936?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/EchoOfMedivhCheats",
"html_url": "https://github.com/EchoOfMedivhCheats",
"followers_url": "https://... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2024-07-07T05:43:01 | 2024-07-07T05:43:01 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Terminal color/bold codes (needs 3 and 3.%)
Custom terminal app (Vulkan for stuff like svgs? Maybe for videos?))
Selinux for English (nouns verbs pronouns)
Voice command and output
Don’t know that much assembly but I think you can daisy chain stuff
I know this is written in go, but a scripting language (if x = che... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5525/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5525/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/8690 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8690/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8690/comments | https://api.github.com/repos/ollama/ollama/issues/8690/events | https://github.com/ollama/ollama/issues/8690 | 2,820,660,880 | I_kwDOJ0Z1Ps6oH96Q | 8,690 | Deepseek-671B: Error: timed out waiting for llama runner to start - progress 0.00 on 8x L40S | {
"login": "orlyandico",
"id": 1325420,
"node_id": "MDQ6VXNlcjEzMjU0MjA=",
"avatar_url": "https://avatars.githubusercontent.com/u/1325420?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/orlyandico",
"html_url": "https://github.com/orlyandico",
"followers_url": "https://api.github.com/users... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 1 | 2025-01-30T12:08:46 | 2025-01-30T12:12:22 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Ollama (0.5.7) appears to be correctly calculating how many layers to offload to the GPU with default settings. This is on a g6e.48xlarge which has 1.5TB of RAM.
```
Jan 30 11:56:19 ip-172-31-21-180 ollama[3237]: time=2025-01-30T11:56:19.283Z level=INFO source=memory.go:356 msg="offload to cuda... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8690/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8690/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/106 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/106/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/106/comments | https://api.github.com/repos/ollama/ollama/issues/106/events | https://github.com/ollama/ollama/pull/106 | 1,810,840,013 | PR_kwDOJ0Z1Ps5V1VZH | 106 | README typo fix | {
"login": "isaac-mcfadyen",
"id": 6243993,
"node_id": "MDQ6VXNlcjYyNDM5OTM=",
"avatar_url": "https://avatars.githubusercontent.com/u/6243993?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/isaac-mcfadyen",
"html_url": "https://github.com/isaac-mcfadyen",
"followers_url": "https://api.gith... | [] | closed | false | null | [] | null | 1 | 2023-07-18T23:00:59 | 2023-07-18T23:24:58 | 2023-07-18T23:24:58 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/106",
"html_url": "https://github.com/ollama/ollama/pull/106",
"diff_url": "https://github.com/ollama/ollama/pull/106.diff",
"patch_url": "https://github.com/ollama/ollama/pull/106.patch",
"merged_at": "2023-07-18T23:24:58"
} | - Simple typo fix in README.md
- `13` changed to `13B` (the `B` was missing) on Nous-Hermes model in models table
- Also edited `hous-hermes` to `nous-hermes` (should be n instead of h) | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/106/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/106/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3936 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3936/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3936/comments | https://api.github.com/repos/ollama/ollama/issues/3936/events | https://github.com/ollama/ollama/pull/3936 | 2,265,316,091 | PR_kwDOJ0Z1Ps5t0VM1 | 3,936 | pulled model not found fix - added to FAQ | {
"login": "lambdaofgod",
"id": 3647577,
"node_id": "MDQ6VXNlcjM2NDc1Nzc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3647577?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lambdaofgod",
"html_url": "https://github.com/lambdaofgod",
"followers_url": "https://api.github.com/us... | [] | closed | false | null | [] | null | 5 | 2024-04-26T09:04:22 | 2024-07-03T17:00:00 | 2024-07-03T16:59:59 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3936",
"html_url": "https://github.com/ollama/ollama/pull/3936",
"diff_url": "https://github.com/ollama/ollama/pull/3936.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3936.patch",
"merged_at": null
} | I've added a fix by [sridvijay](https://github.com/sridvijay) from #3876 to the documentation. | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3936/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3936/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7864 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7864/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7864/comments | https://api.github.com/repos/ollama/ollama/issues/7864/events | https://github.com/ollama/ollama/issues/7864 | 2,698,937,631 | I_kwDOJ0Z1Ps6g3oUf | 7,864 | Please sync with llama.cpp for the update of bert_base like models. | {
"login": "BeNhNp",
"id": 33339730,
"node_id": "MDQ6VXNlcjMzMzM5NzMw",
"avatar_url": "https://avatars.githubusercontent.com/u/33339730?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BeNhNp",
"html_url": "https://github.com/BeNhNp",
"followers_url": "https://api.github.com/users/BeNhNp/fo... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 0 | 2024-11-27T15:34:28 | 2024-11-27T15:34:28 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
currently [ollama_llama_server](https://github.com/ollama/ollama/blob/main/llm/server.go#L894) can't return properly with google bert_base models, while llama.cpp have already supported, I've varified with `transformers`.
### OS
_No response_
### GPU
_No response_
### CPU
_No response_
##... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7864/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7864/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/3601 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3601/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3601/comments | https://api.github.com/repos/ollama/ollama/issues/3601/events | https://github.com/ollama/ollama/issues/3601 | 2,238,409,901 | I_kwDOJ0Z1Ps6Fa2yt | 3,601 | Docker 0.1.31 the 2nd Ollama cannot use its designated GPU | {
"login": "ww2283",
"id": 21956104,
"node_id": "MDQ6VXNlcjIxOTU2MTA0",
"avatar_url": "https://avatars.githubusercontent.com/u/21956104?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ww2283",
"html_url": "https://github.com/ww2283",
"followers_url": "https://api.github.com/users/ww2283/fo... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2024-04-11T19:26:22 | 2024-04-11T19:36:45 | 2024-04-11T19:36:45 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I'm on Ubuntu 20.04 with two Ada6000 cards. I use docker compose to host two instance of Ollama, each has its own model because I want to use them for AutoGen. For each Ollama container I want to assign an individual GPU.
In real use, I noticed that ollama1 is working perfectly fine, while oll... | {
"login": "ww2283",
"id": 21956104,
"node_id": "MDQ6VXNlcjIxOTU2MTA0",
"avatar_url": "https://avatars.githubusercontent.com/u/21956104?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ww2283",
"html_url": "https://github.com/ww2283",
"followers_url": "https://api.github.com/users/ww2283/fo... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3601/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3601/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6696 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6696/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6696/comments | https://api.github.com/repos/ollama/ollama/issues/6696/events | https://github.com/ollama/ollama/pull/6696 | 2,512,217,526 | PR_kwDOJ0Z1Ps56wsWl | 6,696 | Update README.md | {
"login": "rapidarchitect",
"id": 126218667,
"node_id": "U_kgDOB4Xxqw",
"avatar_url": "https://avatars.githubusercontent.com/u/126218667?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rapidarchitect",
"html_url": "https://github.com/rapidarchitect",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | 0 | 2024-09-08T06:08:35 | 2024-09-08T07:36:00 | 2024-09-08T07:36:00 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/6696",
"html_url": "https://github.com/ollama/ollama/pull/6696",
"diff_url": "https://github.com/ollama/ollama/pull/6696.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6696.patch",
"merged_at": "2024-09-08T07:36:00"
} | added crewai with mesop anew since there was a conflict previously | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6696/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6696/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6029 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6029/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6029/comments | https://api.github.com/repos/ollama/ollama/issues/6029/events | https://github.com/ollama/ollama/issues/6029 | 2,434,070,538 | I_kwDOJ0Z1Ps6RFPgK | 6,029 | Prompt evaluation progress indicator | {
"login": "drazdra",
"id": 133811709,
"node_id": "U_kgDOB_nN_Q",
"avatar_url": "https://avatars.githubusercontent.com/u/133811709?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/drazdra",
"html_url": "https://github.com/drazdra",
"followers_url": "https://api.github.com/users/drazdra/foll... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 1 | 2024-07-28T19:05:02 | 2024-08-14T05:23:01 | null | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Prompt evaluation can take huge time, especially with long context. it can literary be hours.
Right now Ollama just hangs in that phase. Is it possible to have some messages updating on the progress? It's tough when after an hour of 100% load you still don't know if it's at 15% or at 99%..
| null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6029/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6029/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/823 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/823/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/823/comments | https://api.github.com/repos/ollama/ollama/issues/823/events | https://github.com/ollama/ollama/issues/823 | 1,948,000,329 | I_kwDOJ0Z1Ps50HCBJ | 823 | failed to verify certificate: x509: certificate signed by unknown authority | {
"login": "sureshpatel66",
"id": 87146534,
"node_id": "MDQ6VXNlcjg3MTQ2NTM0",
"avatar_url": "https://avatars.githubusercontent.com/u/87146534?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sureshpatel66",
"html_url": "https://github.com/sureshpatel66",
"followers_url": "https://api.githu... | [] | closed | false | null | [] | null | 18 | 2023-10-17T18:21:33 | 2024-11-21T10:18:36 | 2023-10-25T19:10:43 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ~$ docker exec -it ollama ollama run llama2
pulling manifest
Error: pull model manifest: Get "https://registry.ollama.ai/v2/library/llama2/manifests/latest": tls: failed to verify certificate: x509: certificate signed by unknown authority
please guide to solve this issue | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/823/reactions",
"total_count": 7,
"+1": 7,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/823/timeline | null | not_planned | false |
https://api.github.com/repos/ollama/ollama/issues/7281 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7281/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7281/comments | https://api.github.com/repos/ollama/ollama/issues/7281/events | https://github.com/ollama/ollama/pull/7281 | 2,600,936,244 | PR_kwDOJ0Z1Ps5_PgOY | 7,281 | added formatting to ollama ls | {
"login": "robbiemu",
"id": 248927,
"node_id": "MDQ6VXNlcjI0ODkyNw==",
"avatar_url": "https://avatars.githubusercontent.com/u/248927?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/robbiemu",
"html_url": "https://github.com/robbiemu",
"followers_url": "https://api.github.com/users/robbiem... | [] | open | false | null | [] | null | 1 | 2024-10-20T22:44:50 | 2024-11-06T22:54:30 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7281",
"html_url": "https://github.com/ollama/ollama/pull/7281",
"diff_url": "https://github.com/ollama/ollama/pull/7281.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7281.patch",
"merged_at": null
} | This groups output by model on disk, and does not pollute the filesize in the listing when the manifests point to the same model.
```
./ollama ls
NAME ID SIZE MODIFIED
qwen2.5:32b-instruct-q4_K_M 9f13ba1299af 19 GB 9 hours ago
bge-m3... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7281/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7281/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1685 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1685/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1685/comments | https://api.github.com/repos/ollama/ollama/issues/1685/events | https://github.com/ollama/ollama/issues/1685 | 2,054,605,354 | I_kwDOJ0Z1Ps56dsoq | 1,685 | Ollama mobile app appeared on iOS app store | {
"login": "netsap",
"id": 41242395,
"node_id": "MDQ6VXNlcjQxMjQyMzk1",
"avatar_url": "https://avatars.githubusercontent.com/u/41242395?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/netsap",
"html_url": "https://github.com/netsap",
"followers_url": "https://api.github.com/users/netsap/fo... | [] | closed | false | null | [] | null | 5 | 2023-12-23T01:05:23 | 2024-05-10T00:27:44 | 2024-05-10T00:27:44 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Accessible here: https://apps.apple.com/us/app/ollama/id6471840114
I'm assuming this is not legitimate? | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1685/reactions",
"total_count": 2,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 2,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1685/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2043 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2043/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2043/comments | https://api.github.com/repos/ollama/ollama/issues/2043/events | https://github.com/ollama/ollama/pull/2043 | 2,087,600,434 | PR_kwDOJ0Z1Ps5kY2C0 | 2,043 | Dockerfile: use variables for package version | {
"login": "stevenbecht",
"id": 9442836,
"node_id": "MDQ6VXNlcjk0NDI4MzY=",
"avatar_url": "https://avatars.githubusercontent.com/u/9442836?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/stevenbecht",
"html_url": "https://github.com/stevenbecht",
"followers_url": "https://api.github.com/us... | [] | closed | false | null | [] | null | 2 | 2024-01-18T06:23:39 | 2024-02-21T00:26:49 | 2024-02-21T00:26:49 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2043",
"html_url": "https://github.com/ollama/ollama/pull/2043",
"diff_url": "https://github.com/ollama/ollama/pull/2043.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2043.patch",
"merged_at": null
} | Update Dockerfile to use variables instead of hardcoded values | {
"login": "stevenbecht",
"id": 9442836,
"node_id": "MDQ6VXNlcjk0NDI4MzY=",
"avatar_url": "https://avatars.githubusercontent.com/u/9442836?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/stevenbecht",
"html_url": "https://github.com/stevenbecht",
"followers_url": "https://api.github.com/us... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2043/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2043/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5285 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5285/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5285/comments | https://api.github.com/repos/ollama/ollama/issues/5285/events | https://github.com/ollama/ollama/pull/5285 | 2,373,797,637 | PR_kwDOJ0Z1Ps5zjnWI | 5,285 | OpenAI: /v1/embeddings compatibility | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjha... | [] | closed | false | null | [] | null | 1 | 2024-06-25T22:47:36 | 2024-07-16T20:36:11 | 2024-07-16T20:36:09 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5285",
"html_url": "https://github.com/ollama/ollama/pull/5285",
"diff_url": "https://github.com/ollama/ollama/pull/5285.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5285.patch",
"merged_at": "2024-07-16T20:36:08"
} | In anticipation of #5127
Resolves #2416
```
curl http://localhost:11434/v1/embeddings \
-H "Content-Type: application/json" \
-d '{
"input": ["hello", "hi"],
"model": "all-minilm"
}'
```
```
{
"object": "list",
"data": [
{
"object": "embedding",
"embedding": [
... | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjha... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5285/reactions",
"total_count": 3,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 2,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/ollama/ollama/issues/5285/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7969 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7969/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7969/comments | https://api.github.com/repos/ollama/ollama/issues/7969/events | https://github.com/ollama/ollama/issues/7969 | 2,723,362,146 | I_kwDOJ0Z1Ps6iUzVi | 7,969 | Administrative / silent install is borked | {
"login": "Lukas-UAUX",
"id": 43141094,
"node_id": "MDQ6VXNlcjQzMTQxMDk0",
"avatar_url": "https://avatars.githubusercontent.com/u/43141094?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Lukas-UAUX",
"html_url": "https://github.com/Lukas-UAUX",
"followers_url": "https://api.github.com/use... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 5860134234,
"node_id": ... | open | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 1 | 2024-12-06T15:50:06 | 2024-12-10T17:30:35 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
For deployment scenarios like for classrooms or to regular managed devices the setup needs to perform an administrative / unattended install. For InnoSetup built installers this can normally be done via the command-line switches /SILENT (or /VERYSILENT) or via a response file via /LOADINF. (See:... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7969/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7969/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/2811 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2811/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2811/comments | https://api.github.com/repos/ollama/ollama/issues/2811/events | https://github.com/ollama/ollama/issues/2811 | 2,159,235,157 | I_kwDOJ0Z1Ps6As1BV | 2,811 | Mistral Instruct models prompt does not use <s> or </s> | {
"login": "louisabraham",
"id": 13174805,
"node_id": "MDQ6VXNlcjEzMTc0ODA1",
"avatar_url": "https://avatars.githubusercontent.com/u/13174805?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/louisabraham",
"html_url": "https://github.com/louisabraham",
"followers_url": "https://api.github.c... | [] | closed | false | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api... | null | 5 | 2024-02-28T15:29:34 | 2024-07-18T22:45:02 | 2024-07-18T22:45:02 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | from https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1#instruction-format
> \<s> [INST] Instruction [/INST] Model answer\</s> [INST] Follow-up instruction [/INST]
I didn't see `<s>` as part of the prompt when using `OLLAMA_DEBUG=1 ollama start`. Is that expected? | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2811/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2811/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6611 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6611/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6611/comments | https://api.github.com/repos/ollama/ollama/issues/6611/events | https://github.com/ollama/ollama/pull/6611 | 2,503,463,117 | PR_kwDOJ0Z1Ps56S5Sr | 6,611 | Make stall duration timeout configurable | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 1 | 2024-09-03T18:04:25 | 2024-09-05T21:00:08 | 2024-09-05T21:00:08 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/6611",
"html_url": "https://github.com/ollama/ollama/pull/6611",
"diff_url": "https://github.com/ollama/ollama/pull/6611.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6611.patch",
"merged_at": "2024-09-05T21:00:08"
} | With the new very large parameter models, some users are willing to wait for a very long time for models to load.
Fixes #6031 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6611/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6611/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/329 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/329/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/329/comments | https://api.github.com/repos/ollama/ollama/issues/329/events | https://github.com/ollama/ollama/pull/329 | 1,846,179,798 | PR_kwDOJ0Z1Ps5XsTL- | 329 | Add tutorials for using Langchain with ollama | {
"login": "technovangelist",
"id": 633681,
"node_id": "MDQ6VXNlcjYzMzY4MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/technovangelist",
"html_url": "https://github.com/technovangelist",
"followers_url": "https://api.git... | [] | closed | false | null | [] | null | 1 | 2023-08-11T04:30:50 | 2024-01-05T18:40:54 | 2023-08-11T22:19:39 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/329",
"html_url": "https://github.com/ollama/ollama/pull/329",
"diff_url": "https://github.com/ollama/ollama/pull/329.diff",
"patch_url": "https://github.com/ollama/ollama/pull/329.patch",
"merged_at": "2023-08-11T22:19:39"
} | null | {
"login": "technovangelist",
"id": 633681,
"node_id": "MDQ6VXNlcjYzMzY4MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/technovangelist",
"html_url": "https://github.com/technovangelist",
"followers_url": "https://api.git... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/329/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/329/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7241 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7241/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7241/comments | https://api.github.com/repos/ollama/ollama/issues/7241/events | https://github.com/ollama/ollama/issues/7241 | 2,595,035,489 | I_kwDOJ0Z1Ps6arRlh | 7,241 | add module/ | {
"login": "malv-c",
"id": 19170213,
"node_id": "MDQ6VXNlcjE5MTcwMjEz",
"avatar_url": "https://avatars.githubusercontent.com/u/19170213?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/malv-c",
"html_url": "https://github.com/malv-c",
"followers_url": "https://api.github.com/users/malv-c/fo... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 1 | 2024-10-17T15:12:19 | 2024-10-20T06:06:27 | 2024-10-20T06:06:27 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | null | {
"login": "malv-c",
"id": 19170213,
"node_id": "MDQ6VXNlcjE5MTcwMjEz",
"avatar_url": "https://avatars.githubusercontent.com/u/19170213?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/malv-c",
"html_url": "https://github.com/malv-c",
"followers_url": "https://api.github.com/users/malv-c/fo... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7241/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7241/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2602 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2602/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2602/comments | https://api.github.com/repos/ollama/ollama/issues/2602/events | https://github.com/ollama/ollama/issues/2602 | 2,143,418,344 | I_kwDOJ0Z1Ps5_wffo | 2,602 | Windows preview - please let us set the location where models are stored | {
"login": "axefrog",
"id": 298883,
"node_id": "MDQ6VXNlcjI5ODg4Mw==",
"avatar_url": "https://avatars.githubusercontent.com/u/298883?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/axefrog",
"html_url": "https://github.com/axefrog",
"followers_url": "https://api.github.com/users/axefrog/fo... | [] | closed | false | null | [] | null | 2 | 2024-02-20T02:36:27 | 2024-03-06T18:06:53 | 2024-02-20T03:38:25 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | My drive C is a bit ancient right now. It's an old 250GB SSD and at any given time seems to have about 5-10gb free, so I'm forever doing cleanups to stop it running out of space. In contrast, I have about 2-3TB of free space on my other drives. I would _much_ prefer it if ollama would store models on one of those drive... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2602/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2602/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7647 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7647/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7647/comments | https://api.github.com/repos/ollama/ollama/issues/7647/events | https://github.com/ollama/ollama/issues/7647 | 2,654,957,622 | I_kwDOJ0Z1Ps6eP3A2 | 7,647 | Sending "hidden" messages for attachments | {
"login": "robotom",
"id": 45123215,
"node_id": "MDQ6VXNlcjQ1MTIzMjE1",
"avatar_url": "https://avatars.githubusercontent.com/u/45123215?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/robotom",
"html_url": "https://github.com/robotom",
"followers_url": "https://api.github.com/users/roboto... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 1 | 2024-11-13T10:20:06 | 2024-11-13T19:53:25 | 2024-11-13T19:53:24 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I'm building a frontend and trying to send attachment contents and an accompanying prompt (e.g. "analyze this and wait for user prompt before responding") as a "hidden" message so that it's not rendered to the user on the frontend. I'm trying to avoid seeing 10 pages of a doc rendered in the chat window.
Perhaps th... | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7647/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7647/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8349 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8349/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8349/comments | https://api.github.com/repos/ollama/ollama/issues/8349/events | https://github.com/ollama/ollama/issues/8349 | 2,775,993,379 | I_kwDOJ0Z1Ps6ldkwj | 8,349 | can't use gpu after update | {
"login": "Heart-eartH",
"id": 75364983,
"node_id": "MDQ6VXNlcjc1MzY0OTgz",
"avatar_url": "https://avatars.githubusercontent.com/u/75364983?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Heart-eartH",
"html_url": "https://github.com/Heart-eartH",
"followers_url": "https://api.github.com/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg... | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 7 | 2025-01-08T17:57:28 | 2025-01-28T20:46:29 | 2025-01-28T20:46:29 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Previously, it run on the GPU by default, but after the update, it run on the CPU. What can I do to change it
### OS
Windows
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.5.4 | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8349/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8349/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1552 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1552/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1552/comments | https://api.github.com/repos/ollama/ollama/issues/1552/events | https://github.com/ollama/ollama/pull/1552 | 2,044,274,427 | PR_kwDOJ0Z1Ps5iI-Kk | 1,552 | add lint and test on pull_request | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-12-15T19:34:07 | 2024-01-11T17:37:47 | 2024-01-11T17:37:46 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1552",
"html_url": "https://github.com/ollama/ollama/pull/1552",
"diff_url": "https://github.com/ollama/ollama/pull/1552.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1552.patch",
"merged_at": "2024-01-11T17:37:46"
} | fixes a bug with generate where `get_flags` errors on ubuntu (and likely windows) when building cuda on a cuda-less system
fixes a bug in windows where `/api/list` does not return models correctly
both lint and test requires go generate results so do it once then propagate the artifacts to the rest of the pipelin... | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1552/reactions",
"total_count": 2,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 1,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1552/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6272 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6272/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6272/comments | https://api.github.com/repos/ollama/ollama/issues/6272/events | https://github.com/ollama/ollama/issues/6272 | 2,457,035,391 | I_kwDOJ0Z1Ps6Sc2J_ | 6,272 | Ollama Creat 手动部署 报错 Error: invalid file magic | {
"login": "JaminYan",
"id": 83393121,
"node_id": "MDQ6VXNlcjgzMzkzMTIx",
"avatar_url": "https://avatars.githubusercontent.com/u/83393121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/JaminYan",
"html_url": "https://github.com/JaminYan",
"followers_url": "https://api.github.com/users/Jam... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 22 | 2024-08-09T03:47:19 | 2024-11-11T09:41:36 | 2024-08-25T20:15:11 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
按飞书文档找的modelfile 文件,ollama creat 后输出报错
Error: invalid file magic ,因此无法部署在ollama
### OS
Windows
### GPU
Nvidia
### CPU
AMD
### Ollama version
0.34 | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6272/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6272/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5633 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5633/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5633/comments | https://api.github.com/repos/ollama/ollama/issues/5633/events | https://github.com/ollama/ollama/pull/5633 | 2,403,487,442 | PR_kwDOJ0Z1Ps51HNIc | 5,633 | Update README.md >> Extension & Plugins :Terraform AWS Ollama & Open WebUI | {
"login": "xuyangbocn",
"id": 11894567,
"node_id": "MDQ6VXNlcjExODk0NTY3",
"avatar_url": "https://avatars.githubusercontent.com/u/11894567?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/xuyangbocn",
"html_url": "https://github.com/xuyangbocn",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 0 | 2024-07-11T15:29:25 | 2024-11-21T10:28:58 | 2024-11-21T10:28:58 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5633",
"html_url": "https://github.com/ollama/ollama/pull/5633",
"diff_url": "https://github.com/ollama/ollama/pull/5633.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5633.patch",
"merged_at": "2024-11-21T10:28:58"
} | A Terraform module to deploy on AWS a ready-to-use Ollama service, together with its front end Open WebUI service. | {
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/us... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5633/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5633/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3733 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3733/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3733/comments | https://api.github.com/repos/ollama/ollama/issues/3733/events | https://github.com/ollama/ollama/issues/3733 | 2,250,701,058 | I_kwDOJ0Z1Ps6GJvkC | 3,733 | I wrote an uninstallation script. | {
"login": "TikaFlow",
"id": 33306473,
"node_id": "MDQ6VXNlcjMzMzA2NDcz",
"avatar_url": "https://avatars.githubusercontent.com/u/33306473?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/TikaFlow",
"html_url": "https://github.com/TikaFlow",
"followers_url": "https://api.github.com/users/Tik... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2024-04-18T13:26:28 | 2024-04-19T01:59:32 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi, I wrote an uninstallation script and [this](https://github.com/TikaFlow/tika_grocery/tree/main/delOllama) is my script.
If you think it's not bad, you might consider referencing my script in your [documentation](https://github.com/ollama/ollama/blob/main/docs/linux.md#uninstall).
Thanks! | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3733/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3733/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/4409 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4409/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4409/comments | https://api.github.com/repos/ollama/ollama/issues/4409/events | https://github.com/ollama/ollama/issues/4409 | 2,293,631,935 | I_kwDOJ0Z1Ps6Itgu_ | 4,409 | Windows security: ALERT: Trojan:Win32/Wacatac.H!ml detected | {
"login": "boessu",
"id": 2807976,
"node_id": "MDQ6VXNlcjI4MDc5NzY=",
"avatar_url": "https://avatars.githubusercontent.com/u/2807976?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/boessu",
"html_url": "https://github.com/boessu",
"followers_url": "https://api.github.com/users/boessu/foll... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2024-05-13T19:17:50 | 2024-05-13T20:03:24 | 2024-05-13T20:03:24 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
There is an alarm in Windows Defender which sets the following file under quarantine:
file: %USERPROFILE%\AppData\Local\Programs\Ollama\ollama app.exe
Windows Version: Windows 11 Pro, 23H2
You'll get the warning with update and also if you install it new.
### OS
Windows
### GPU
AMD
#... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4409/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4409/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5743 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5743/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5743/comments | https://api.github.com/repos/ollama/ollama/issues/5743/events | https://github.com/ollama/ollama/issues/5743 | 2,413,191,986 | I_kwDOJ0Z1Ps6P1mMy | 5,743 | process cant't kill on windows 10 | {
"login": "wszgrcy",
"id": 9607121,
"node_id": "MDQ6VXNlcjk2MDcxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/9607121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/wszgrcy",
"html_url": "https://github.com/wszgrcy",
"followers_url": "https://api.github.com/users/wszgrcy/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg... | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 5 | 2024-07-17T10:04:20 | 2024-07-19T03:29:24 | 2024-07-19T03:29:24 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I use child process in node to start ollama,and when I kill process,The process will be partially retained(`ollama.exe` is killed ollama_llama_server retained)

Ubuntu will correctly kill all processes
... | {
"login": "wszgrcy",
"id": 9607121,
"node_id": "MDQ6VXNlcjk2MDcxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/9607121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/wszgrcy",
"html_url": "https://github.com/wszgrcy",
"followers_url": "https://api.github.com/users/wszgrcy/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5743/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5743/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7662 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7662/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7662/comments | https://api.github.com/repos/ollama/ollama/issues/7662/events | https://github.com/ollama/ollama/issues/7662 | 2,657,523,102 | I_kwDOJ0Z1Ps6eZpWe | 7,662 | How can I remove specific PARAMETER from model? | {
"login": "wwjCMP",
"id": 32979859,
"node_id": "MDQ6VXNlcjMyOTc5ODU5",
"avatar_url": "https://avatars.githubusercontent.com/u/32979859?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/wwjCMP",
"html_url": "https://github.com/wwjCMP",
"followers_url": "https://api.github.com/users/wwjCMP/fo... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 2 | 2024-11-14T04:27:14 | 2024-11-14T14:17:11 | 2024-11-14T14:17:11 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | How can I remove specific PARAMETER from model
For example, delete PARAMETER num_gpu 30 from qwen2.5-coder:32b-instruct-q8_0
ollama show qwen2.5-coder:32b-instruct-q8_0
Model
architecture qwen2
parameters 32.8B
context length 32768
embedding length 5120
quantiza... | {
"login": "wwjCMP",
"id": 32979859,
"node_id": "MDQ6VXNlcjMyOTc5ODU5",
"avatar_url": "https://avatars.githubusercontent.com/u/32979859?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/wwjCMP",
"html_url": "https://github.com/wwjCMP",
"followers_url": "https://api.github.com/users/wwjCMP/fo... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7662/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7662/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3330 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3330/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3330/comments | https://api.github.com/repos/ollama/ollama/issues/3330/events | https://github.com/ollama/ollama/issues/3330 | 2,204,580,969 | I_kwDOJ0Z1Ps6DZzxp | 3,330 | Failed to open the file /tmp/tmp.RbBP0lFvPD/ollama: | {
"login": "hansaskov",
"id": 70580312,
"node_id": "MDQ6VXNlcjcwNTgwMzEy",
"avatar_url": "https://avatars.githubusercontent.com/u/70580312?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hansaskov",
"html_url": "https://github.com/hansaskov",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2024-03-24T22:08:56 | 2024-03-24T22:17:04 | 2024-03-24T22:17:04 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I am unable to download and install ollama on ubuntu 22.04. I used the following command
```bash
sudo curl -fsSL https://ollama.com/install.sh | sh
```
And i get the following output
```
>>> Downloading ollama...
Warning: Failed to open the file /tmp/tmp.uSBa9PbYEo/ollama: No such ... | {
"login": "hansaskov",
"id": 70580312,
"node_id": "MDQ6VXNlcjcwNTgwMzEy",
"avatar_url": "https://avatars.githubusercontent.com/u/70580312?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hansaskov",
"html_url": "https://github.com/hansaskov",
"followers_url": "https://api.github.com/users/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3330/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3330/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8271 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8271/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8271/comments | https://api.github.com/repos/ollama/ollama/issues/8271/events | https://github.com/ollama/ollama/issues/8271 | 2,763,974,531 | I_kwDOJ0Z1Ps6kvueD | 8,271 | llama runner process terminated: CUDA error | {
"login": "iplayfast",
"id": 751306,
"node_id": "MDQ6VXNlcjc1MTMwNg==",
"avatar_url": "https://avatars.githubusercontent.com/u/751306?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/iplayfast",
"html_url": "https://github.com/iplayfast",
"followers_url": "https://api.github.com/users/ipla... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q... | closed | false | null | [] | null | 14 | 2024-12-31T05:34:54 | 2025-01-01T07:33:02 | 2025-01-01T07:33:02 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I'm making a little embedding example. The text is chunked fine, and embedded. But on retrieval, the I get this weird message.
Then if I go to the cli and try running the same model, I get the same message.
I'm pretty sure it has to do with having the embedded model in memory at the same ... | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8271/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8271/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7839 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7839/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7839/comments | https://api.github.com/repos/ollama/ollama/issues/7839/events | https://github.com/ollama/ollama/pull/7839 | 2,693,432,609 | PR_kwDOJ0Z1Ps6DI6lo | 7,839 | fix(docs): Replace hard coded num of make threads w/ CPU cores | {
"login": "sammcj",
"id": 862951,
"node_id": "MDQ6VXNlcjg2Mjk1MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sammcj",
"html_url": "https://github.com/sammcj",
"followers_url": "https://api.github.com/users/sammcj/follow... | [] | closed | false | null | [] | null | 0 | 2024-11-26T06:54:28 | 2024-12-06T22:51:09 | 2024-12-06T22:51:04 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7839",
"html_url": "https://github.com/ollama/ollama/pull/7839",
"diff_url": "https://github.com/ollama/ollama/pull/7839.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7839.patch",
"merged_at": null
} | Replace the seemingly random number of make threads with the number CPU cores / 2. | {
"login": "sammcj",
"id": 862951,
"node_id": "MDQ6VXNlcjg2Mjk1MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sammcj",
"html_url": "https://github.com/sammcj",
"followers_url": "https://api.github.com/users/sammcj/follow... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7839/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7839/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3914 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3914/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3914/comments | https://api.github.com/repos/ollama/ollama/issues/3914/events | https://github.com/ollama/ollama/pull/3914 | 2,264,038,325 | PR_kwDOJ0Z1Ps5twARO | 3,914 | Improve mac parallel performance | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-04-25T16:40:14 | 2024-04-25T23:28:34 | 2024-04-25T23:28:31 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3914",
"html_url": "https://github.com/ollama/ollama/pull/3914",
"diff_url": "https://github.com/ollama/ollama/pull/3914.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3914.patch",
"merged_at": "2024-04-25T23:28:31"
} | Carries #3900
Testing underway... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3914/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3914/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4048 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4048/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4048/comments | https://api.github.com/repos/ollama/ollama/issues/4048/events | https://github.com/ollama/ollama/issues/4048 | 2,271,291,702 | I_kwDOJ0Z1Ps6HYSk2 | 4,048 | Windows installation is behaving like malware | {
"login": "eabase",
"id": 52289379,
"node_id": "MDQ6VXNlcjUyMjg5Mzc5",
"avatar_url": "https://avatars.githubusercontent.com/u/52289379?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/eabase",
"html_url": "https://github.com/eabase",
"followers_url": "https://api.github.com/users/eabase/fo... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 8 | 2024-04-30T12:06:32 | 2024-05-08T12:43:37 | 2024-05-01T16:34:27 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
This is beyond bad.
1. It installs stuff all over the place, without informing the user.
```
C:\Users\xxxx\AppData\Roaming\Microsoft\Windows\Start Menu\Programs
C:\Users\xxxx\AppData\Local\Temp\ollamaNNNNNN
C:\Users\xxxx\.ollama\id_ed25519
C:\Users\xxxx\AppData\Local\Programs\Oll... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4048/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4048/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3025 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3025/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3025/comments | https://api.github.com/repos/ollama/ollama/issues/3025/events | https://github.com/ollama/ollama/pull/3025 | 2,177,367,153 | PR_kwDOJ0Z1Ps5pJu-X | 3,025 | Rename ROCm deps file to avoid confusion | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-03-09T19:22:33 | 2024-03-10T03:50:45 | 2024-03-10T01:48:39 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3025",
"html_url": "https://github.com/ollama/ollama/pull/3025",
"diff_url": "https://github.com/ollama/ollama/pull/3025.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3025.patch",
"merged_at": "2024-03-10T01:48:39"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3025/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3025/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7111 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7111/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7111/comments | https://api.github.com/repos/ollama/ollama/issues/7111/events | https://github.com/ollama/ollama/issues/7111 | 2,569,542,877 | I_kwDOJ0Z1Ps6ZKBzd | 7,111 | Any plans to add nvidia/NVLM-D-72B new model to ollama? | {
"login": "SilverLaius",
"id": 26542231,
"node_id": "MDQ6VXNlcjI2NTQyMjMx",
"avatar_url": "https://avatars.githubusercontent.com/u/26542231?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/SilverLaius",
"html_url": "https://github.com/SilverLaius",
"followers_url": "https://api.github.com/... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 1 | 2024-10-07T07:08:00 | 2024-10-07T07:09:13 | 2024-10-07T07:08:51 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Nvidia just released a state of the art model that would be a good addition to ollama. Any plans to add it? [https://huggingface.co/nvidia/NVLM-D-72B/tree/main](https://huggingface.co/nvidia/NVLM-D-72B/tree/main) | {
"login": "SilverLaius",
"id": 26542231,
"node_id": "MDQ6VXNlcjI2NTQyMjMx",
"avatar_url": "https://avatars.githubusercontent.com/u/26542231?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/SilverLaius",
"html_url": "https://github.com/SilverLaius",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7111/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7111/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7619 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7619/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7619/comments | https://api.github.com/repos/ollama/ollama/issues/7619/events | https://github.com/ollama/ollama/issues/7619 | 2,648,701,609 | I_kwDOJ0Z1Ps6d3_qp | 7,619 | llama3.2-vision on multi gpu error | {
"login": "18600709862",
"id": 46145698,
"node_id": "MDQ6VXNlcjQ2MTQ1Njk4",
"avatar_url": "https://avatars.githubusercontent.com/u/46145698?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/18600709862",
"html_url": "https://github.com/18600709862",
"followers_url": "https://api.github.com/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg... | closed | false | null | [] | null | 7 | 2024-11-11T09:26:43 | 2024-11-12T20:59:09 | 2024-11-12T20:58:58 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
multi gpu
ollama run llama3.2-vision
>>> The image is a book cover. Output should be in this format - <Name of the Book>: <Name of the Author>. Do not output anything else /media/root/ssd2t/data/pro/tmp/o
... l/new/FastChat/image.png
Added image '/media/root/ssd2t/data/pro/tmp/ol/new/FastCha... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7619/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7619/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/748 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/748/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/748/comments | https://api.github.com/repos/ollama/ollama/issues/748/events | https://github.com/ollama/ollama/issues/748 | 1,935,466,908 | I_kwDOJ0Z1Ps5zXOGc | 748 | Enabling langchain | {
"login": "suoko",
"id": 3659980,
"node_id": "MDQ6VXNlcjM2NTk5ODA=",
"avatar_url": "https://avatars.githubusercontent.com/u/3659980?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/suoko",
"html_url": "https://github.com/suoko",
"followers_url": "https://api.github.com/users/suoko/follower... | [] | closed | false | null | [] | null | 2 | 2023-10-10T14:13:06 | 2023-10-11T15:45:32 | 2023-10-11T15:45:32 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I found this video about implementing langchain with ollama. If available it should be easy to use it via ollama-ui
https://www.youtube.com/watch?v=CPgp8MhmGVY | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/748/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/748/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2844 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2844/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2844/comments | https://api.github.com/repos/ollama/ollama/issues/2844/events | https://github.com/ollama/ollama/issues/2844 | 2,162,130,461 | I_kwDOJ0Z1Ps6A334d | 2,844 | OpenAI package compatibility | {
"login": "eliranwong",
"id": 25262722,
"node_id": "MDQ6VXNlcjI1MjYyNzIy",
"avatar_url": "https://avatars.githubusercontent.com/u/25262722?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/eliranwong",
"html_url": "https://github.com/eliranwong",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 6 | 2024-02-29T21:26:43 | 2024-05-02T01:23:23 | 2024-03-01T01:08:58 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I read the example in https://ollama.com/blog/openai-compatibility
The example works, but it doesn't when I add "response_format={ "type": "json_object" },"
https://platform.openai.com/docs/guides/text-generation/json-mode | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2844/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2844/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/285 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/285/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/285/comments | https://api.github.com/repos/ollama/ollama/issues/285/events | https://github.com/ollama/ollama/pull/285 | 1,837,254,131 | PR_kwDOJ0Z1Ps5XOXrT | 285 | Update README.md | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 0 | 2023-08-04T19:43:08 | 2023-08-04T19:45:56 | 2023-08-04T19:45:55 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/285",
"html_url": "https://github.com/ollama/ollama/pull/285",
"diff_url": "https://github.com/ollama/ollama/pull/285.diff",
"patch_url": "https://github.com/ollama/ollama/pull/285.patch",
"merged_at": "2023-08-04T19:45:55"
} | Ollama now supports Intel Macs | {
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/us... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/285/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/285/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2392 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2392/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2392/comments | https://api.github.com/repos/ollama/ollama/issues/2392/events | https://github.com/ollama/ollama/issues/2392 | 2,123,591,240 | I_kwDOJ0Z1Ps5-k25I | 2,392 | unable to initialize llm library Radeon card detected | {
"login": "sigmaya",
"id": 159318222,
"node_id": "U_kgDOCX8Azg",
"avatar_url": "https://avatars.githubusercontent.com/u/159318222?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sigmaya",
"html_url": "https://github.com/sigmaya",
"followers_url": "https://api.github.com/users/sigmaya/foll... | [] | closed | false | null | [] | null | 6 | 2024-02-07T18:04:01 | 2024-02-29T00:50:07 | 2024-02-07T22:47:19 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hello,
I am trying to run as user and manually, i get this error:
time=2024-02-07T19:00:18.967+01:00 level=INFO source=payload_common.go:106 msg="Extracting dynamic libraries..."
Error: unable to initialize llm library Radeon card detected, but permissions not set up properly. Either run ollama as root, or add you ... | {
"login": "sigmaya",
"id": 159318222,
"node_id": "U_kgDOCX8Azg",
"avatar_url": "https://avatars.githubusercontent.com/u/159318222?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sigmaya",
"html_url": "https://github.com/sigmaya",
"followers_url": "https://api.github.com/users/sigmaya/foll... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2392/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2392/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5488 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5488/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5488/comments | https://api.github.com/repos/ollama/ollama/issues/5488/events | https://github.com/ollama/ollama/issues/5488 | 2,391,347,345 | I_kwDOJ0Z1Ps6OiRCR | 5,488 | falcon2 fails to genarate code in almalinux9 | {
"login": "olumolu",
"id": 162728301,
"node_id": "U_kgDOCbMJbQ",
"avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/olumolu",
"html_url": "https://github.com/olumolu",
"followers_url": "https://api.github.com/users/olumolu/foll... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 0 | 2024-07-04T18:55:46 | 2024-09-26T12:39:47 | 2024-09-26T12:39:47 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I got a error that it could not able to genarate any code but it failed and just skipped and jumped 5lines of free space and ask for input again.
### OS
Linux
### GPU
_No response_
### CPU
Intel
### Ollama version
1.48 | {
"login": "olumolu",
"id": 162728301,
"node_id": "U_kgDOCbMJbQ",
"avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/olumolu",
"html_url": "https://github.com/olumolu",
"followers_url": "https://api.github.com/users/olumolu/foll... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5488/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5488/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3287 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3287/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3287/comments | https://api.github.com/repos/ollama/ollama/issues/3287/events | https://github.com/ollama/ollama/issues/3287 | 2,200,586,023 | I_kwDOJ0Z1Ps6DKkcn | 3,287 | Allow system message to be on its own | {
"login": "jackielii",
"id": 360983,
"node_id": "MDQ6VXNlcjM2MDk4Mw==",
"avatar_url": "https://avatars.githubusercontent.com/u/360983?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jackielii",
"html_url": "https://github.com/jackielii",
"followers_url": "https://api.github.com/users/jack... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api... | null | 3 | 2024-03-21T16:05:18 | 2024-09-04T04:42:37 | 2024-09-04T04:42:37 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What are you trying to do?
Current prompt rendering assumes there is only one system message. With the [Command-R model from Cohere AI](https://huggingface.co/CohereForAI/c4ai-command-r-v01), there can be multiple system message: https://docs.cohere.com/docs/prompting-command-r#structured-prompt-templates. There i... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3287/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3287/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3711 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3711/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3711/comments | https://api.github.com/repos/ollama/ollama/issues/3711/events | https://github.com/ollama/ollama/issues/3711 | 2,249,202,461 | I_kwDOJ0Z1Ps6GEBsd | 3,711 | CUDA malloc fails on newly supported models in 0.1.32 (dual-GPU setup with 72GB VRAM and 128GB RAM) | {
"login": "mz2",
"id": 71363,
"node_id": "MDQ6VXNlcjcxMzYz",
"avatar_url": "https://avatars.githubusercontent.com/u/71363?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mz2",
"html_url": "https://github.com/mz2",
"followers_url": "https://api.github.com/users/mz2/followers",
"following... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 16 | 2024-04-17T20:18:35 | 2024-06-22T00:10:31 | 2024-06-22T00:10:31 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I am getting cuda malloc errors with v0.1.32 (as well as with the current head of main branch) when trying any of the new big models: wizardlm2, mixtral:8x22b, dbrx (command-r+ does work) with my dual GPU setup (A6000 + RTX 3090, i.e. combined 72GB VRAM) with a 24-core 13th gen Intel CPU (128GB ... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3711/reactions",
"total_count": 8,
"+1": 8,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3711/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3779 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3779/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3779/comments | https://api.github.com/repos/ollama/ollama/issues/3779/events | https://github.com/ollama/ollama/issues/3779 | 2,254,549,576 | I_kwDOJ0Z1Ps6GYbJI | 3,779 | How to check memory utilization rate | {
"login": "taozhiyuai",
"id": 146583103,
"node_id": "U_kgDOCLyuPw",
"avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/taozhiyuai",
"html_url": "https://github.com/taozhiyuai",
"followers_url": "https://api.github.com/users/tao... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 2 | 2024-04-20T13:19:09 | 2024-05-02T10:55:11 | 2024-05-02T10:55:11 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | after loading a model,How to check v-memory utilization rate | {
"login": "taozhiyuai",
"id": 146583103,
"node_id": "U_kgDOCLyuPw",
"avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/taozhiyuai",
"html_url": "https://github.com/taozhiyuai",
"followers_url": "https://api.github.com/users/tao... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3779/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3779/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/431 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/431/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/431/comments | https://api.github.com/repos/ollama/ollama/issues/431/events | https://github.com/ollama/ollama/issues/431 | 1,868,486,927 | I_kwDOJ0Z1Ps5vXtkP | 431 | UTF-8 characters | {
"login": "arpecop",
"id": 168974,
"node_id": "MDQ6VXNlcjE2ODk3NA==",
"avatar_url": "https://avatars.githubusercontent.com/u/168974?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/arpecop",
"html_url": "https://github.com/arpecop",
"followers_url": "https://api.github.com/users/arpecop/fo... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA... | closed | false | null | [] | null | 4 | 2023-08-27T12:56:34 | 2023-08-30T21:11:50 | 2023-08-30T21:11:50 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | How to render properly such characters , is this per-model issue ... or font related ?

| {
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/us... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/431/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/431/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1546 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1546/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1546/comments | https://api.github.com/repos/ollama/ollama/issues/1546/events | https://github.com/ollama/ollama/issues/1546 | 2,044,086,055 | I_kwDOJ0Z1Ps551kcn | 1,546 | Is running behind a proxy fully supported? | {
"login": "pardeike",
"id": 853584,
"node_id": "MDQ6VXNlcjg1MzU4NA==",
"avatar_url": "https://avatars.githubusercontent.com/u/853584?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pardeike",
"html_url": "https://github.com/pardeike",
"followers_url": "https://api.github.com/users/pardeik... | [] | closed | false | null | [] | null | 4 | 2023-12-15T17:16:41 | 2023-12-28T12:17:29 | 2023-12-15T17:24:09 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi,
when I run with a configured proxy, I get the error `lookup registry.ollama.ai: no such host`:
```
> HTTPS_PROXY=http://REDACTED.XXX:10000 ollama run llama2
pulling manifest
Error: pull model manifest: Get "https://registry.ollama.ai/v2/library/llama2/manifests/latest": dial tcp: lookup registry.ollama.ai: ... | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1546/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1546/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1151 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1151/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1151/comments | https://api.github.com/repos/ollama/ollama/issues/1151/events | https://github.com/ollama/ollama/pull/1151 | 1,996,326,933 | PR_kwDOJ0Z1Ps5fmfF0 | 1,151 | install: fix enable contrib on debian 12 | {
"login": "yanndegat",
"id": 17246629,
"node_id": "MDQ6VXNlcjE3MjQ2NjI5",
"avatar_url": "https://avatars.githubusercontent.com/u/17246629?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yanndegat",
"html_url": "https://github.com/yanndegat",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 1 | 2023-11-16T08:23:44 | 2023-11-16T20:53:07 | 2023-11-16T20:53:07 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1151",
"html_url": "https://github.com/ollama/ollama/pull/1151",
"diff_url": "https://github.com/ollama/ollama/pull/1151.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1151.patch",
"merged_at": "2023-11-16T20:53:07"
} | On debian 12, sources definitions have moved from
/etc/apt/sources.list to /etc/apt/sources.list.d/debian.sources | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1151/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1151/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7571 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7571/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7571/comments | https://api.github.com/repos/ollama/ollama/issues/7571/events | https://github.com/ollama/ollama/issues/7571 | 2,643,262,341 | I_kwDOJ0Z1Ps6djPuF | 7,571 | Please add Qwen-VL Qwen2.5-VL Qwen2-VL multimodal large model | {
"login": "smileyboy2019",
"id": 59221294,
"node_id": "MDQ6VXNlcjU5MjIxMjk0",
"avatar_url": "https://avatars.githubusercontent.com/u/59221294?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/smileyboy2019",
"html_url": "https://github.com/smileyboy2019",
"followers_url": "https://api.githu... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 2 | 2024-11-08T08:19:59 | 2024-11-12T00:40:52 | 2024-11-12T00:40:52 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7571/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7571/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4353 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4353/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4353/comments | https://api.github.com/repos/ollama/ollama/issues/4353/events | https://github.com/ollama/ollama/issues/4353 | 2,290,802,195 | I_kwDOJ0Z1Ps6Iit4T | 4,353 | Unable to determine if Ollama can be run on GPU after running Linux commands | {
"login": "jerry123cao",
"id": 124961434,
"node_id": "U_kgDOB3LCmg",
"avatar_url": "https://avatars.githubusercontent.com/u/124961434?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jerry123cao",
"html_url": "https://github.com/jerry123cao",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 1 | 2024-05-11T09:18:56 | 2024-05-21T23:21:51 | 2024-05-21T23:21:48 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hello, I have a question. If I use the command curl - fsSL https://ollama.com/install.sh |Can Ollama run on GPU after installing the Nvidia driver and cuda toolkit in SH? | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4353/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4353/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5292 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5292/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5292/comments | https://api.github.com/repos/ollama/ollama/issues/5292/events | https://github.com/ollama/ollama/issues/5292 | 2,374,357,161 | I_kwDOJ0Z1Ps6NhdCp | 5,292 | 请上架InternVL | {
"login": "enryteam",
"id": 20081090,
"node_id": "MDQ6VXNlcjIwMDgxMDkw",
"avatar_url": "https://avatars.githubusercontent.com/u/20081090?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/enryteam",
"html_url": "https://github.com/enryteam",
"followers_url": "https://api.github.com/users/enr... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | open | false | null | [] | null | 1 | 2024-06-26T05:58:42 | 2025-01-28T13:33:00 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ERROR: type should be string, got "\r\nhttps://github.com/OpenGVLab/InternVL\r\n\r\nthanks 谢谢\r\n\r\n\r\n\r\nollama0.1.43\r\nerror format not yet support!\r\n错误格式尚不支持!\r\n错误格式尚不支持!\r\n搞了多次 均报错" | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5292/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5292/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/367 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/367/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/367/comments | https://api.github.com/repos/ollama/ollama/issues/367/events | https://github.com/ollama/ollama/pull/367 | 1,854,193,882 | PR_kwDOJ0Z1Ps5YHWrd | 367 | fix nous-hermes model file size listing in readme | {
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/us... | [] | closed | false | null | [] | null | 0 | 2023-08-17T03:41:42 | 2023-08-17T03:42:01 | 2023-08-17T03:42:00 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/367",
"html_url": "https://github.com/ollama/ollama/pull/367",
"diff_url": "https://github.com/ollama/ollama/pull/367.diff",
"patch_url": "https://github.com/ollama/ollama/pull/367.patch",
"merged_at": "2023-08-17T03:42:00"
} | fix nous-hermes model file size listing in readme | {
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/us... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/367/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/367/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6661 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6661/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6661/comments | https://api.github.com/repos/ollama/ollama/issues/6661/events | https://github.com/ollama/ollama/pull/6661 | 2,508,691,339 | PR_kwDOJ0Z1Ps56kwXL | 6,661 | commit | {
"login": "rpreslar4765",
"id": 89657947,
"node_id": "MDQ6VXNlcjg5NjU3OTQ3",
"avatar_url": "https://avatars.githubusercontent.com/u/89657947?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rpreslar4765",
"html_url": "https://github.com/rpreslar4765",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | 1 | 2024-09-05T20:53:42 | 2024-09-05T21:16:01 | 2024-09-05T21:16:01 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/6661",
"html_url": "https://github.com/ollama/ollama/pull/6661",
"diff_url": "https://github.com/ollama/ollama/pull/6661.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6661.patch",
"merged_at": null
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6661/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6661/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8345 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8345/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8345/comments | https://api.github.com/repos/ollama/ollama/issues/8345/events | https://github.com/ollama/ollama/issues/8345 | 2,775,164,105 | I_kwDOJ0Z1Ps6laaTJ | 8,345 | llama3.2-vision does not support tools | {
"login": "kwaa",
"id": 50108258,
"node_id": "MDQ6VXNlcjUwMTA4MjU4",
"avatar_url": "https://avatars.githubusercontent.com/u/50108258?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kwaa",
"html_url": "https://github.com/kwaa",
"followers_url": "https://api.github.com/users/kwaa/followers"... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/... | [
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "htt... | null | 17 | 2025-01-08T11:34:22 | 2025-01-18T09:30:37 | 2025-01-15T21:25:43 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I get above error when I use `llama3.2-vision:11b` with tool calling.
I think `llama3.2-vision:11b` should support tool calls: https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/vision_prompt_format.md#builtin-and-zero-shot-tool-calling
- [llama3.2:3b template](https://ol... | {
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8345/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8345/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1421 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1421/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1421/comments | https://api.github.com/repos/ollama/ollama/issues/1421/events | https://github.com/ollama/ollama/pull/1421 | 2,031,577,402 | PR_kwDOJ0Z1Ps5hd4Rw | 1,421 | fix redundant newline | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-12-07T21:44:54 | 2023-12-07T21:47:24 | 2023-12-07T21:47:23 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1421",
"html_url": "https://github.com/ollama/ollama/pull/1421",
"diff_url": "https://github.com/ollama/ollama/pull/1421.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1421.patch",
"merged_at": "2023-12-07T21:47:23"
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1421/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1421/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8020 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8020/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8020/comments | https://api.github.com/repos/ollama/ollama/issues/8020/events | https://github.com/ollama/ollama/issues/8020 | 2,728,619,855 | I_kwDOJ0Z1Ps6io29P | 8,020 | Nancy finds security vulnerabilities | {
"login": "mitar",
"id": 585279,
"node_id": "MDQ6VXNlcjU4NTI3OQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/585279?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mitar",
"html_url": "https://github.com/mitar",
"followers_url": "https://api.github.com/users/mitar/followers"... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 0 | 2024-12-10T00:29:13 | 2024-12-10T00:29:13 | null | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
It finds:
* [CVE-2024-8063](https://ossindex.sonatype.org/vulnerability/CVE-2024-8063?component-type=golang&component-name=github.com%2Follama%2Follama&utm_source=nancy-client&utm_medium=integration&utm_content=1.0.46) CWE-369: Divide By Zero
* [CVE-2024-39719](https://ossindex.sonatype.org/vu... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8020/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8020/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/2467 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2467/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2467/comments | https://api.github.com/repos/ollama/ollama/issues/2467/events | https://github.com/ollama/ollama/pull/2467 | 2,130,595,963 | PR_kwDOJ0Z1Ps5mqQ3_ | 2,467 | Check image filetype in api handlers | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 0 | 2024-02-12T17:17:19 | 2024-02-12T19:16:21 | 2024-02-12T19:16:20 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2467",
"html_url": "https://github.com/ollama/ollama/pull/2467",
"diff_url": "https://github.com/ollama/ollama/pull/2467.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2467.patch",
"merged_at": "2024-02-12T19:16:20"
} | Fixes: https://github.com/ollama/ollama/issues/2456 | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2467/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2467/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1067 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1067/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1067/comments | https://api.github.com/repos/ollama/ollama/issues/1067/events | https://github.com/ollama/ollama/pull/1067 | 1,986,694,006 | PR_kwDOJ0Z1Ps5fF3bO | 1,067 | add custom ollama-runner | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 1 | 2023-11-10T01:45:17 | 2023-12-06T23:54:38 | 2023-11-21T20:14:50 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1067",
"html_url": "https://github.com/ollama/ollama/pull/1067",
"diff_url": "https://github.com/ollama/ollama/pull/1067.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1067.patch",
"merged_at": null
} | - update llama.cpp examples with custom ollama-runner
- update llama.cpp gguf version to latest
This change adds a custom inference server to llama.cpp based on the server we use in the current version, but with excess features removed. This allows us to have a more stable interface to build on when llama.cpp updat... | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1067/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1067/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1420 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1420/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1420/comments | https://api.github.com/repos/ollama/ollama/issues/1420/events | https://github.com/ollama/ollama/pull/1420 | 2,031,437,051 | PR_kwDOJ0Z1Ps5hdZct | 1,420 | os specific ctrl-z | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 3 | 2023-12-07T20:03:08 | 2023-12-14T06:46:26 | 2023-12-11T15:48:15 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1420",
"html_url": "https://github.com/ollama/ollama/pull/1420",
"diff_url": "https://github.com/ollama/ollama/pull/1420.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1420.patch",
"merged_at": "2023-12-11T15:48:15"
} | Add OS specific readline functions. Windows does not support these suspend system calls, so make ctrl-z a no-op on windows. This fixes development windows native builds.
resolves #1414 | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1420/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1420/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7824 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7824/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7824/comments | https://api.github.com/repos/ollama/ollama/issues/7824/events | https://github.com/ollama/ollama/issues/7824 | 2,688,581,053 | I_kwDOJ0Z1Ps6gQH29 | 7,824 | Tool calling parsing for llama3.2 | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | {
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/... | [
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "htt... | null | 0 | 2024-11-24T23:12:47 | 2024-11-24T23:12:47 | null | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Llama 3.2 tool call outputs [are not in JSON](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/text_prompt_format.md) and so Ollama's tool parsing needs to be updated
### OS
_No response_
### GPU
_No response_
### CPU
_No response_
### Ollama version
_No response_ | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7824/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7824/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/2586 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2586/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2586/comments | https://api.github.com/repos/ollama/ollama/issues/2586/events | https://github.com/ollama/ollama/issues/2586 | 2,141,379,273 | I_kwDOJ0Z1Ps5_otrJ | 2,586 | Question on model updating | {
"login": "owenzhao",
"id": 2182896,
"node_id": "MDQ6VXNlcjIxODI4OTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/2182896?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/owenzhao",
"html_url": "https://github.com/owenzhao",
"followers_url": "https://api.github.com/users/owenz... | [
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] | closed | false | null | [] | null | 2 | 2024-02-19T02:42:14 | 2024-03-11T20:45:20 | 2024-03-11T20:45:11 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I know I could download the latest model when a model is not installed. Say:
```swift
ollama run llama2
```
After some days, say one month, the model will be updated. Can ollama updates to the latest model automatically? I didn't find a way to do that.
Please help. | {
"login": "hoyyeva",
"id": 63033505,
"node_id": "MDQ6VXNlcjYzMDMzNTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoyyeva",
"html_url": "https://github.com/hoyyeva",
"followers_url": "https://api.github.com/users/hoyyev... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2586/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2586/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2247 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2247/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2247/comments | https://api.github.com/repos/ollama/ollama/issues/2247/events | https://github.com/ollama/ollama/pull/2247 | 2,104,479,078 | PR_kwDOJ0Z1Ps5lRYk5 | 2,247 | Update import instructions to use convert and quantize tooling from llama.cpp submodule | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 0 | 2024-01-29T01:45:12 | 2024-02-05T05:50:45 | 2024-02-05T05:50:44 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2247",
"html_url": "https://github.com/ollama/ollama/pull/2247",
"diff_url": "https://github.com/ollama/ollama/pull/2247.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2247.patch",
"merged_at": "2024-02-05T05:50:44"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2247/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2247/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5994 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5994/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5994/comments | https://api.github.com/repos/ollama/ollama/issues/5994/events | https://github.com/ollama/ollama/pull/5994 | 2,432,953,146 | PR_kwDOJ0Z1Ps52nmhq | 5,994 | server: fix race conditions during download | {
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers"... | [] | closed | false | null | [] | null | 0 | 2024-07-26T20:44:18 | 2024-07-26T21:24:25 | 2024-07-26T21:24:24 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5994",
"html_url": "https://github.com/ollama/ollama/pull/5994",
"diff_url": "https://github.com/ollama/ollama/pull/5994.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5994.patch",
"merged_at": "2024-07-26T21:24:24"
} | This fixes various data races scattered throughout the download/pull client where the client was accessing the download state concurrently.
This commit is mostly a hot-fix and will be replaced by a new client one day soon.
Also, remove the unnecessary opts argument from downloadChunk. | {
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers"... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5994/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 1,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5994/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3839 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3839/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3839/comments | https://api.github.com/repos/ollama/ollama/issues/3839/events | https://github.com/ollama/ollama/issues/3839 | 2,258,206,018 | I_kwDOJ0Z1Ps6GmX1C | 3,839 | Feature Request: Detect Truncation Due to Exceeding Context Size | {
"login": "guoxf",
"id": 3947566,
"node_id": "MDQ6VXNlcjM5NDc1NjY=",
"avatar_url": "https://avatars.githubusercontent.com/u/3947566?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/guoxf",
"html_url": "https://github.com/guoxf",
"followers_url": "https://api.github.com/users/guoxf/follower... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 2 | 2024-04-23T07:54:37 | 2024-07-01T17:15:46 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Understanding whether model output has been truncated due to exceeding context size is crucial for trusting the model to provide complete and accurate information. Here are some specific examples that illustrate why it's necessary to know if the output has been truncated:
1. **Question-Answering Systems**: When buildi... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3839/reactions",
"total_count": 4,
"+1": 4,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3839/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/6044 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6044/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6044/comments | https://api.github.com/repos/ollama/ollama/issues/6044/events | https://github.com/ollama/ollama/issues/6044 | 2,435,056,414 | I_kwDOJ0Z1Ps6RJAMe | 6,044 | Illegal istruction in ollama_llama_server runner | {
"login": "SnowyCoder",
"id": 10945206,
"node_id": "MDQ6VXNlcjEwOTQ1MjA2",
"avatar_url": "https://avatars.githubusercontent.com/u/10945206?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/SnowyCoder",
"html_url": "https://github.com/SnowyCoder",
"followers_url": "https://api.github.com/use... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2024-07-29T10:34:52 | 2024-07-29T10:46:24 | 2024-07-29T10:46:23 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I tried to run llama3 model with ollama.
Reproduction (with my CPU: `AMD Ryzen 7 7735HS`)
1. Start server: `ollama serve`
2. Try to run llama3: `ollama run llama3
(The same error occurs with llama3.1)
The server will execute a new runner with the following arguments:
```/tmp/ollama996131... | {
"login": "SnowyCoder",
"id": 10945206,
"node_id": "MDQ6VXNlcjEwOTQ1MjA2",
"avatar_url": "https://avatars.githubusercontent.com/u/10945206?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/SnowyCoder",
"html_url": "https://github.com/SnowyCoder",
"followers_url": "https://api.github.com/use... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6044/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6044/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2368 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2368/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2368/comments | https://api.github.com/repos/ollama/ollama/issues/2368/events | https://github.com/ollama/ollama/pull/2368 | 2,119,989,649 | PR_kwDOJ0Z1Ps5mGd7d | 2,368 | Replace `reflect` usage in option parsing | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 3 | 2024-02-06T05:25:58 | 2024-08-11T16:52:11 | 2024-08-11T16:52:11 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2368",
"html_url": "https://github.com/ollama/ollama/pull/2368",
"diff_url": "https://github.com/ollama/ollama/pull/2368.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2368.patch",
"merged_at": null
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2368/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2368/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6929 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6929/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6929/comments | https://api.github.com/repos/ollama/ollama/issues/6929/events | https://github.com/ollama/ollama/issues/6929 | 2,544,478,139 | I_kwDOJ0Z1Ps6Xqae7 | 6,929 | support qwen2.5:72b-instruct function call | {
"login": "chenfolingithub",
"id": 9414696,
"node_id": "MDQ6VXNlcjk0MTQ2OTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/9414696?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/chenfolingithub",
"html_url": "https://github.com/chenfolingithub",
"followers_url": "https://api.g... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 2 | 2024-09-24T06:36:16 | 2024-11-06T00:18:09 | 2024-11-06T00:18:09 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | qwen2.5:72b-instruct now support function call
Returned its own unique data format, but currently ollama 0.3.11 does not correctly return formatted data for function calls
| {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6929/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6929/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4588 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4588/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4588/comments | https://api.github.com/repos/ollama/ollama/issues/4588/events | https://github.com/ollama/ollama/issues/4588 | 2,312,622,323 | I_kwDOJ0Z1Ps6J19Dz | 4,588 | Tool calling feature support | {
"login": "Goekdeniz-Guelmez",
"id": 60228478,
"node_id": "MDQ6VXNlcjYwMjI4NDc4",
"avatar_url": "https://avatars.githubusercontent.com/u/60228478?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Goekdeniz-Guelmez",
"html_url": "https://github.com/Goekdeniz-Guelmez",
"followers_url": "https... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 2 | 2024-05-23T11:06:39 | 2024-05-23T21:39:55 | 2024-05-23T21:39:47 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | since there are more and more models that support function/tool calling. I think it would be a great feature for that support. It can be hard because the models could need Special tokens. My thought is that you can be able to set these spetial tokens or customizations in the Modelfile, something like:
```text
PARAM... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4588/reactions",
"total_count": 4,
"+1": 4,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4588/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2424 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2424/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2424/comments | https://api.github.com/repos/ollama/ollama/issues/2424/events | https://github.com/ollama/ollama/issues/2424 | 2,126,622,691 | I_kwDOJ0Z1Ps5-wa_j | 2,424 | Always getting a timeout error while querying using mistral using Ollama | {
"login": "Chakit22",
"id": 118890138,
"node_id": "U_kgDOBxYemg",
"avatar_url": "https://avatars.githubusercontent.com/u/118890138?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Chakit22",
"html_url": "https://github.com/Chakit22",
"followers_url": "https://api.github.com/users/Chakit22/... | [] | closed | false | {
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers"... | [
{
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/... | null | 19 | 2024-02-09T07:36:05 | 2025-01-16T09:16:45 | 2024-05-07T23:53:24 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ```
Traceback (most recent call last):
File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpcore/_exceptions.py", line 10, in map_exceptions
yield
File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpcore/_backends/sync.py", line 126, in read
return self._sock.... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2424/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2424/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1025 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1025/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1025/comments | https://api.github.com/repos/ollama/ollama/issues/1025/events | https://github.com/ollama/ollama/issues/1025 | 1,980,404,467 | I_kwDOJ0Z1Ps52CpLz | 1,025 | How to stop a model aftering it's served? | {
"login": "paulwongx",
"id": 20178251,
"node_id": "MDQ6VXNlcjIwMTc4MjUx",
"avatar_url": "https://avatars.githubusercontent.com/u/20178251?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/paulwongx",
"html_url": "https://github.com/paulwongx",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 1 | 2023-11-07T02:12:01 | 2023-11-07T14:11:09 | 2023-11-07T14:11:09 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | When the modelling is already being served, what is the command to stop it so that I can serve it again with different variables? I want to run `OLLAMA_ORIGINS=https://XXXXX ollama serve` but I'm getting an error that the address is already in use.
Running `ollama --help` doesn't show a command to unserve...?
tld... | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1025/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1025/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4544 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4544/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4544/comments | https://api.github.com/repos/ollama/ollama/issues/4544/events | https://github.com/ollama/ollama/pull/4544 | 2,306,532,844 | PR_kwDOJ0Z1Ps5v_RLM | 4,544 | update langchain-python-rag-document example to use latest langchain apis | {
"login": "pixelsoccupied",
"id": 19938874,
"node_id": "MDQ6VXNlcjE5OTM4ODc0",
"avatar_url": "https://avatars.githubusercontent.com/u/19938874?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pixelsoccupied",
"html_url": "https://github.com/pixelsoccupied",
"followers_url": "https://api.gi... | [] | open | false | null | [] | null | 0 | 2024-05-20T18:27:53 | 2024-10-29T07:15:48 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4544",
"html_url": "https://github.com/ollama/ollama/pull/4544",
"diff_url": "https://github.com/ollama/ollama/pull/4544.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4544.patch",
"merged_at": null
} | Hello!
Langchain library went through a lot changes since this helpful example was first added and users will find that this will not work out of box anymore.
I have updated this to work with all the latest changes and also removed any deprecated functions.
Hope others find this useful! | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4544/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4544/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3174 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3174/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3174/comments | https://api.github.com/repos/ollama/ollama/issues/3174/events | https://github.com/ollama/ollama/pull/3174 | 2,189,661,685 | PR_kwDOJ0Z1Ps5pzx1y | 3,174 | llama: remove server static assets | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 0 | 2024-03-16T00:42:03 | 2024-03-16T02:24:13 | 2024-03-16T02:24:12 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3174",
"html_url": "https://github.com/ollama/ollama/pull/3174",
"diff_url": "https://github.com/ollama/ollama/pull/3174.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3174.patch",
"merged_at": "2024-03-16T02:24:12"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3174/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3174/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1439 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1439/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1439/comments | https://api.github.com/repos/ollama/ollama/issues/1439/events | https://github.com/ollama/ollama/issues/1439 | 2,033,191,401 | I_kwDOJ0Z1Ps55MAnp | 1,439 | Serious bug | {
"login": "m-c-frank",
"id": 61345033,
"node_id": "MDQ6VXNlcjYxMzQ1MDMz",
"avatar_url": "https://avatars.githubusercontent.com/u/61345033?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/m-c-frank",
"html_url": "https://github.com/m-c-frank",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 0 | 2023-12-08T19:20:22 | 2023-12-08T19:23:58 | 2023-12-08T19:23:58 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | (base) mcfrank@interface:~$ orca show
The following are not valid: show
Warning: Could not load keyboard geometry for :0
BadName (named color or font does not exist)
Resulting keymap file will not describe geometry
^C(base) mcfrank@interface:~$
| {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1439/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1439/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7411 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7411/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7411/comments | https://api.github.com/repos/ollama/ollama/issues/7411/events | https://github.com/ollama/ollama/pull/7411 | 2,622,487,493 | PR_kwDOJ0Z1Ps6ATg6p | 7,411 | add more tests for getting the optimal tiled canvas | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | [] | closed | false | null | [] | null | 0 | 2024-10-29T23:02:00 | 2024-10-29T23:28:04 | 2024-10-29T23:28:02 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7411",
"html_url": "https://github.com/ollama/ollama/pull/7411",
"diff_url": "https://github.com/ollama/ollama/pull/7411.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7411.patch",
"merged_at": "2024-10-29T23:28:02"
} | null | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7411/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7411/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/291 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/291/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/291/comments | https://api.github.com/repos/ollama/ollama/issues/291/events | https://github.com/ollama/ollama/issues/291 | 1,837,538,185 | I_kwDOJ0Z1Ps5thpuJ | 291 | add `ollama serve` to the quickstart doc | {
"login": "marafa-sugarcrm",
"id": 69474017,
"node_id": "MDQ6VXNlcjY5NDc0MDE3",
"avatar_url": "https://avatars.githubusercontent.com/u/69474017?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/marafa-sugarcrm",
"html_url": "https://github.com/marafa-sugarcrm",
"followers_url": "https://api... | [] | closed | false | null | [] | null | 0 | 2023-08-05T02:51:59 | 2023-08-05T02:53:17 | 2023-08-05T02:53:17 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | i downloaded using `brew` and attempted to follow the quick start doc
it is missing `ollama serve` | {
"login": "marafa-sugarcrm",
"id": 69474017,
"node_id": "MDQ6VXNlcjY5NDc0MDE3",
"avatar_url": "https://avatars.githubusercontent.com/u/69474017?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/marafa-sugarcrm",
"html_url": "https://github.com/marafa-sugarcrm",
"followers_url": "https://api... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/291/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/291/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5881 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5881/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5881/comments | https://api.github.com/repos/ollama/ollama/issues/5881/events | https://github.com/ollama/ollama/issues/5881 | 2,425,609,012 | I_kwDOJ0Z1Ps6Qk9s0 | 5,881 | Is llama 3.1 already supported (on 2.8) or should we wait another update ? | {
"login": "Qualzz",
"id": 35169816,
"node_id": "MDQ6VXNlcjM1MTY5ODE2",
"avatar_url": "https://avatars.githubusercontent.com/u/35169816?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Qualzz",
"html_url": "https://github.com/Qualzz",
"followers_url": "https://api.github.com/users/Qualzz/fo... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 20 | 2024-07-23T16:22:57 | 2024-09-02T00:15:39 | 2024-09-02T00:15:38 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
The model page seems to already exists in ollama website, but the model is clearly behaving erratically, which makes me wonder if we should wait for an update before using llama 3.1.
### OS
Windows
### GPU
Nvidia
### CPU
AMD
### Ollama version
2.8 | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5881/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5881/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1445 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1445/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1445/comments | https://api.github.com/repos/ollama/ollama/issues/1445/events | https://github.com/ollama/ollama/pull/1445 | 2,033,535,254 | PR_kwDOJ0Z1Ps5hkiJ6 | 1,445 | fix: parallel queueing race condition caused silent failure | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 1 | 2023-12-09T01:58:05 | 2023-12-09T19:14:03 | 2023-12-09T19:14:02 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1445",
"html_url": "https://github.com/ollama/ollama/pull/1445",
"diff_url": "https://github.com/ollama/ollama/pull/1445.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1445.patch",
"merged_at": "2023-12-09T19:14:02"
} | As of the most recent llama.cpp update concurrent requests had a race condition that would result in an empty response.
This was not easy to observe since the response from the llm runner subprocess was a 200 with the error {"content":"slot unavailable"} in the response stream, which just silently closed the channel... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1445/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1445/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3483 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3483/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3483/comments | https://api.github.com/repos/ollama/ollama/issues/3483/events | https://github.com/ollama/ollama/issues/3483 | 2,224,525,531 | I_kwDOJ0Z1Ps6El5Db | 3,483 | Ollama hangs on CUDA devices when running multi-modal models | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2024-04-04T05:24:22 | 2024-04-04T06:30:58 | 2024-04-04T06:30:58 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
```
Apr 04 05:15:04 gpu.us-central1-a.c.ollama.internal ollama[5042]: {"function":"launch_slot_with_data","level":"INFO","line":804,"msg":"slot is processing task","slot_id":0,"task_id":29930,"tid":"140079034640064","timestamp":1712207704}
Apr 04 05:15:04 gpu.us-central1-a.c.ollama.internal ol... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3483/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3483/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/167 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/167/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/167/comments | https://api.github.com/repos/ollama/ollama/issues/167/events | https://github.com/ollama/ollama/pull/167 | 1,816,432,518 | PR_kwDOJ0Z1Ps5WIaj5 | 167 | partial decode ggml bin for more info | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 1 | 2023-07-21T22:10:33 | 2023-08-11T00:22:41 | 2023-08-11T00:22:40 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/167",
"html_url": "https://github.com/ollama/ollama/pull/167",
"diff_url": "https://github.com/ollama/ollama/pull/167.diff",
"patch_url": "https://github.com/ollama/ollama/pull/167.patch",
"merged_at": "2023-08-11T00:22:40"
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/167/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/167/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/27 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/27/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/27/comments | https://api.github.com/repos/ollama/ollama/issues/27/events | https://github.com/ollama/ollama/issues/27 | 1,782,711,141 | I_kwDOJ0Z1Ps5qQgNl | 27 | show a loading bar for model loading | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 5667396205,
"node_id": ... | closed | false | null | [] | null | 0 | 2023-06-30T15:17:59 | 2023-09-07T11:11:10 | 2023-09-07T11:11:10 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | currently we show a spinner – even better would be to show a loader similar to the download one | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/27/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/27/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4749 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4749/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4749/comments | https://api.github.com/repos/ollama/ollama/issues/4749/events | https://github.com/ollama/ollama/issues/4749 | 2,327,508,143 | I_kwDOJ0Z1Ps6KuvSv | 4,749 | OLLAMA_MODELS not applied on initial start or on restart after upgrade on macOS | {
"login": "vernonstinebaker",
"id": 8304954,
"node_id": "MDQ6VXNlcjgzMDQ5NTQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/8304954?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/vernonstinebaker",
"html_url": "https://github.com/vernonstinebaker",
"followers_url": "https://ap... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 4 | 2024-05-31T10:32:39 | 2024-07-26T03:15:28 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | The addition of OLLAMA_MODELS is much appreciated, since it allows specifying a different location, such as an external disk, where more space might be available.
One issue, however, is if we put specify OLLAMA_MODELS in our .zshrc, for example, the .zshrc file isn't read when Ollama starts initially or when Ollama ... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4749/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4749/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/6280 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6280/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6280/comments | https://api.github.com/repos/ollama/ollama/issues/6280/events | https://github.com/ollama/ollama/issues/6280 | 2,457,405,573 | I_kwDOJ0Z1Ps6SeQiF | 6,280 | Need qwen2:math !! | {
"login": "jsrdcht",
"id": 22300065,
"node_id": "MDQ6VXNlcjIyMzAwMDY1",
"avatar_url": "https://avatars.githubusercontent.com/u/22300065?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jsrdcht",
"html_url": "https://github.com/jsrdcht",
"followers_url": "https://api.github.com/users/jsrdch... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 2 | 2024-08-09T08:32:20 | 2024-09-02T23:12:26 | 2024-09-02T23:12:26 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | https://github.com/QwenLM/Qwen2-Math | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6280/reactions",
"total_count": 4,
"+1": 4,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6280/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8303 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8303/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8303/comments | https://api.github.com/repos/ollama/ollama/issues/8303/events | https://github.com/ollama/ollama/issues/8303 | 2,768,632,208 | I_kwDOJ0Z1Ps6lBfmQ | 8,303 | Warning: could not connect to a running Ollama instance | {
"login": "rcjdut",
"id": 140966542,
"node_id": "U_kgDOCGb6jg",
"avatar_url": "https://avatars.githubusercontent.com/u/140966542?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rcjdut",
"html_url": "https://github.com/rcjdut",
"followers_url": "https://api.github.com/users/rcjdut/follower... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2025-01-04T08:22:23 | 2025-01-04T08:40:22 | 2025-01-04T08:40:22 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I manually installed Ollama and successfully ran the command 'ollama serve' in a terminal,but when i try to ran the command 'ollama -v',it showed 'Warning: could not connect to a running Ollama instance
Warning: client version is 0.5.4'
 is disappeared after `go test github.com/ollama/ollama/gpu`. Our `%buildroot` (`/usr/src/tmp/ollama-buildroot`) is inside of `TMPDIR` (`/usr/src/tmp`).
I am not completely investi... | {
"login": "joshyan1",
"id": 76125168,
"node_id": "MDQ6VXNlcjc2MTI1MTY4",
"avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joshyan1",
"html_url": "https://github.com/joshyan1",
"followers_url": "https://api.github.com/users/jos... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5129/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5129/timeline | null | completed | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.