url
stringlengths
51
54
repository_url
stringclasses
1 value
labels_url
stringlengths
65
68
comments_url
stringlengths
60
63
events_url
stringlengths
58
61
html_url
stringlengths
39
44
id
int64
1.78B
2.82B
node_id
stringlengths
18
19
number
int64
1
8.69k
title
stringlengths
1
382
user
dict
labels
listlengths
0
5
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
2
milestone
null
comments
int64
0
323
created_at
timestamp[s]
updated_at
timestamp[s]
closed_at
timestamp[s]
author_association
stringclasses
4 values
sub_issues_summary
dict
active_lock_reason
null
draft
bool
2 classes
pull_request
dict
body
stringlengths
2
118k
closed_by
dict
reactions
dict
timeline_url
stringlengths
60
63
performed_via_github_app
null
state_reason
stringclasses
4 values
is_pull_request
bool
2 classes
https://api.github.com/repos/ollama/ollama/issues/7049
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7049/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7049/comments
https://api.github.com/repos/ollama/ollama/issues/7049/events
https://github.com/ollama/ollama/issues/7049
2,557,134,462
I_kwDOJ0Z1Ps6YasZ-
7,049
ollama does not detect Quadro RTX 4000 - cuda driver library failed to get device context 801
{ "login": "mfzhsn", "id": 5251972, "node_id": "MDQ6VXNlcjUyNTE5NzI=", "avatar_url": "https://avatars.githubusercontent.com/u/5251972?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mfzhsn", "html_url": "https://github.com/mfzhsn", "followers_url": "https://api.github.com/users/mfzhsn/foll...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5755339642, "node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg...
open
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
10
2024-09-30T16:24:28
2024-11-25T19:04:01
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Hi All, I installed ollama both (on machine/docker) both with same behaviour of not detecting the GPU. Have LM Studio on the same machine which picks up GPU without any issues. ``` root@d50a3f8d8474:/# ollama run phi3.5:3.8b-mini-instruct-q2_K "" root@d50a3f8d8474:/# ollama ps NAME ...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7049/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7049/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2914
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2914/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2914/comments
https://api.github.com/repos/ollama/ollama/issues/2914/events
https://github.com/ollama/ollama/issues/2914
2,167,049,192
I_kwDOJ0Z1Ps6BKovo
2,914
ollama run starcoder2:15b
{ "login": "limaolin2017", "id": 28923721, "node_id": "MDQ6VXNlcjI4OTIzNzIx", "avatar_url": "https://avatars.githubusercontent.com/u/28923721?v=4", "gravatar_id": "", "url": "https://api.github.com/users/limaolin2017", "html_url": "https://github.com/limaolin2017", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
2
2024-03-04T14:57:44
2024-03-04T15:18:58
2024-03-04T15:12:38
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I have encountered an error with Apple silicon M1 pro: ollama run starcoder2:15b Error: Post "http://127.0.0.1:11434/api/chat": EOF
{ "login": "limaolin2017", "id": 28923721, "node_id": "MDQ6VXNlcjI4OTIzNzIx", "avatar_url": "https://avatars.githubusercontent.com/u/28923721?v=4", "gravatar_id": "", "url": "https://api.github.com/users/limaolin2017", "html_url": "https://github.com/limaolin2017", "followers_url": "https://api.github.c...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2914/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2914/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5251
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5251/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5251/comments
https://api.github.com/repos/ollama/ollama/issues/5251/events
https://github.com/ollama/ollama/issues/5251
2,369,624,310
I_kwDOJ0Z1Ps6NPZj2
5,251
how to install this in my steam deck?
{ "login": "olumolu", "id": 162728301, "node_id": "U_kgDOCbMJbQ", "avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4", "gravatar_id": "", "url": "https://api.github.com/users/olumolu", "html_url": "https://github.com/olumolu", "followers_url": "https://api.github.com/users/olumolu/foll...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
1
2024-06-24T08:46:38
2024-06-25T16:19:14
2024-06-25T16:18:58
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? cant install this in steam os 3 i think this is the issue for fedora silverblue like os opensuse aeon also ### OS Linux ### GPU AMD ### CPU AMD ### Ollama version _No response_
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5251/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5251/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7377
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7377/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7377/comments
https://api.github.com/repos/ollama/ollama/issues/7377/events
https://github.com/ollama/ollama/pull/7377
2,616,184,675
PR_kwDOJ0Z1Ps5__7QO
7,377
readme: add TextCraft to community integrations
{ "login": "suncloudsmoon", "id": 34616349, "node_id": "MDQ6VXNlcjM0NjE2MzQ5", "avatar_url": "https://avatars.githubusercontent.com/u/34616349?v=4", "gravatar_id": "", "url": "https://api.github.com/users/suncloudsmoon", "html_url": "https://github.com/suncloudsmoon", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
0
2024-10-26T22:34:06
2024-11-04T00:53:51
2024-11-04T00:53:51
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7377", "html_url": "https://github.com/ollama/ollama/pull/7377", "diff_url": "https://github.com/ollama/ollama/pull/7377.diff", "patch_url": "https://github.com/ollama/ollama/pull/7377.patch", "merged_at": "2024-11-04T00:53:51" }
Hey everyone! I've recently been working on an extension for Word that aims to be a local, privacy-friendly alternative to Microsoft 365 Copilot by utilizing Ollama as the backend. I would like to introduce TextCraft, which is an add-in for Word that seamlessly integrates essential AI tools, including text generation, ...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7377/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7377/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6713
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6713/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6713/comments
https://api.github.com/repos/ollama/ollama/issues/6713/events
https://github.com/ollama/ollama/issues/6713
2,514,609,406
I_kwDOJ0Z1Ps6V4eT-
6,713
Talking to Mistral-Nemo via OpenAI tool calling - fails
{ "login": "ChristianWeyer", "id": 888718, "node_id": "MDQ6VXNlcjg4ODcxOA==", "avatar_url": "https://avatars.githubusercontent.com/u/888718?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ChristianWeyer", "html_url": "https://github.com/ChristianWeyer", "followers_url": "https://api.github...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q...
open
false
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/...
[ { "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "htt...
null
10
2024-09-09T18:12:02
2025-01-16T14:54:23
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? With this curl command: ``` curl http://localhost:11434/v1/chat/completions \ -H "Content-Type: application/json" \ -d '{"model":"mistral-nemo:12b-instruct-2407-fp16", "messages": [ { "role": "user", "content": "What is the weather like in Boston?" } ], "to...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6713/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6713/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/4253
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4253/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4253/comments
https://api.github.com/repos/ollama/ollama/issues/4253/events
https://github.com/ollama/ollama/issues/4253
2,284,818,276
I_kwDOJ0Z1Ps6IL49k
4,253
A repeatable hang issue on Linux - dual radeon
{ "login": "eliranwong", "id": 25262722, "node_id": "MDQ6VXNlcjI1MjYyNzIy", "avatar_url": "https://avatars.githubusercontent.com/u/25262722?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eliranwong", "html_url": "https://github.com/eliranwong", "followers_url": "https://api.github.com/use...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6433346500, "node_id": "LA_kwDOJ0Z1Ps8AAAABf3UTxA...
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
5
2024-05-08T06:49:58
2024-05-09T22:30:37
2024-05-09T22:08:58
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Experience a hang issue consistently. Device information: OS: Ubuntu, CPU: AMD Threadripper [AMD Ryzen Threadripper 7960X, 24 Cores, 48 Threads, 4.2GHz Base, 5.3GHz Turbo], Memory: 256GB RAM, Two GPUs: AMD RX 7900XTX + AMD RX 7900XTX To reproduce the hang issue: 1. ollama run command...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4253/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4253/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3000
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3000/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3000/comments
https://api.github.com/repos/ollama/ollama/issues/3000/events
https://github.com/ollama/ollama/issues/3000
2,175,452,464
I_kwDOJ0Z1Ps6BqsUw
3,000
Server hangs with no responsewhen running `gemma`
{ "login": "songsh", "id": 2272252, "node_id": "MDQ6VXNlcjIyNzIyNTI=", "avatar_url": "https://avatars.githubusercontent.com/u/2272252?v=4", "gravatar_id": "", "url": "https://api.github.com/users/songsh", "html_url": "https://github.com/songsh", "followers_url": "https://api.github.com/users/songsh/foll...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q...
closed
false
null
[]
null
7
2024-03-08T07:29:25
2024-05-02T22:33:21
2024-05-02T22:33:21
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
serve dead, how i to check the problem ,where is logs
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3000/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3000/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7798
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7798/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7798/comments
https://api.github.com/repos/ollama/ollama/issues/7798/events
https://github.com/ollama/ollama/issues/7798
2,683,215,299
I_kwDOJ0Z1Ps6f7p3D
7,798
Is this a bug? (2GB model -> up to 20GB pagefile)
{ "login": "sebkont", "id": 189359503, "node_id": "U_kgDOC0lljw", "avatar_url": "https://avatars.githubusercontent.com/u/189359503?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sebkont", "html_url": "https://github.com/sebkont", "followers_url": "https://api.github.com/users/sebkont/foll...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
8
2024-11-22T13:22:13
2024-12-02T15:36:27
2024-12-02T15:36:27
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? My GPU is old (GTX 1070) with 8GB, but should still be enough for running a model based on Phi 3 Mini? [This one ](https://huggingface.co/v8karlo/UNCENSORED-Phi-3-mini-4k-geminified-Q4_K_M-GGUF) Unfortunately what happens is 'ollama ps' says 20 GB 63%/37% CPU/GPU + C:/ drive instantly gets ...
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7798/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7798/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5635
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5635/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5635/comments
https://api.github.com/repos/ollama/ollama/issues/5635/events
https://github.com/ollama/ollama/issues/5635
2,403,793,783
I_kwDOJ0Z1Ps6PRvt3
5,635
ollama not use all GPUs
{ "login": "mavershang", "id": 8919917, "node_id": "MDQ6VXNlcjg5MTk5MTc=", "avatar_url": "https://avatars.githubusercontent.com/u/8919917?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mavershang", "html_url": "https://github.com/mavershang", "followers_url": "https://api.github.com/users...
[ { "id": 5667396191, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aXw", "url": "https://api.github.com/repos/ollama/ollama/labels/documentation", "name": "documentation", "color": "0075ca", "default": true, "description": "Improvements or additions to documentation" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
2
2024-07-11T18:12:48
2024-07-29T21:25:42
2024-07-29T21:25:42
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I ran ollama on a server with 4x A100. It only uses 1 of them. Is there some setting need to be changed? Thanks ![image](https://github.com/ollama/ollama/assets/8919917/8a49b4a4-fd85-4715-8ff0-1ea0c1276939) ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version ...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5635/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5635/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/49
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/49/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/49/comments
https://api.github.com/repos/ollama/ollama/issues/49/events
https://github.com/ollama/ollama/pull/49
1,792,432,386
PR_kwDOJ0Z1Ps5U26Op
49
Go run
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
1
2023-07-06T23:03:39
2023-07-07T00:19:03
2023-07-07T00:18:58
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/49", "html_url": "https://github.com/ollama/ollama/pull/49", "diff_url": "https://github.com/ollama/ollama/pull/49.diff", "patch_url": "https://github.com/ollama/ollama/pull/49.patch", "merged_at": "2023-07-07T00:18:58" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/49/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/49/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5098
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5098/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5098/comments
https://api.github.com/repos/ollama/ollama/issues/5098/events
https://github.com/ollama/ollama/pull/5098
2,357,217,804
PR_kwDOJ0Z1Ps5yr-W8
5,098
feat: support setting the KV cache quant type
{ "login": "sammcj", "id": 862951, "node_id": "MDQ6VXNlcjg2Mjk1MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sammcj", "html_url": "https://github.com/sammcj", "followers_url": "https://api.github.com/users/sammcj/follow...
[]
closed
false
null
[]
null
3
2024-06-17T12:24:22
2024-06-29T01:17:07
2024-06-28T21:50:53
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
true
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5098", "html_url": "https://github.com/ollama/ollama/pull/5098", "diff_url": "https://github.com/ollama/ollama/pull/5098.diff", "patch_url": "https://github.com/ollama/ollama/pull/5098.patch", "merged_at": null }
WIP Testing adding configuration to allow setting the KV cache type re: #5091 --- - Allow setting the KV cache type in the env and params. - Allow setting flashattention in params (as well as the existing env).
{ "login": "sammcj", "id": 862951, "node_id": "MDQ6VXNlcjg2Mjk1MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sammcj", "html_url": "https://github.com/sammcj", "followers_url": "https://api.github.com/users/sammcj/follow...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5098/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 2, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5098/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/637
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/637/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/637/comments
https://api.github.com/repos/ollama/ollama/issues/637/events
https://github.com/ollama/ollama/pull/637
1,918,203,306
PR_kwDOJ0Z1Ps5bex-l
637
windows runner fixes
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
[]
closed
false
null
[]
null
0
2023-09-28T20:13:13
2023-09-29T15:47:56
2023-09-29T15:47:55
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/637", "html_url": "https://github.com/ollama/ollama/pull/637", "diff_url": "https://github.com/ollama/ollama/pull/637.diff", "patch_url": "https://github.com/ollama/ollama/pull/637.patch", "merged_at": "2023-09-29T15:47:55" }
- use filepath for runner files - get embedded files with unix filepath - runner is only available is embedded directories have files
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/637/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/637/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2490
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2490/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2490/comments
https://api.github.com/repos/ollama/ollama/issues/2490/events
https://github.com/ollama/ollama/issues/2490
2,134,150,180
I_kwDOJ0Z1Ps5_NIwk
2,490
[Question] Do not offload to CPU RAM
{ "login": "freQuensy23-coder", "id": 64750224, "node_id": "MDQ6VXNlcjY0NzUwMjI0", "avatar_url": "https://avatars.githubusercontent.com/u/64750224?v=4", "gravatar_id": "", "url": "https://api.github.com/users/freQuensy23-coder", "html_url": "https://github.com/freQuensy23-coder", "followers_url": "https...
[]
closed
false
null
[]
null
3
2024-02-14T11:32:17
2024-03-16T19:39:26
2024-03-11T18:28:34
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
By default, after some time of inactivity, ollama will automatically be offloaded from GPU memory, that caused some latency, especially to large models)
{ "login": "hoyyeva", "id": 63033505, "node_id": "MDQ6VXNlcjYzMDMzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoyyeva", "html_url": "https://github.com/hoyyeva", "followers_url": "https://api.github.com/users/hoyyev...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2490/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2490/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/238
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/238/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/238/comments
https://api.github.com/repos/ollama/ollama/issues/238/events
https://github.com/ollama/ollama/issues/238
1,827,345,140
I_kwDOJ0Z1Ps5s6xL0
238
Ability to download LLAMA2 7b 32k context
{ "login": "jlarmstrongiv", "id": 20903247, "node_id": "MDQ6VXNlcjIwOTAzMjQ3", "avatar_url": "https://avatars.githubusercontent.com/u/20903247?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jlarmstrongiv", "html_url": "https://github.com/jlarmstrongiv", "followers_url": "https://api.githu...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 5789807732, "node_id": ...
closed
false
null
[]
null
2
2023-07-29T06:23:31
2023-12-04T19:02:13
2023-12-04T19:02:12
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
- https://together.ai/blog/llama-2-7b-32k - https://github.com/togethercomputer/OpenChatKit - https://huggingface.co/togethercomputer/LLaMA-2-7B-32K
{ "login": "technovangelist", "id": 633681, "node_id": "MDQ6VXNlcjYzMzY4MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4", "gravatar_id": "", "url": "https://api.github.com/users/technovangelist", "html_url": "https://github.com/technovangelist", "followers_url": "https://api.git...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/238/reactions", "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/238/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1365
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1365/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1365/comments
https://api.github.com/repos/ollama/ollama/issues/1365/events
https://github.com/ollama/ollama/issues/1365
2,022,633,560
I_kwDOJ0Z1Ps54jvBY
1,365
llama_print_timings have disappeared from the logs.
{ "login": "madsamjp", "id": 49611363, "node_id": "MDQ6VXNlcjQ5NjExMzYz", "avatar_url": "https://avatars.githubusercontent.com/u/49611363?v=4", "gravatar_id": "", "url": "https://api.github.com/users/madsamjp", "html_url": "https://github.com/madsamjp", "followers_url": "https://api.github.com/users/mad...
[]
closed
false
null
[]
null
3
2023-12-03T17:26:16
2024-01-20T00:18:22
2024-01-20T00:18:21
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
In a previous version of Ollama, following the logs (on Linux using `journalctl -t ollama -f`) would give helpful information after the model has finished with its response (such as tokens per second). e.g. this: ``` Dec 03 14:58:42 osm-server ollama[20658]: llama server listening at http://127.0.0.1:54457 Dec 03...
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1365/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1365/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6356
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6356/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6356/comments
https://api.github.com/repos/ollama/ollama/issues/6356/events
https://github.com/ollama/ollama/issues/6356
2,465,338,496
I_kwDOJ0Z1Ps6S8hSA
6,356
AMD Multiple GPU support
{ "login": "VitalickS", "id": 10177561, "node_id": "MDQ6VXNlcjEwMTc3NTYx", "avatar_url": "https://avatars.githubusercontent.com/u/10177561?v=4", "gravatar_id": "", "url": "https://api.github.com/users/VitalickS", "html_url": "https://github.com/VitalickS", "followers_url": "https://api.github.com/users/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg...
open
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
6
2024-08-14T09:25:57
2024-10-16T00:15:13
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### Hi, I think the current AMD ROCm doesn’t work well with multiple video cards. I have an XTX 7900 (24GB) and an XT 7900 (20GB). My processor also has a small integrated GPU, but that shouldn’t be a problem. When I try to load the model llama3.1:70b (39GB): 1. It doesn’t crash, but it has an infinite load ti...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6356/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6356/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/8661
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8661/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8661/comments
https://api.github.com/repos/ollama/ollama/issues/8661/events
https://github.com/ollama/ollama/issues/8661
2,818,282,626
I_kwDOJ0Z1Ps6n-5SC
8,661
Will Ollama run on the NPU(ANE) of Apple M silicon?
{ "login": "imJack6", "id": 58357771, "node_id": "MDQ6VXNlcjU4MzU3Nzcx", "avatar_url": "https://avatars.githubusercontent.com/u/58357771?v=4", "gravatar_id": "", "url": "https://api.github.com/users/imJack6", "html_url": "https://github.com/imJack6", "followers_url": "https://api.github.com/users/imJack...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
0
2025-01-29T13:50:08
2025-01-29T13:50:08
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
RT
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8661/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8661/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/3119
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3119/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3119/comments
https://api.github.com/repos/ollama/ollama/issues/3119/events
https://github.com/ollama/ollama/issues/3119
2,184,596,868
I_kwDOJ0Z1Ps6CNk2E
3,119
Tensor `token_embed.weight` has wrong shape
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
4
2024-03-13T18:01:27
2024-03-13T20:30:47
2024-03-13T18:21:34
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
![image](https://github.com/ollama/ollama/assets/5853428/543d8714-16de-421f-b669-2afd9b87fe81) split from #2753
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3119/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3119/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4932
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4932/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4932/comments
https://api.github.com/repos/ollama/ollama/issues/4932/events
https://github.com/ollama/ollama/issues/4932
2,341,666,209
I_kwDOJ0Z1Ps6Lkv2h
4,932
Cant see installed models
{ "login": "ahgsql", "id": 35695543, "node_id": "MDQ6VXNlcjM1Njk1NTQz", "avatar_url": "https://avatars.githubusercontent.com/u/35695543?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ahgsql", "html_url": "https://github.com/ahgsql", "followers_url": "https://api.github.com/users/ahgsql/fo...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q...
closed
false
null
[]
null
4
2024-06-08T13:51:57
2024-08-10T05:41:42
2024-08-09T23:51:33
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I have 7 models installed and was using them till yesterday. But now it re-tries to download them, even i have all manifests files and my blobs folder is over 18 GB. After shutdown and restart of WSL, ollama is not running and i m trying with ollama serve cmd ### OS Windows ### ...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4932/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4932/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5323
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5323/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5323/comments
https://api.github.com/repos/ollama/ollama/issues/5323/events
https://github.com/ollama/ollama/issues/5323
2,377,959,722
I_kwDOJ0Z1Ps6NvMkq
5,323
Weird output with any typos in accepted commands
{ "login": "yoshimario", "id": 8993080, "node_id": "MDQ6VXNlcjg5OTMwODA=", "avatar_url": "https://avatars.githubusercontent.com/u/8993080?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yoshimario", "html_url": "https://github.com/yoshimario", "followers_url": "https://api.github.com/users...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
3
2024-06-27T11:59:02
2024-06-27T22:31:00
2024-06-27T22:29:45
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Problem is program outputs weird text if the program experiences commands that are not in the program list. This should result in an error message instead of a never ending loop of erraneous output. This should be handled better instead of forcing the process to close using ctr + c. `>>> e...
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjha...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5323/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5323/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4153
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4153/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4153/comments
https://api.github.com/repos/ollama/ollama/issues/4153/events
https://github.com/ollama/ollama/pull/4153
2,279,118,097
PR_kwDOJ0Z1Ps5ujE7R
4,153
Add GPU usage
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
2
2024-05-04T17:10:08
2024-05-08T23:39:14
2024-05-08T23:39:11
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4153", "html_url": "https://github.com/ollama/ollama/pull/4153", "diff_url": "https://github.com/ollama/ollama/pull/4153.diff", "patch_url": "https://github.com/ollama/ollama/pull/4153.patch", "merged_at": "2024-05-08T23:39:11" }
Help users understand how much of the model fit into their GPU without having to resort to inspecting the server log A few examples from different systems and models ``` eval rate: 4.40 tokens/s gpu usage: 1 GPU (14/27 layers) 3.2 GB (2.0 GB GPU) eval rate: 6.64 tokens/s gpu...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4153/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4153/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7815
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7815/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7815/comments
https://api.github.com/repos/ollama/ollama/issues/7815/events
https://github.com/ollama/ollama/issues/7815
2,687,656,860
I_kwDOJ0Z1Ps6gMmOc
7,815
Any fine-tuning ways?
{ "login": "Niifuji", "id": 111742025, "node_id": "U_kgDOBqkMSQ", "avatar_url": "https://avatars.githubusercontent.com/u/111742025?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Niifuji", "html_url": "https://github.com/Niifuji", "followers_url": "https://api.github.com/users/Niifuji/foll...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2024-11-24T12:14:14
2024-12-23T07:57:42
2024-12-23T07:57:42
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
As I mentioned in the title, I want to "continuously" fine-tune a pre-trained model with my custom dataset and explore adding some "emotion" to it (not sure why this idea came to mind). If you have any features or suggestions for this, I would appreciate your input.
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7815/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7815/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5187
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5187/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5187/comments
https://api.github.com/repos/ollama/ollama/issues/5187/events
https://github.com/ollama/ollama/pull/5187
2,364,767,566
PR_kwDOJ0Z1Ps5zF1G4
5,187
fix: skip os.removeAll() in assets.go if no PID
{ "login": "joshyan1", "id": 76125168, "node_id": "MDQ6VXNlcjc2MTI1MTY4", "avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joshyan1", "html_url": "https://github.com/joshyan1", "followers_url": "https://api.github.com/users/jos...
[]
closed
false
null
[]
null
1
2024-06-20T15:48:16
2024-06-20T15:53:26
2024-06-20T15:49:39
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5187", "html_url": "https://github.com/ollama/ollama/pull/5187", "diff_url": "https://github.com/ollama/ollama/pull/5187.diff", "patch_url": "https://github.com/ollama/ollama/pull/5187.patch", "merged_at": null }
we accidentally deleted every directory in $TMPDIR that was in the form "ollama*". Used errorcheck with PID to ensure directory is ours before deleting. Resolves: https://github.com/ollama/ollama/issues/5129
{ "login": "joshyan1", "id": 76125168, "node_id": "MDQ6VXNlcjc2MTI1MTY4", "avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joshyan1", "html_url": "https://github.com/joshyan1", "followers_url": "https://api.github.com/users/jos...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5187/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5187/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4131
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4131/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4131/comments
https://api.github.com/repos/ollama/ollama/issues/4131/events
https://github.com/ollama/ollama/issues/4131
2,278,060,992
I_kwDOJ0Z1Ps6HyHPA
4,131
Error "timed out waiting for llama runner to start: " on larger models.
{ "login": "CalvesGEH", "id": 42101564, "node_id": "MDQ6VXNlcjQyMTAxNTY0", "avatar_url": "https://avatars.githubusercontent.com/u/42101564?v=4", "gravatar_id": "", "url": "https://api.github.com/users/CalvesGEH", "html_url": "https://github.com/CalvesGEH", "followers_url": "https://api.github.com/users/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
45
2024-05-03T16:45:32
2024-12-18T05:50:45
2024-07-03T23:28:05
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I just setup Ollama on a fresh machine and am running into an issue starting Ollama on larger models. I am running Ubuntu 22.04.4 LTS with 2 Nvidia Tesla P40 GPUs with Driver Version: 535.161.08 and CUDA Version: 12.2. Small 8b models work great and have no issues but when I try something ...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4131/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4131/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5454
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5454/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5454/comments
https://api.github.com/repos/ollama/ollama/issues/5454/events
https://github.com/ollama/ollama/issues/5454
2,387,738,883
I_kwDOJ0Z1Ps6OUgED
5,454
When can we perform function calls like OpenAI?
{ "login": "qq1005894049", "id": 48113255, "node_id": "MDQ6VXNlcjQ4MTEzMjU1", "avatar_url": "https://avatars.githubusercontent.com/u/48113255?v=4", "gravatar_id": "", "url": "https://api.github.com/users/qq1005894049", "html_url": "https://github.com/qq1005894049", "followers_url": "https://api.github.c...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-07-03T05:57:43
2024-07-30T17:25:18
2024-07-30T17:25:18
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
<img width="1341" alt="image" src="https://github.com/ollama/ollama/assets/48113255/70ca615f-aae2-4b48-bd8f-c913b2ede23e">
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjha...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5454/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5454/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8110
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8110/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8110/comments
https://api.github.com/repos/ollama/ollama/issues/8110/events
https://github.com/ollama/ollama/issues/8110
2,741,302,826
I_kwDOJ0Z1Ps6jZPYq
8,110
Support llama.cpp's Control Vector Functionality
{ "login": "amyb-asu", "id": 156008468, "node_id": "U_kgDOCUyAFA", "avatar_url": "https://avatars.githubusercontent.com/u/156008468?v=4", "gravatar_id": "", "url": "https://api.github.com/users/amyb-asu", "html_url": "https://github.com/amyb-asu", "followers_url": "https://api.github.com/users/amyb-asu/...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
3
2024-12-16T04:30:35
2024-12-18T00:30:05
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
llama.cpp added support for control vectors a while ago https://github.com/ggerganov/llama.cpp/pull/5970 They can be loaded via `llama_control_vector_load` and `llama_control_vector_apply` which can take a vector in the form of a `.gguf` https://github.com/ollama/ollama/blob/main/llama/common.h#L645 https://gith...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8110/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8110/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/64
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/64/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/64/comments
https://api.github.com/repos/ollama/ollama/issues/64/events
https://github.com/ollama/ollama/pull/64
1,796,593,917
PR_kwDOJ0Z1Ps5VE8wr
64
Do not seg fault on client disconnect
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
[]
closed
false
null
[]
null
0
2023-07-10T11:46:30
2023-07-11T14:19:33
2023-07-10T15:00:45
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/64", "html_url": "https://github.com/ollama/ollama/pull/64", "diff_url": "https://github.com/ollama/ollama/pull/64.diff", "patch_url": "https://github.com/ollama/ollama/pull/64.patch", "merged_at": "2023-07-10T15:00:44" }
This was nicer to fix on the revised `b2` branch, so this is a pull request into that simplified change
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/64/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/64/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3494
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3494/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3494/comments
https://api.github.com/repos/ollama/ollama/issues/3494/events
https://github.com/ollama/ollama/pull/3494
2,226,051,077
PR_kwDOJ0Z1Ps5rvMkF
3,494
Fail fast if mingw missing on windows
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
1
2024-04-04T16:52:06
2024-04-04T17:15:44
2024-04-04T17:15:40
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3494", "html_url": "https://github.com/ollama/ollama/pull/3494", "diff_url": "https://github.com/ollama/ollama/pull/3494.diff", "patch_url": "https://github.com/ollama/ollama/pull/3494.patch", "merged_at": "2024-04-04T17:15:40" }
null
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3494/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3494/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4974
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4974/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4974/comments
https://api.github.com/repos/ollama/ollama/issues/4974/events
https://github.com/ollama/ollama/issues/4974
2,345,961,529
I_kwDOJ0Z1Ps6L1Ig5
4,974
panic: runtime error: invalid memory address or nil pointer dereference
{ "login": "wywself", "id": 8843053, "node_id": "MDQ6VXNlcjg4NDMwNTM=", "avatar_url": "https://avatars.githubusercontent.com/u/8843053?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wywself", "html_url": "https://github.com/wywself", "followers_url": "https://api.github.com/users/wywself/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-06-11T09:52:18
2024-06-12T02:04:31
2024-06-12T02:04:31
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I am using Tesla M60, which is on the GPU card support list. However, when I execute the following command to start the model, an error is reported as follows. ``` # ollama run qwen:7b Error: Post "http://127.0.0.1:11434/api/chat": EOF ``` The log is as follows: ![image](https://github.co...
{ "login": "wywself", "id": 8843053, "node_id": "MDQ6VXNlcjg4NDMwNTM=", "avatar_url": "https://avatars.githubusercontent.com/u/8843053?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wywself", "html_url": "https://github.com/wywself", "followers_url": "https://api.github.com/users/wywself/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4974/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4974/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/860
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/860/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/860/comments
https://api.github.com/repos/ollama/ollama/issues/860/events
https://github.com/ollama/ollama/issues/860
1,954,877,675
I_kwDOJ0Z1Ps50hRDr
860
bug: the `-v` for `--version` should be capital `-V`
{ "login": "coolaj86", "id": 122831, "node_id": "MDQ6VXNlcjEyMjgzMQ==", "avatar_url": "https://avatars.githubusercontent.com/u/122831?v=4", "gravatar_id": "", "url": "https://api.github.com/users/coolaj86", "html_url": "https://github.com/coolaj86", "followers_url": "https://api.github.com/users/coolaj8...
[]
closed
false
null
[]
null
2
2023-10-20T19:05:59
2023-10-20T21:55:16
2023-10-20T19:38:26
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I just noticed that there's a typo in the shorthand of the `--version` flag. big `-V` is for `--version` (little `-v` is for `--verbose`)
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/860/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/860/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7010
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7010/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7010/comments
https://api.github.com/repos/ollama/ollama/issues/7010/events
https://github.com/ollama/ollama/pull/7010
2,553,850,810
PR_kwDOJ0Z1Ps58-L8j
7,010
llama: Fix directory for conditional flash attention patch
{ "login": "jessegross", "id": 6468499, "node_id": "MDQ6VXNlcjY0Njg0OTk=", "avatar_url": "https://avatars.githubusercontent.com/u/6468499?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jessegross", "html_url": "https://github.com/jessegross", "followers_url": "https://api.github.com/users...
[]
closed
false
null
[]
null
1
2024-09-27T22:52:40
2024-10-10T21:38:26
2024-09-30T19:41:33
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7010", "html_url": "https://github.com/ollama/ollama/pull/7010", "diff_url": "https://github.com/ollama/ollama/pull/7010.diff", "patch_url": "https://github.com/ollama/ollama/pull/7010.patch", "merged_at": null }
Patches are against the llama.cpp directory structure, otherwise sync.sh can't apply them.
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7010/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7010/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/523
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/523/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/523/comments
https://api.github.com/repos/ollama/ollama/issues/523/events
https://github.com/ollama/ollama/issues/523
1,894,010,611
I_kwDOJ0Z1Ps5w5E7z
523
LLM falcon:text infinity loop
{ "login": "dcasota", "id": 14890243, "node_id": "MDQ6VXNlcjE0ODkwMjQz", "avatar_url": "https://avatars.githubusercontent.com/u/14890243?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dcasota", "html_url": "https://github.com/dcasota", "followers_url": "https://api.github.com/users/dcasot...
[]
closed
false
null
[]
null
3
2023-09-13T08:26:09
2023-09-13T14:32:53
2023-09-13T13:56:46
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi, I was trying to run falcon, but it responds ... weired. Setup recipe. ``` git clone https://github.com/jmorganca/ollama cd .\ollama mkdir ..\.ollama go generate .\... go build . ``` Then, start the server component of ollama. `start "Ollama server component" ollama.exe serve` Download the select...
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/523/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/523/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3441
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3441/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3441/comments
https://api.github.com/repos/ollama/ollama/issues/3441/events
https://github.com/ollama/ollama/issues/3441
2,218,582,612
I_kwDOJ0Z1Ps6EPOJU
3,441
Download/Archive and move models offline
{ "login": "Solomin0", "id": 37559666, "node_id": "MDQ6VXNlcjM3NTU5NjY2", "avatar_url": "https://avatars.githubusercontent.com/u/37559666?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Solomin0", "html_url": "https://github.com/Solomin0", "followers_url": "https://api.github.com/users/Sol...
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
null
[]
null
10
2024-04-01T16:09:15
2024-10-21T08:22:19
2024-05-10T20:18:11
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What are you trying to do? I would like to be able to move ollama models between environments that are offline. There does not seem to be a supported official way to do this. ### How should we solve this? An ollama archive command would be great! Then the user could just ollama pull from the path the arc...
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3441/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3441/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2051
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2051/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2051/comments
https://api.github.com/repos/ollama/ollama/issues/2051/events
https://github.com/ollama/ollama/issues/2051
2,088,450,843
I_kwDOJ0Z1Ps58ezsb
2,051
Mixtral : How to connect to the Web
{ "login": "ymoymo", "id": 10183941, "node_id": "MDQ6VXNlcjEwMTgzOTQx", "avatar_url": "https://avatars.githubusercontent.com/u/10183941?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ymoymo", "html_url": "https://github.com/ymoymo", "followers_url": "https://api.github.com/users/ymoymo/fo...
[]
closed
false
null
[]
null
2
2024-01-18T14:50:38
2024-03-11T18:13:47
2024-03-11T18:13:47
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi, I want to modify scipt to get this service, but I can't find the docker id or name that run Mixtral instance. sudo docker ps return nothing while Mixtral is running. Is there somthing I don't understand ? Thx for any help. Linux Pop Os
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2051/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2051/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3651
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3651/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3651/comments
https://api.github.com/repos/ollama/ollama/issues/3651/events
https://github.com/ollama/ollama/pull/3651
2,243,436,782
PR_kwDOJ0Z1Ps5sqeHc
3,651
If OLLAMA_CONTAINER_MANAGER is set, only install NVIDIA drivers
{ "login": "ericcurtin", "id": 1694275, "node_id": "MDQ6VXNlcjE2OTQyNzU=", "avatar_url": "https://avatars.githubusercontent.com/u/1694275?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ericcurtin", "html_url": "https://github.com/ericcurtin", "followers_url": "https://api.github.com/users...
[]
open
false
null
[]
null
2
2024-04-15T11:33:41
2024-04-16T08:22:58
null
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3651", "html_url": "https://github.com/ollama/ollama/pull/3651", "diff_url": "https://github.com/ollama/ollama/pull/3651.diff", "patch_url": "https://github.com/ollama/ollama/pull/3651.patch", "merged_at": null }
If installing for containerized environment, we should not have to install ollama binary, configure systemd, install rocm, etc. Intended to be run like this: curl -fsSL https://ollama.com/install.sh | OLLAMA_CONTAINER_MANAGER=podman sh
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3651/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3651/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3142
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3142/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3142/comments
https://api.github.com/repos/ollama/ollama/issues/3142/events
https://github.com/ollama/ollama/pull/3142
2,186,567,894
PR_kwDOJ0Z1Ps5ppF8X
3,142
doc: faq gpu compatibility
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
[]
closed
false
null
[]
null
3
2024-03-14T14:47:11
2024-03-21T09:21:35
2024-03-21T09:21:34
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3142", "html_url": "https://github.com/ollama/ollama/pull/3142", "diff_url": "https://github.com/ollama/ollama/pull/3142.diff", "patch_url": "https://github.com/ollama/ollama/pull/3142.patch", "merged_at": "2024-03-21T09:21:34" }
Add some information about GPU compatibility to the FAQs.
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3142/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3142/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/13
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/13/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/13/comments
https://api.github.com/repos/ollama/ollama/issues/13/events
https://github.com/ollama/ollama/pull/13
1,779,612,155
PR_kwDOJ0Z1Ps5ULXaq
13
update development.md
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
0
2023-06-28T19:30:01
2023-06-28T19:44:59
2023-06-28T19:44:56
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/13", "html_url": "https://github.com/ollama/ollama/pull/13", "diff_url": "https://github.com/ollama/ollama/pull/13.diff", "patch_url": "https://github.com/ollama/ollama/pull/13.patch", "merged_at": "2023-06-28T19:44:56" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/13/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/13/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1838
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1838/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1838/comments
https://api.github.com/repos/ollama/ollama/issues/1838/events
https://github.com/ollama/ollama/issues/1838
2,069,059,120
I_kwDOJ0Z1Ps57U1Yw
1,838
Cuda Error with 2GB VRAM: `Error: Post "http://127.0.0.1:11434/api/generate": EOF`
{ "login": "falaimo", "id": 29931008, "node_id": "MDQ6VXNlcjI5OTMxMDA4", "avatar_url": "https://avatars.githubusercontent.com/u/29931008?v=4", "gravatar_id": "", "url": "https://api.github.com/users/falaimo", "html_url": "https://github.com/falaimo", "followers_url": "https://api.github.com/users/falaim...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
8
2024-01-07T09:38:40
2024-01-08T21:42:01
2024-01-08T21:42:01
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hello everyone, in Ollama version 0.1.18, I'm encountering the error "Error: Post "http://127.0.0.1:11434/api/generate": EOF" when starting Ollama with any model. I think it depends of cuda... [logs_ollama.txt](https://github.com/jmorganca/ollama/files/13852832/logs_ollama.txt)
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1838/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1838/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6061
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6061/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6061/comments
https://api.github.com/repos/ollama/ollama/issues/6061/events
https://github.com/ollama/ollama/issues/6061
2,436,319,840
I_kwDOJ0Z1Ps6RN0pg
6,061
[Feature Request] Force function calling for a model
{ "login": "mak448a", "id": 94062293, "node_id": "U_kgDOBZtG1Q", "avatar_url": "https://avatars.githubusercontent.com/u/94062293?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mak448a", "html_url": "https://github.com/mak448a", "followers_url": "https://api.github.com/users/mak448a/follow...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
4
2024-07-29T20:43:00
2025-01-06T07:17:25
2025-01-06T07:17:24
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Not sure if this belongs in ollama-python or here, but I'll open it here. Could you add a way to use function calling on any model, or is this something that the model itself has to support?
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6061/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6061/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2096
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2096/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2096/comments
https://api.github.com/repos/ollama/ollama/issues/2096/events
https://github.com/ollama/ollama/issues/2096
2,090,734,716
I_kwDOJ0Z1Ps58nhR8
2,096
How is Tinyllama on Ollama trained?
{ "login": "oliverbob", "id": 23272429, "node_id": "MDQ6VXNlcjIzMjcyNDI5", "avatar_url": "https://avatars.githubusercontent.com/u/23272429?v=4", "gravatar_id": "", "url": "https://api.github.com/users/oliverbob", "html_url": "https://github.com/oliverbob", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
9
2024-01-19T14:58:44
2024-03-18T20:47:59
2024-02-20T22:51:52
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi everyone, as always, thank you for the great work you have done with this project for the good of humanity. I have tried importing gguf file using tintyllama on huggingface, but when I chat with it using ollama, it returns gibberish talk. But when I download the one from Ollama with ollama pull/run tinyllama, it wor...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2096/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2096/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3765
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3765/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3765/comments
https://api.github.com/repos/ollama/ollama/issues/3765/events
https://github.com/ollama/ollama/issues/3765
2,253,970,292
I_kwDOJ0Z1Ps6GWNt0
3,765
CUDA error: out of memory - other VRAM consumers not detected in available memory
{ "login": "martinus", "id": 14386, "node_id": "MDQ6VXNlcjE0Mzg2", "avatar_url": "https://avatars.githubusercontent.com/u/14386?v=4", "gravatar_id": "", "url": "https://api.github.com/users/martinus", "html_url": "https://github.com/martinus", "followers_url": "https://api.github.com/users/martinus/foll...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5755339642, "node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg...
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
18
2024-04-19T20:47:24
2024-06-14T22:35:02
2024-06-14T22:35:02
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? When I try the llama3 model I get out of memory errors. I have 64GB of RAM and 24GB on the GPU. ``` ❯ ollama run llama3:70b-instruct-q2_K --verbose "write a constexpr GCD that is not recursive in C++17" Error: an unknown error was encountered while running the model CUDA error: out of memo...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3765/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3765/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8319
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8319/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8319/comments
https://api.github.com/repos/ollama/ollama/issues/8319/events
https://github.com/ollama/ollama/pull/8319
2,770,591,352
PR_kwDOJ0Z1Ps6G0bib
8,319
Add Safetensor Conversion for Granite Models
{ "login": "alex-jw-brooks", "id": 10740300, "node_id": "MDQ6VXNlcjEwNzQwMzAw", "avatar_url": "https://avatars.githubusercontent.com/u/10740300?v=4", "gravatar_id": "", "url": "https://api.github.com/users/alex-jw-brooks", "html_url": "https://github.com/alex-jw-brooks", "followers_url": "https://api.gi...
[]
open
false
null
[]
null
0
2025-01-06T12:48:04
2025-01-16T04:20:40
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8319", "html_url": "https://github.com/ollama/ollama/pull/8319", "diff_url": "https://github.com/ollama/ollama/pull/8319.diff", "patch_url": "https://github.com/ollama/ollama/pull/8319.patch", "merged_at": null }
This PR fixes the unrecognized architecture for converting Granite Models (`GraniteForCausalLM`) for use from safetensors.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8319/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8319/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1430
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1430/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1430/comments
https://api.github.com/repos/ollama/ollama/issues/1430/events
https://github.com/ollama/ollama/issues/1430
2,031,816,063
I_kwDOJ0Z1Ps55Gw1_
1,430
coda error 222 after building
{ "login": "rhettg", "id": 50074, "node_id": "MDQ6VXNlcjUwMDc0", "avatar_url": "https://avatars.githubusercontent.com/u/50074?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rhettg", "html_url": "https://github.com/rhettg", "followers_url": "https://api.github.com/users/rhettg/followers", ...
[ { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
3
2023-12-08T02:14:32
2024-02-01T23:15:41
2024-02-01T23:15:41
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
This might be a llama.cpp question, but I'm struggling to get Ollama to work when I build it myself. The release builds work fine for me: ```console $ sudo -u ollama /usr/bin/ollama serve 2023/12/07 17:52:41 images.go:779: total blobs: 10 2023/12/07 17:52:41 images.go:786: total unused blobs removed: 0 2023/...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1430/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1430/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8662
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8662/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8662/comments
https://api.github.com/repos/ollama/ollama/issues/8662/events
https://github.com/ollama/ollama/pull/8662
2,818,376,666
PR_kwDOJ0Z1Ps6JXtvk
8,662
Update README.md Adding DeepSeek to the table of models
{ "login": "teymuur", "id": 64795612, "node_id": "MDQ6VXNlcjY0Nzk1NjEy", "avatar_url": "https://avatars.githubusercontent.com/u/64795612?v=4", "gravatar_id": "", "url": "https://api.github.com/users/teymuur", "html_url": "https://github.com/teymuur", "followers_url": "https://api.github.com/users/teymuu...
[]
closed
false
null
[]
null
0
2025-01-29T14:25:25
2025-01-29T14:33:35
2025-01-29T14:33:28
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8662", "html_url": "https://github.com/ollama/ollama/pull/8662", "diff_url": "https://github.com/ollama/ollama/pull/8662.diff", "patch_url": "https://github.com/ollama/ollama/pull/8662.patch", "merged_at": null }
This is just a minor change, I added DeepSeek R1 to the model library table. Only changed `README.md`
{ "login": "teymuur", "id": 64795612, "node_id": "MDQ6VXNlcjY0Nzk1NjEy", "avatar_url": "https://avatars.githubusercontent.com/u/64795612?v=4", "gravatar_id": "", "url": "https://api.github.com/users/teymuur", "html_url": "https://github.com/teymuur", "followers_url": "https://api.github.com/users/teymuu...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8662/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8662/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5631
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5631/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5631/comments
https://api.github.com/repos/ollama/ollama/issues/5631/events
https://github.com/ollama/ollama/pull/5631
2,403,403,309
PR_kwDOJ0Z1Ps51G6eZ
5,631
Refactor linux packaging
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
3
2024-07-11T14:56:45
2024-08-17T17:16:53
2024-08-17T17:16:45
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5631", "html_url": "https://github.com/ollama/ollama/pull/5631", "diff_url": "https://github.com/ollama/ollama/pull/5631.diff", "patch_url": "https://github.com/ollama/ollama/pull/5631.patch", "merged_at": null }
This adjusts linux to follow a similar model to windows with a discrete archive (zip/tgz) to cary the primary executable, and dependent libraries. Runners are still carried as payloads inside the main binary. As Darwin has no significant dependent libraries, it still functions as a discrete stand-alone executable ca...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5631/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5631/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1112
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1112/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1112/comments
https://api.github.com/repos/ollama/ollama/issues/1112/events
https://github.com/ollama/ollama/issues/1112
1,991,164,963
I_kwDOJ0Z1Ps52rsQj
1,112
Support `ollama create` with PyTorch
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2023-11-13T17:53:57
2024-05-06T23:26:01
2024-05-06T23:26:01
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Currently, create a model via a `Modelfile` supports importing GGUF format model binaries. Ollama should also support importing PyTorch models directly via `ollama create` Related: * https://github.com/jmorganca/ollama/issues/1037 * https://github.com/jmorganca/ollama/issues/1097
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1112/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1112/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/112
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/112/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/112/comments
https://api.github.com/repos/ollama/ollama/issues/112/events
https://github.com/ollama/ollama/pull/112
1,811,022,006
PR_kwDOJ0Z1Ps5V18xY
112
resolve modelfile before passing to server
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
0
2023-07-19T02:34:32
2023-07-19T02:36:27
2023-07-19T02:36:25
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/112", "html_url": "https://github.com/ollama/ollama/pull/112", "diff_url": "https://github.com/ollama/ollama/pull/112.diff", "patch_url": "https://github.com/ollama/ollama/pull/112.patch", "merged_at": "2023-07-19T02:36:25" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/112/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/112/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4095
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4095/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4095/comments
https://api.github.com/repos/ollama/ollama/issues/4095/events
https://github.com/ollama/ollama/issues/4095
2,274,734,176
I_kwDOJ0Z1Ps6HlbBg
4,095
Is there a problem with the document?
{ "login": "ggjk616", "id": 168710680, "node_id": "U_kgDOCg5SGA", "avatar_url": "https://avatars.githubusercontent.com/u/168710680?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ggjk616", "html_url": "https://github.com/ggjk616", "followers_url": "https://api.github.com/users/ggjk616/foll...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
0
2024-05-02T06:37:54
2024-05-02T10:16:12
2024-05-02T10:16:12
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Can you help me,In the documentation, I noticed the following statement: “You can set OLLAMA_LLM_LIBRARY to any of the available LLM libraries to bypass autodetection, so for example, if you have a CUDA card, but want to force the CPU LLM library with AVX2 vector support, use: OLLAMA_LLM_LIBRAR...
{ "login": "ggjk616", "id": 168710680, "node_id": "U_kgDOCg5SGA", "avatar_url": "https://avatars.githubusercontent.com/u/168710680?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ggjk616", "html_url": "https://github.com/ggjk616", "followers_url": "https://api.github.com/users/ggjk616/foll...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4095/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4095/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4155
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4155/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4155/comments
https://api.github.com/repos/ollama/ollama/issues/4155/events
https://github.com/ollama/ollama/issues/4155
2,279,169,147
I_kwDOJ0Z1Ps6H2Vx7
4,155
Add option in the install scripts to auto set OLLAMA_HOST environment variable
{ "login": "centopw", "id": 30675552, "node_id": "MDQ6VXNlcjMwNjc1NTUy", "avatar_url": "https://avatars.githubusercontent.com/u/30675552?v=4", "gravatar_id": "", "url": "https://api.github.com/users/centopw", "html_url": "https://github.com/centopw", "followers_url": "https://api.github.com/users/centop...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2024-05-04T19:37:42
2024-05-09T21:16:30
2024-05-09T21:16:30
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
In the installer scripts, add a option that ask if user want to allow other machine on the same network to connect. Base on this docs: [faq](https://github.com/ollama/ollama/blob/main/docs/faq.md#setting-environment-variables-on-mac) I can create a PR if needed
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4155/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4155/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3030
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3030/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3030/comments
https://api.github.com/repos/ollama/ollama/issues/3030/events
https://github.com/ollama/ollama/pull/3030
2,177,439,108
PR_kwDOJ0Z1Ps5pJ9Nn
3,030
Update llama.cpp submodule to `77d1ac7`
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[]
closed
false
null
[]
null
0
2024-03-09T23:10:52
2024-03-09T23:55:35
2024-03-09T23:55:34
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3030", "html_url": "https://github.com/ollama/ollama/pull/3030", "diff_url": "https://github.com/ollama/ollama/pull/3030.diff", "patch_url": "https://github.com/ollama/ollama/pull/3030.patch", "merged_at": "2024-03-09T23:55:34" }
Note we use `-DLLAMA_METAL_EMBED_LIBRARY=on` on arm64 darwin to embed ggml-metal.metal. This change also required us to prepend ggml-common.h to the top of ggml-metal.metal to avoid a runtime lookup error.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3030/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3030/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1157
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1157/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1157/comments
https://api.github.com/repos/ollama/ollama/issues/1157/events
https://github.com/ollama/ollama/issues/1157
1,997,826,755
I_kwDOJ0Z1Ps53FGrD
1,157
[Linux] - Instructions for exposing Ollama doesn't work
{ "login": "SoloBSD", "id": 17459633, "node_id": "MDQ6VXNlcjE3NDU5NjMz", "avatar_url": "https://avatars.githubusercontent.com/u/17459633?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SoloBSD", "html_url": "https://github.com/SoloBSD", "followers_url": "https://api.github.com/users/SoloBS...
[]
closed
false
null
[]
null
8
2023-11-16T21:33:43
2023-11-17T01:09:35
2023-11-17T00:55:21
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Instructions for Linux on how to expose ollama doesn't work. https://github.com/jmorganca/ollama/blob/main/docs/faq.md#how-can-i-expose-ollama-on-my-network For some reason when Ollama gets installed on Linux it creates: /etc/systemd/system/ollama.service So it seems it never processes /etc/systemd/system/o...
{ "login": "SoloBSD", "id": 17459633, "node_id": "MDQ6VXNlcjE3NDU5NjMz", "avatar_url": "https://avatars.githubusercontent.com/u/17459633?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SoloBSD", "html_url": "https://github.com/SoloBSD", "followers_url": "https://api.github.com/users/SoloBS...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1157/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1157/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2466
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2466/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2466/comments
https://api.github.com/repos/ollama/ollama/issues/2466/events
https://github.com/ollama/ollama/pull/2466
2,130,496,639
PR_kwDOJ0Z1Ps5mp7JN
2,466
Added NextJS web interface for Ollama models to readme.md
{ "login": "jakobhoeg", "id": 114422072, "node_id": "U_kgDOBtHxOA", "avatar_url": "https://avatars.githubusercontent.com/u/114422072?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jakobhoeg", "html_url": "https://github.com/jakobhoeg", "followers_url": "https://api.github.com/users/jakobh...
[]
closed
false
null
[]
null
1
2024-02-12T16:26:42
2024-02-20T02:57:36
2024-02-20T02:57:36
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2466", "html_url": "https://github.com/ollama/ollama/pull/2466", "diff_url": "https://github.com/ollama/ollama/pull/2466.diff", "patch_url": "https://github.com/ollama/ollama/pull/2466.patch", "merged_at": "2024-02-20T02:57:36" }
Added [nextjs-ollama-llm-ui](https://github.com/jakobhoeg/nextjs-ollama-llm-ui) to the readme file.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2466/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2466/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2659
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2659/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2659/comments
https://api.github.com/repos/ollama/ollama/issues/2659/events
https://github.com/ollama/ollama/issues/2659
2,148,066,317
I_kwDOJ0Z1Ps6ACOQN
2,659
Add phixtral
{ "login": "vprelovac", "id": 4319401, "node_id": "MDQ6VXNlcjQzMTk0MDE=", "avatar_url": "https://avatars.githubusercontent.com/u/4319401?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vprelovac", "html_url": "https://github.com/vprelovac", "followers_url": "https://api.github.com/users/vp...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
0
2024-02-22T02:39:04
2024-03-12T02:02:47
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Currently the best 2B model https://huggingface.co/shadowml/phixtral-4x2_8odd
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2659/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2659/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2850
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2850/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2850/comments
https://api.github.com/repos/ollama/ollama/issues/2850/events
https://github.com/ollama/ollama/issues/2850
2,162,455,308
I_kwDOJ0Z1Ps6A5HMM
2,850
`ollama push` and `ollama pull` are slow or hang on windows
{ "login": "ewebgh33", "id": 123797054, "node_id": "U_kgDOB2D-Pg", "avatar_url": "https://avatars.githubusercontent.com/u/123797054?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ewebgh33", "html_url": "https://github.com/ewebgh33", "followers_url": "https://api.github.com/users/ewebgh33/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677370291, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCVsw...
closed
false
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[ { "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/...
null
10
2024-03-01T02:21:35
2024-08-06T18:10:32
2024-08-06T18:10:32
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Can't download ANY models. What is happening? Not my internet, speed test blasts. Your servers OK? Happening on Windows version buggy still? Using latest, 0.1.27 (Win11). As per docs, set Windows environment variable to: OLLAMA_MODELS = D:\AI\text\ollama-models I am familiar with environment variables and ...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2850/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2850/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5474
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5474/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5474/comments
https://api.github.com/repos/ollama/ollama/issues/5474/events
https://github.com/ollama/ollama/issues/5474
2,389,647,879
I_kwDOJ0Z1Ps6ObyIH
5,474
InternLM2.5 - hallucinations - lot of repetitions etc
{ "login": "Qualzz", "id": 35169816, "node_id": "MDQ6VXNlcjM1MTY5ODE2", "avatar_url": "https://avatars.githubusercontent.com/u/35169816?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Qualzz", "html_url": "https://github.com/Qualzz", "followers_url": "https://api.github.com/users/Qualzz/fo...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[ { "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.g...
null
5
2024-07-03T23:18:06
2024-10-04T17:07:56
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Seems like something is wrong with InternLM2.5, I can't get any meaningful out of it. (tried with 32k context) ### OS Linux ### GPU Nvidia ### CPU AMD ### Ollama version v0.1.48
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5474/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 2 }
https://api.github.com/repos/ollama/ollama/issues/5474/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/1169
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1169/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1169/comments
https://api.github.com/repos/ollama/ollama/issues/1169/events
https://github.com/ollama/ollama/issues/1169
1,998,422,439
I_kwDOJ0Z1Ps53HYGn
1,169
Update the model name in the api doc
{ "login": "shenli", "id": 1192573, "node_id": "MDQ6VXNlcjExOTI1NzM=", "avatar_url": "https://avatars.githubusercontent.com/u/1192573?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shenli", "html_url": "https://github.com/shenli", "followers_url": "https://api.github.com/users/shenli/foll...
[]
closed
false
null
[]
null
1
2023-11-17T07:10:44
2023-11-17T12:18:09
2023-11-17T12:18:09
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi, I am new to Ollama. I followed the [Quickstart](https://github.com/jmorganca/ollama/tree/main#quickstart) to try Ollama with model Llama2. It is very easy to run and a very interesting project. When I explored further in the [API doc](https://github.com/jmorganca/ollama/blob/main/docs/api.md), I found that t...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1169/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1169/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2517
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2517/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2517/comments
https://api.github.com/repos/ollama/ollama/issues/2517/events
https://github.com/ollama/ollama/issues/2517
2,137,087,723
I_kwDOJ0Z1Ps5_YV7r
2,517
parser/parser.go:9:2: package log/slog is not in GOROOT (/usr/local/go120/src/log/slog)
{ "login": "yurivict", "id": 271906, "node_id": "MDQ6VXNlcjI3MTkwNg==", "avatar_url": "https://avatars.githubusercontent.com/u/271906?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yurivict", "html_url": "https://github.com/yurivict", "followers_url": "https://api.github.com/users/yurivic...
[]
closed
false
null
[]
null
1
2024-02-15T17:36:41
2024-02-15T19:51:08
2024-02-15T19:51:08
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Build fails: ``` ===> Building for ollama-0.1.25 (cd /usr/ports/misc/ollama/work/github.com/ollama/ollama@v0.1.25; for t in ./cmd; do out=$(/usr/bin/basename $(echo ${t} | /usr/bin/sed -Ee 's/^[^:]*:([^:]+).*$/\1/' -e 's/^\.$/ollama/')); pkg=$(echo ${t} | /usr/bin/sed -Ee 's/^([^:]*).*$/\1/' -e 's/^ollama$/./'...
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2517/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2517/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7287
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7287/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7287/comments
https://api.github.com/repos/ollama/ollama/issues/7287/events
https://github.com/ollama/ollama/issues/7287
2,601,723,323
I_kwDOJ0Z1Ps6bEyW7
7,287
Version v0.3.14 impacted CPU inference performance
{ "login": "closesim", "id": 9018799, "node_id": "MDQ6VXNlcjkwMTg3OTk=", "avatar_url": "https://avatars.githubusercontent.com/u/9018799?v=4", "gravatar_id": "", "url": "https://api.github.com/users/closesim", "html_url": "https://github.com/closesim", "followers_url": "https://api.github.com/users/close...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677677816, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgVG-A...
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
9
2024-10-21T08:10:06
2024-10-30T22:05:47
2024-10-30T22:05:47
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Hi, I just updated my docker container where I run my small models to the latest version, as I use to every 15 days or so. I'm using a Quad Core CPU (no GPU) and with this new version I noticed that LLama 3.1 8b performance was very slow. I Initially thought it was a hardware issue, like overhea...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7287/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7287/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1088
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1088/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1088/comments
https://api.github.com/repos/ollama/ollama/issues/1088/events
https://github.com/ollama/ollama/issues/1088
1,989,023,621
I_kwDOJ0Z1Ps52jheF
1,088
Problems installing the docker image.
{ "login": "pdavis68", "id": 2781885, "node_id": "MDQ6VXNlcjI3ODE4ODU=", "avatar_url": "https://avatars.githubusercontent.com/u/2781885?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdavis68", "html_url": "https://github.com/pdavis68", "followers_url": "https://api.github.com/users/pdavi...
[]
closed
false
null
[]
null
0
2023-11-11T16:17:04
2023-11-11T16:19:32
2023-11-11T16:19:32
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
null
{ "login": "pdavis68", "id": 2781885, "node_id": "MDQ6VXNlcjI3ODE4ODU=", "avatar_url": "https://avatars.githubusercontent.com/u/2781885?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdavis68", "html_url": "https://github.com/pdavis68", "followers_url": "https://api.github.com/users/pdavi...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1088/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1088/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3262
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3262/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3262/comments
https://api.github.com/repos/ollama/ollama/issues/3262/events
https://github.com/ollama/ollama/issues/3262
2,196,585,763
I_kwDOJ0Z1Ps6C7T0j
3,262
Ollama can support windows 7?
{ "login": "zhaosd", "id": 5444416, "node_id": "MDQ6VXNlcjU0NDQ0MTY=", "avatar_url": "https://avatars.githubusercontent.com/u/5444416?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zhaosd", "html_url": "https://github.com/zhaosd", "followers_url": "https://api.github.com/users/zhaosd/foll...
[]
closed
false
null
[]
null
6
2024-03-20T03:12:53
2024-11-18T23:47:12
2024-03-20T07:41:50
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
null
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3262/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3262/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7124
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7124/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7124/comments
https://api.github.com/repos/ollama/ollama/issues/7124/events
https://github.com/ollama/ollama/pull/7124
2,571,787,113
PR_kwDOJ0Z1Ps594dhh
7,124
llama: Decouple patching script from submodule
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
0
2024-10-08T00:29:54
2024-10-08T16:21:35
2024-10-08T15:54:00
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7124", "html_url": "https://github.com/ollama/ollama/pull/7124", "diff_url": "https://github.com/ollama/ollama/pull/7124.diff", "patch_url": "https://github.com/ollama/ollama/pull/7124.patch", "merged_at": null }
Replaced by #7139 on main
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7124/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7124/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/862
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/862/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/862/comments
https://api.github.com/repos/ollama/ollama/issues/862/events
https://github.com/ollama/ollama/pull/862
1,955,129,810
PR_kwDOJ0Z1Ps5dbU1e
862
fix/Predict: A prediction should use the options sent with the request
{ "login": "CyrilPeponnet", "id": 2277387, "node_id": "MDQ6VXNlcjIyNzczODc=", "avatar_url": "https://avatars.githubusercontent.com/u/2277387?v=4", "gravatar_id": "", "url": "https://api.github.com/users/CyrilPeponnet", "html_url": "https://github.com/CyrilPeponnet", "followers_url": "https://api.github....
[]
closed
false
null
[]
null
2
2023-10-20T23:26:14
2023-10-26T15:07:42
2023-10-26T15:07:42
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/862", "html_url": "https://github.com/ollama/ollama/pull/862", "diff_url": "https://github.com/ollama/ollama/pull/862.diff", "patch_url": "https://github.com/ollama/ollama/pull/862.patch", "merged_at": null }
Consecutive query to the same running model should use the client request parameters instead of the one set during the model loading.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/862/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/862/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6799
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6799/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6799/comments
https://api.github.com/repos/ollama/ollama/issues/6799/events
https://github.com/ollama/ollama/issues/6799
2,526,152,209
I_kwDOJ0Z1Ps6WkgYR
6,799
Is it possible to configure ollama deployed in docker?
{ "login": "wizounovziki", "id": 42036658, "node_id": "MDQ6VXNlcjQyMDM2NjU4", "avatar_url": "https://avatars.githubusercontent.com/u/42036658?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wizounovziki", "html_url": "https://github.com/wizounovziki", "followers_url": "https://api.github.c...
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" }, { "id": 6677677816, "node_id": "LA_kwDOJ0Z1Ps8AAAABjg...
closed
false
null
[]
null
1
2024-09-14T09:29:47
2024-09-25T21:23:15
2024-09-25T21:23:15
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I pull docker image from dockerhub and launched a few models and then found the num of user requests was limited. In the documentation it shows that this could be solved by set up OLLAMA_NUM_PARALLEL by systemctl commands. How can I do this since systemctl is not included in the docker container?
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6799/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6799/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1499
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1499/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1499/comments
https://api.github.com/repos/ollama/ollama/issues/1499/events
https://github.com/ollama/ollama/issues/1499
2,039,445,848
I_kwDOJ0Z1Ps55j3lY
1,499
Add mistral's new 7B-instruct-v0.2
{ "login": "tarek-ayed", "id": 45576986, "node_id": "MDQ6VXNlcjQ1NTc2OTg2", "avatar_url": "https://avatars.githubusercontent.com/u/45576986?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tarek-ayed", "html_url": "https://github.com/tarek-ayed", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
4
2023-12-13T10:51:12
2023-12-14T03:11:15
2023-12-13T16:17:43
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Along with many releases, Mistral vastly improved their existing 7B model with a version named `v0.2`. It has 32k context instead of 8k and better benchmark scores: https://x.com/dchaplot/status/1734198245067243629?s=20 More can be found here: https://docs.mistral.ai/platform/endpoints (see "Mistral Tiny") The w...
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/us...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1499/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1499/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5399
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5399/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5399/comments
https://api.github.com/repos/ollama/ollama/issues/5399/events
https://github.com/ollama/ollama/issues/5399
2,382,949,738
I_kwDOJ0Z1Ps6OCO1q
5,399
Please support models of rerank type
{ "login": "yushengliao", "id": 29765903, "node_id": "MDQ6VXNlcjI5NzY1OTAz", "avatar_url": "https://avatars.githubusercontent.com/u/29765903?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yushengliao", "html_url": "https://github.com/yushengliao", "followers_url": "https://api.github.com/...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
10
2024-07-01T06:23:13
2024-09-02T20:51:51
2024-09-02T20:51:50
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
There are so many users for the Ollama project, why hasn't it been so long to support Renanker Similar software such as Localai、xinreference already supports rerank https://localai.io/features/reranker/ https://inference.readthedocs.io/en/latest/models/builtin/rerank/index.html ![image](https://github.com/ollam...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5399/reactions", "total_count": 10, "+1": 9, "-1": 0, "laugh": 0, "hooray": 1, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5399/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6952
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6952/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6952/comments
https://api.github.com/repos/ollama/ollama/issues/6952/events
https://github.com/ollama/ollama/issues/6952
2,547,557,710
I_kwDOJ0Z1Ps6X2KVO
6,952
codegeex4-----Error: pull model manifest
{ "login": "zylGit-lte", "id": 181957291, "node_id": "U_kgDOCthyqw", "avatar_url": "https://avatars.githubusercontent.com/u/181957291?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zylGit-lte", "html_url": "https://github.com/zylGit-lte", "followers_url": "https://api.github.com/users/zyl...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-09-25T10:09:54
2024-09-25T11:22:21
2024-09-25T11:21:01
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? when i run command "ollama run codegeex4", then print out below log,how can i slove this problem? pulling manifest Error: pull model manifest: Get "https://registry.ollama.ai/v2/library/codegeex4/manifests/latest": dial tcp 172.67.182.229:443: i/o timeout ### OS Linux ### GPU Other ##...
{ "login": "zylGit-lte", "id": 181957291, "node_id": "U_kgDOCthyqw", "avatar_url": "https://avatars.githubusercontent.com/u/181957291?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zylGit-lte", "html_url": "https://github.com/zylGit-lte", "followers_url": "https://api.github.com/users/zyl...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6952/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6952/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1709
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1709/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1709/comments
https://api.github.com/repos/ollama/ollama/issues/1709/events
https://github.com/ollama/ollama/issues/1709
2,055,700,932
I_kwDOJ0Z1Ps56h4HE
1,709
Is there any plan to support FinGPT
{ "login": "waqasakram117", "id": 13805372, "node_id": "MDQ6VXNlcjEzODA1Mzcy", "avatar_url": "https://avatars.githubusercontent.com/u/13805372?v=4", "gravatar_id": "", "url": "https://api.github.com/users/waqasakram117", "html_url": "https://github.com/waqasakram117", "followers_url": "https://api.githu...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
5
2023-12-25T13:55:01
2024-07-03T21:18:02
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hey Team, Is there any plan to support [FinGPT](https://github.com/AI4Finance-Foundation/FinGPT) anytime soon?
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1709/reactions", "total_count": 22, "+1": 7, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 3, "eyes": 12 }
https://api.github.com/repos/ollama/ollama/issues/1709/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6481
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6481/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6481/comments
https://api.github.com/repos/ollama/ollama/issues/6481/events
https://github.com/ollama/ollama/issues/6481
2,483,787,305
I_kwDOJ0Z1Ps6UC5Yp
6,481
gork2
{ "login": "olumolu", "id": 162728301, "node_id": "U_kgDOCbMJbQ", "avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4", "gravatar_id": "", "url": "https://api.github.com/users/olumolu", "html_url": "https://github.com/olumolu", "followers_url": "https://api.github.com/users/olumolu/foll...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
2
2024-08-23T19:46:43
2024-08-24T04:31:25
2024-08-23T20:35:17
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
New gork2 has already published can we have support for that thanks.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6481/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6481/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3844
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3844/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3844/comments
https://api.github.com/repos/ollama/ollama/issues/3844/events
https://github.com/ollama/ollama/issues/3844
2,258,843,905
I_kwDOJ0Z1Ps6GozkB
3,844
api error occurred after some times request
{ "login": "Shiyaoa", "id": 48488459, "node_id": "MDQ6VXNlcjQ4NDg4NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/48488459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Shiyaoa", "html_url": "https://github.com/Shiyaoa", "followers_url": "https://api.github.com/users/Shiyao...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
7
2024-04-23T13:09:20
2025-01-06T03:55:27
2024-04-25T11:50:41
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? i try to post request using the url http://localhost:11434/v1 and model "llama3:8b-instruct-q8_0", it works successfully at the initially first time, but then failed with these information: Error occurred: Error code: 400 - {'error': {'message': 'unexpected server status: 1', 'type': 'api_erro...
{ "login": "Shiyaoa", "id": 48488459, "node_id": "MDQ6VXNlcjQ4NDg4NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/48488459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Shiyaoa", "html_url": "https://github.com/Shiyaoa", "followers_url": "https://api.github.com/users/Shiyao...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3844/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3844/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4275
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4275/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4275/comments
https://api.github.com/repos/ollama/ollama/issues/4275/events
https://github.com/ollama/ollama/issues/4275
2,286,923,121
I_kwDOJ0Z1Ps6IT61x
4,275
Degraded accuracy when using the nomic-embed-text (v1.5) model with Ollama versions 0.1.32 and 0.1.33
{ "login": "Ganesh1030", "id": 48667223, "node_id": "MDQ6VXNlcjQ4NjY3MjIz", "avatar_url": "https://avatars.githubusercontent.com/u/48667223?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Ganesh1030", "html_url": "https://github.com/Ganesh1030", "followers_url": "https://api.github.com/use...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-05-09T05:29:55
2024-06-26T05:47:57
2024-06-25T16:46:35
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? We have an application where we are training the CSV file and using the following things for it:. - 'nomic-embed-text(v1.5) model -  chromadb -  ollama(0.1.31)   At runtime, we are using'similarity_search' and getting good accuracy with ollama version 0.1.31, but when we upgrade ollama vers...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4275/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4275/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/633
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/633/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/633/comments
https://api.github.com/repos/ollama/ollama/issues/633/events
https://github.com/ollama/ollama/pull/633
1,917,657,126
PR_kwDOJ0Z1Ps5bc6m4
633
do not download updates multiple times
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
[]
closed
false
null
[]
null
0
2023-09-28T14:20:30
2023-09-28T19:29:18
2023-09-28T19:29:18
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/633", "html_url": "https://github.com/ollama/ollama/pull/633", "diff_url": "https://github.com/ollama/ollama/pull/633.diff", "patch_url": "https://github.com/ollama/ollama/pull/633.patch", "merged_at": "2023-09-28T19:29:18" }
We've hit a bug in the Electron auto-updater that prevents the toolbar app from restarting after update when `autoUpdater.checkForUpdates()` is called more than once. The root cause of this is not clear, it may be related to [this Electron issue](https://github.com/electron-userland/electron-builder/issues/7800). In an...
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/Br...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/633/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/633/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4395
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4395/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4395/comments
https://api.github.com/repos/ollama/ollama/issues/4395/events
https://github.com/ollama/ollama/issues/4395
2,292,283,708
I_kwDOJ0Z1Ps6IoXk8
4,395
Cannot Use GPU properly
{ "login": "applepieiris", "id": 36785462, "node_id": "MDQ6VXNlcjM2Nzg1NDYy", "avatar_url": "https://avatars.githubusercontent.com/u/36785462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/applepieiris", "html_url": "https://github.com/applepieiris", "followers_url": "https://api.github.c...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg...
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
7
2024-05-13T09:25:56
2024-06-02T00:29:42
2024-06-02T00:29:42
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I installed the Ollama in my linux server according to the official documents: `curl -fsSL https://ollama.com/install.sh | sh` Installation is ok and it returns: ` >>> Downloading ollama... ######################################################################## 100.0%-#O#- # # >>> Ins...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4395/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4395/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8654
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8654/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8654/comments
https://api.github.com/repos/ollama/ollama/issues/8654/events
https://github.com/ollama/ollama/issues/8654
2,817,986,286
I_kwDOJ0Z1Ps6n9w7u
8,654
Available memory check should be disabled when mmap is in use
{ "login": "outis151", "id": 11805613, "node_id": "MDQ6VXNlcjExODA1NjEz", "avatar_url": "https://avatars.githubusercontent.com/u/11805613?v=4", "gravatar_id": "", "url": "https://api.github.com/users/outis151", "html_url": "https://github.com/outis151", "followers_url": "https://api.github.com/users/out...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
1
2025-01-29T11:48:38
2025-01-29T13:07:03
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? With mmap enabled, a model does not need to fit in the system RAM. Therefore the associated check should be disabled in this case. ### OS Linux ### GPU _No response_ ### CPU Intel ### Ollama version 0.5.7
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8654/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8654/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/3396
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3396/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3396/comments
https://api.github.com/repos/ollama/ollama/issues/3396/events
https://github.com/ollama/ollama/issues/3396
2,214,189,195
I_kwDOJ0Z1Ps6D-diL
3,396
exec format error when Running Ollama Container on AMD64 Architecture
{ "login": "joshyorko", "id": 54248591, "node_id": "MDQ6VXNlcjU0MjQ4NTkx", "avatar_url": "https://avatars.githubusercontent.com/u/54248591?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joshyorko", "html_url": "https://github.com/joshyorko", "followers_url": "https://api.github.com/users/...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-03-28T21:26:11
2024-03-29T00:11:02
2024-03-29T00:11:01
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? When attempting to run the Ollama container, I continuously encounter an exec format error. The container fails to start properly and keeps trying to restart, logging the same error multiple times. What did you expect to happen? I expected the Ollama container to start successfully without any ...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3396/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3396/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7553
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7553/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7553/comments
https://api.github.com/repos/ollama/ollama/issues/7553/events
https://github.com/ollama/ollama/issues/7553
2,640,556,716
I_kwDOJ0Z1Ps6dY7Ks
7,553
Unable to load images from network fileshares on Windows
{ "login": "Antsiscool", "id": 4112838, "node_id": "MDQ6VXNlcjQxMTI4Mzg=", "avatar_url": "https://avatars.githubusercontent.com/u/4112838?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Antsiscool", "html_url": "https://github.com/Antsiscool", "followers_url": "https://api.github.com/users...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg...
open
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
4
2024-11-07T10:24:54
2024-11-17T19:50:17
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Using Ollama on Windows via the terminal, if you ask a question and reference an image on a network fileshare, it will give a response about it not been able to see the photo. If you copy the image locally and then reference the local image, it has no problem with analysing the image. Paths s...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7553/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7553/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/8002
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8002/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8002/comments
https://api.github.com/repos/ollama/ollama/issues/8002/events
https://github.com/ollama/ollama/pull/8002
2,725,706,888
PR_kwDOJ0Z1Ps6Edzde
8,002
llama: preserve field order in user-defined JSON schemas
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers"...
[]
closed
false
null
[]
null
3
2024-12-09T01:14:51
2024-12-11T22:07:32
2024-12-11T22:07:30
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8002", "html_url": "https://github.com/ollama/ollama/pull/8002", "diff_url": "https://github.com/ollama/ollama/pull/8002.diff", "patch_url": "https://github.com/ollama/ollama/pull/8002.patch", "merged_at": "2024-12-11T22:07:30" }
llama: preserve field order in user-defined JSON schemas Previously we decoded and re-encoded JSON schemas during validation, which served no purpose since json.RawMessage already validates JSON syntax. Worse, the re-encoding lost field ordering from the original schema, which affects inference quality during ste...
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers"...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8002/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8002/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5985
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5985/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5985/comments
https://api.github.com/repos/ollama/ollama/issues/5985/events
https://github.com/ollama/ollama/pull/5985
2,432,324,831
PR_kwDOJ0Z1Ps52leCz
5,985
Use llama3.1 in tools example
{ "login": "rgbkrk", "id": 836375, "node_id": "MDQ6VXNlcjgzNjM3NQ==", "avatar_url": "https://avatars.githubusercontent.com/u/836375?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rgbkrk", "html_url": "https://github.com/rgbkrk", "followers_url": "https://api.github.com/users/rgbkrk/follow...
[]
closed
false
null
[]
null
0
2024-07-26T14:07:11
2024-08-08T01:31:50
2024-08-07T21:20:51
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5985", "html_url": "https://github.com/ollama/ollama/pull/5985", "diff_url": "https://github.com/ollama/ollama/pull/5985.diff", "patch_url": "https://github.com/ollama/ollama/pull/5985.patch", "merged_at": "2024-08-07T21:20:50" }
Running this example with `mistral` produces the error "mistral does not support tools". What wasn't obvious to me until I made this PR was that my copy of mistral needed upgrading for tools (`ollama pull mistral`). Making the example be `llama3.1` will lead to more success for other long time ollama users.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5985/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5985/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6319
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6319/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6319/comments
https://api.github.com/repos/ollama/ollama/issues/6319/events
https://github.com/ollama/ollama/issues/6319
2,460,663,129
I_kwDOJ0Z1Ps6Sqr1Z
6,319
Models RuGPT3, RuBERT
{ "login": "DewiarQR", "id": 64423698, "node_id": "MDQ6VXNlcjY0NDIzNjk4", "avatar_url": "https://avatars.githubusercontent.com/u/64423698?v=4", "gravatar_id": "", "url": "https://api.github.com/users/DewiarQR", "html_url": "https://github.com/DewiarQR", "followers_url": "https://api.github.com/users/Dew...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
2
2024-08-12T10:51:26
2024-12-21T10:38:07
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
All models currently have pretty poor Russian language support. Is it possible to add RuGPT3, RuBERT models?
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6319/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6319/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/766
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/766/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/766/comments
https://api.github.com/repos/ollama/ollama/issues/766/events
https://github.com/ollama/ollama/issues/766
1,939,890,658
I_kwDOJ0Z1Ps5zoGHi
766
Release mac and linux binaries alongside the desktop packages
{ "login": "Clivern", "id": 1634427, "node_id": "MDQ6VXNlcjE2MzQ0Mjc=", "avatar_url": "https://avatars.githubusercontent.com/u/1634427?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Clivern", "html_url": "https://github.com/Clivern", "followers_url": "https://api.github.com/users/Clivern/...
[]
closed
false
null
[]
null
5
2023-10-12T12:30:39
2023-11-11T22:56:39
2023-10-12T16:07:48
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Maybe ollama is intended to be a desktop app but I believe a lot are using it as an API service. Honestly i couldn't get it to work as desktop app on Intel Mac but works as API service. Assuming the 500% spike in cpu usage is expected with each prompt and model pulling on a decent mac. I guess i need to give it a t...
{ "login": "Clivern", "id": 1634427, "node_id": "MDQ6VXNlcjE2MzQ0Mjc=", "avatar_url": "https://avatars.githubusercontent.com/u/1634427?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Clivern", "html_url": "https://github.com/Clivern", "followers_url": "https://api.github.com/users/Clivern/...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/766/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/766/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6293
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6293/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6293/comments
https://api.github.com/repos/ollama/ollama/issues/6293/events
https://github.com/ollama/ollama/issues/6293
2,458,713,929
I_kwDOJ0Z1Ps6SjP9J
6,293
"The model you are attempting to pull requires a newer version of Ollama" when Ollama is built from the latest source
{ "login": "sammcj", "id": 862951, "node_id": "MDQ6VXNlcjg2Mjk1MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sammcj", "html_url": "https://github.com/sammcj", "followers_url": "https://api.github.com/users/sammcj/follow...
[ { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" } ]
closed
false
null
[]
null
8
2024-08-09T21:59:06
2024-08-15T02:32:17
2024-08-09T22:31:01
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? When trying to pull models from the official Ollama registry - if you're building Ollama from source it now seems to fail with an error that your Ollama version is too old. ``` ollama pull llama3.1:8b-instruct-q8_0 pulling manifest Error: pull model manifest: 412: The model you are atte...
{ "login": "sammcj", "id": 862951, "node_id": "MDQ6VXNlcjg2Mjk1MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sammcj", "html_url": "https://github.com/sammcj", "followers_url": "https://api.github.com/users/sammcj/follow...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6293/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6293/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/269
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/269/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/269/comments
https://api.github.com/repos/ollama/ollama/issues/269/events
https://github.com/ollama/ollama/issues/269
1,835,339,821
I_kwDOJ0Z1Ps5tZRAt
269
Pressing enter during `ollama pull` causes newlines to be printed repeatedly
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2023-08-03T16:02:52
2023-12-24T21:39:30
2023-12-24T21:39:30
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
<img width="1531" alt="Screenshot 2023-08-03 at 11 59 21 AM" src="https://github.com/jmorganca/ollama/assets/251292/1e782cfa-75f2-4bc3-84da-567c685ef36c">
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/269/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/269/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1251
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1251/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1251/comments
https://api.github.com/repos/ollama/ollama/issues/1251/events
https://github.com/ollama/ollama/issues/1251
2,007,405,923
I_kwDOJ0Z1Ps53ppVj
1,251
How can I disable automatic model offloading from GPU memory
{ "login": "anan-dad", "id": 30836142, "node_id": "MDQ6VXNlcjMwODM2MTQy", "avatar_url": "https://avatars.githubusercontent.com/u/30836142?v=4", "gravatar_id": "", "url": "https://api.github.com/users/anan-dad", "html_url": "https://github.com/anan-dad", "followers_url": "https://api.github.com/users/ana...
[]
closed
false
null
[]
null
1
2023-11-23T02:58:48
2023-11-23T03:05:50
2023-11-23T03:05:18
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
First of all, thank you for your great work with ollama! I found that ollama will automatically offload models from GPU memory (very frequently, even after 2-minute inactive use). But the loading process takes too much time, how can I forge ollama keep the model loading in GPU memory? Thanks
{ "login": "anan-dad", "id": 30836142, "node_id": "MDQ6VXNlcjMwODM2MTQy", "avatar_url": "https://avatars.githubusercontent.com/u/30836142?v=4", "gravatar_id": "", "url": "https://api.github.com/users/anan-dad", "html_url": "https://github.com/anan-dad", "followers_url": "https://api.github.com/users/ana...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1251/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1251/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2175
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2175/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2175/comments
https://api.github.com/repos/ollama/ollama/issues/2175/events
https://github.com/ollama/ollama/pull/2175
2,098,909,748
PR_kwDOJ0Z1Ps5k_KNU
2,175
refactor tensor read
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
0
2024-01-24T19:10:03
2024-01-25T17:22:43
2024-01-25T17:22:42
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2175", "html_url": "https://github.com/ollama/ollama/pull/2175", "diff_url": "https://github.com/ollama/ollama/pull/2175.diff", "patch_url": "https://github.com/ollama/ollama/pull/2175.patch", "merged_at": "2024-01-25T17:22:42" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2175/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2175/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8486
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8486/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8486/comments
https://api.github.com/repos/ollama/ollama/issues/8486/events
https://github.com/ollama/ollama/issues/8486
2,797,681,012
I_kwDOJ0Z1Ps6mwTl0
8,486
Add Tool Calling to the Generate Function
{ "login": "twright-0x1", "id": 13889385, "node_id": "MDQ6VXNlcjEzODg5Mzg1", "avatar_url": "https://avatars.githubusercontent.com/u/13889385?v=4", "gravatar_id": "", "url": "https://api.github.com/users/twright-0x1", "html_url": "https://github.com/twright-0x1", "followers_url": "https://api.github.com/...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
0
2025-01-19T15:21:05
2025-01-19T15:21:05
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
It appears from the API documentation and code examples available that tool calling is only possible with chat(). If this capability is feasible to add to generate() it would be much appreciated!
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8486/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8486/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2908
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2908/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2908/comments
https://api.github.com/repos/ollama/ollama/issues/2908/events
https://github.com/ollama/ollama/issues/2908
2,166,302,209
I_kwDOJ0Z1Ps6BHyYB
2,908
How to specify the installation directory
{ "login": "yuanjie-ai", "id": 20265321, "node_id": "MDQ6VXNlcjIwMjY1MzIx", "avatar_url": "https://avatars.githubusercontent.com/u/20265321?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yuanjie-ai", "html_url": "https://github.com/yuanjie-ai", "followers_url": "https://api.github.com/use...
[]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
3
2024-03-04T08:56:56
2024-05-26T09:21:51
2024-03-21T11:36:12
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
How to specify the installation directory
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2908/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2908/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4135
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4135/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4135/comments
https://api.github.com/repos/ollama/ollama/issues/4135/events
https://github.com/ollama/ollama/pull/4135
2,278,248,414
PR_kwDOJ0Z1Ps5ugQRD
4,135
Skip PhysX cudart library
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[]
closed
false
null
[]
null
1
2024-05-03T18:56:55
2024-05-06T20:34:03
2024-05-06T20:34:00
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4135", "html_url": "https://github.com/ollama/ollama/pull/4135", "diff_url": "https://github.com/ollama/ollama/pull/4135.diff", "patch_url": "https://github.com/ollama/ollama/pull/4135.patch", "merged_at": "2024-05-06T20:34:00" }
For some reason this library gives incorrect GPU information, so skip it I'm not convinced yet this is the optimal fix, but queuing this up in case we get ready to cut a new release and haven't found a better solution yet. Fixes #4008
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4135/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4135/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3089
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3089/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3089/comments
https://api.github.com/repos/ollama/ollama/issues/3089/events
https://github.com/ollama/ollama/issues/3089
2,182,976,031
I_kwDOJ0Z1Ps6CHZIf
3,089
Error when requesting ollama api from another pc (windows)
{ "login": "insooneelife", "id": 8437769, "node_id": "MDQ6VXNlcjg0Mzc3Njk=", "avatar_url": "https://avatars.githubusercontent.com/u/8437769?v=4", "gravatar_id": "", "url": "https://api.github.com/users/insooneelife", "html_url": "https://github.com/insooneelife", "followers_url": "https://api.github.com...
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
null
[]
null
16
2024-03-13T02:06:05
2024-05-13T21:15:11
2024-03-15T13:36:22
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I plan to set up ollama on another PC and proceed with the work on the current PC. However, when sending a request to ollama from a PC, I entered the IP address of the PC and sent it, but there is no reply. Can you tell me what the problem is? request url http://localhost:11434/api/chat -> http://172.168.10.1...
{ "login": "insooneelife", "id": 8437769, "node_id": "MDQ6VXNlcjg0Mzc3Njk=", "avatar_url": "https://avatars.githubusercontent.com/u/8437769?v=4", "gravatar_id": "", "url": "https://api.github.com/users/insooneelife", "html_url": "https://github.com/insooneelife", "followers_url": "https://api.github.com...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3089/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3089/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5452
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5452/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5452/comments
https://api.github.com/repos/ollama/ollama/issues/5452/events
https://github.com/ollama/ollama/issues/5452
2,387,458,446
I_kwDOJ0Z1Ps6OTbmO
5,452
MARKDOWN!!
{ "login": "ashercn97", "id": 131724380, "node_id": "U_kgDOB9n0XA", "avatar_url": "https://avatars.githubusercontent.com/u/131724380?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ashercn97", "html_url": "https://github.com/ashercn97", "followers_url": "https://api.github.com/users/asherc...
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
5
2024-07-03T01:42:42
2024-10-17T17:32:45
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I think it would be so cool if this could render markdown in the terminal. It is kind of hard to read some of the stuff, and I would love if it could use something like glow or mdcat. Thanks!
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5452/reactions", "total_count": 10, "+1": 10, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5452/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2168
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2168/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2168/comments
https://api.github.com/repos/ollama/ollama/issues/2168/events
https://github.com/ollama/ollama/issues/2168
2,097,815,632
I_kwDOJ0Z1Ps59CiBQ
2,168
Issues Running Ollama Container Behind Proxy - No Error Logs Found
{ "login": "OM-EL", "id": 36996895, "node_id": "MDQ6VXNlcjM2OTk2ODk1", "avatar_url": "https://avatars.githubusercontent.com/u/36996895?v=4", "gravatar_id": "", "url": "https://api.github.com/users/OM-EL", "html_url": "https://github.com/OM-EL", "followers_url": "https://api.github.com/users/OM-EL/follow...
[]
closed
false
null
[]
null
10
2024-01-24T09:26:08
2024-10-17T07:06:36
2024-03-11T19:02:35
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I'm encountering issues while trying to run an Ollama container behind a proxy. Here are the steps I've taken and the issues I've faced: 1. **Creating an Image with Certificate**: ``` cat Dockerfile FROM ollama/ollama COPY my-ca.pem /usr/local/share/ca-certificates/my-ca.crt RUN update-ca-c...
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2168/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2168/timeline
null
not_planned
false
https://api.github.com/repos/ollama/ollama/issues/5988
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5988/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5988/comments
https://api.github.com/repos/ollama/ollama/issues/5988/events
https://github.com/ollama/ollama/issues/5988
2,432,490,642
I_kwDOJ0Z1Ps6Q_NyS
5,988
GPU with 12GB VRAM couldn't load 8B model under WSL2
{ "login": "hoangminh1109", "id": 20716428, "node_id": "MDQ6VXNlcjIwNzE2NDI4", "avatar_url": "https://avatars.githubusercontent.com/u/20716428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoangminh1109", "html_url": "https://github.com/hoangminh1109", "followers_url": "https://api.githu...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg...
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
6
2024-07-26T15:37:39
2024-08-03T10:55:00
2024-08-03T10:55:00
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I'm unable to run any of the small model (8B model) on my RTX 3060 12GB. Ollama is installed in WSL2 under Win10. ![image](https://github.com/user-attachments/assets/56675022-afb1-4361-b7e8-30add303f8c1) Server log uploaded [ollama_log_error.txt](https://github.com/user-attachments/files/1639...
{ "login": "hoangminh1109", "id": 20716428, "node_id": "MDQ6VXNlcjIwNzE2NDI4", "avatar_url": "https://avatars.githubusercontent.com/u/20716428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoangminh1109", "html_url": "https://github.com/hoangminh1109", "followers_url": "https://api.githu...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5988/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5988/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3742
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3742/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3742/comments
https://api.github.com/repos/ollama/ollama/issues/3742/events
https://github.com/ollama/ollama/issues/3742
2,251,904,090
I_kwDOJ0Z1Ps6GOVRa
3,742
Slow Performance with Llama2 on a Dual-GPU System - Seeking Advice
{ "login": "AkiMatsushita", "id": 5045321, "node_id": "MDQ6VXNlcjUwNDUzMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/5045321?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AkiMatsushita", "html_url": "https://github.com/AkiMatsushita", "followers_url": "https://api.github....
[ { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhilt...
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.gi...
null
5
2024-04-19T01:41:32
2024-04-22T22:41:38
2024-04-22T22:39:29
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hello ollama Community, I'm encountering extremely slow performance while running ollama on my PC, specifically with models like Llama2 13B. The issue isn't just the slow output speed (around 1 token/min), but I'm also concerned that my GPUs might not be utilized properly. Below are my PC specs: - CPU: Intel Core...
{ "login": "AkiMatsushita", "id": 5045321, "node_id": "MDQ6VXNlcjUwNDUzMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/5045321?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AkiMatsushita", "html_url": "https://github.com/AkiMatsushita", "followers_url": "https://api.github....
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3742/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3742/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5912
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5912/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5912/comments
https://api.github.com/repos/ollama/ollama/issues/5912/events
https://github.com/ollama/ollama/pull/5912
2,427,590,112
PR_kwDOJ0Z1Ps52V55j
5,912
Server tls 3203
{ "login": "gabe-l-hart", "id": 1254484, "node_id": "MDQ6VXNlcjEyNTQ0ODQ=", "avatar_url": "https://avatars.githubusercontent.com/u/1254484?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gabe-l-hart", "html_url": "https://github.com/gabe-l-hart", "followers_url": "https://api.github.com/us...
[]
closed
false
null
[]
null
1
2024-07-24T13:25:29
2024-10-03T16:04:14
2024-10-03T16:04:14
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5912", "html_url": "https://github.com/ollama/ollama/pull/5912", "diff_url": "https://github.com/ollama/ollama/pull/5912.diff", "patch_url": "https://github.com/ollama/ollama/pull/5912.patch", "merged_at": null }
**Disclaimer!** This PR started as a small feature addition and resulted in some significant scope creep when I added the unit tests. I'm certainly open to trying to remove some of that refactoring for `ServerNonBlocking` if that's preferred, but figured it was worth presenting as-is to start the discussion. ## I...
{ "login": "gabe-l-hart", "id": 1254484, "node_id": "MDQ6VXNlcjEyNTQ0ODQ=", "avatar_url": "https://avatars.githubusercontent.com/u/1254484?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gabe-l-hart", "html_url": "https://github.com/gabe-l-hart", "followers_url": "https://api.github.com/us...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5912/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5912/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/236
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/236/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/236/comments
https://api.github.com/repos/ollama/ollama/issues/236/events
https://github.com/ollama/ollama/pull/236
1,826,971,486
PR_kwDOJ0Z1Ps5WryWE
236
check os.Walk err
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
0
2023-07-28T19:15:46
2023-07-28T21:14:22
2023-07-28T21:14:21
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/236", "html_url": "https://github.com/ollama/ollama/pull/236", "diff_url": "https://github.com/ollama/ollama/pull/236.diff", "patch_url": "https://github.com/ollama/ollama/pull/236.patch", "merged_at": "2023-07-28T21:14:21" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/236/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/236/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1541
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1541/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1541/comments
https://api.github.com/repos/ollama/ollama/issues/1541/events
https://github.com/ollama/ollama/pull/1541
2,042,976,921
PR_kwDOJ0Z1Ps5iEfYP
1,541
add API create/copy handlers
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
[]
closed
false
null
[]
null
0
2023-12-15T06:20:11
2023-12-15T19:59:19
2023-12-15T19:59:18
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1541", "html_url": "https://github.com/ollama/ollama/pull/1541", "diff_url": "https://github.com/ollama/ollama/pull/1541.diff", "patch_url": "https://github.com/ollama/ollama/pull/1541.patch", "merged_at": "2023-12-15T19:59:18" }
This change adds a test for calling `POST /api/create` which creates a new model.
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1541/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1541/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1889
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1889/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1889/comments
https://api.github.com/repos/ollama/ollama/issues/1889/events
https://github.com/ollama/ollama/issues/1889
2,074,013,731
I_kwDOJ0Z1Ps57nvAj
1,889
Phi2/dolphin-phi Disobedient on system prompt Biblical topics:
{ "login": "oliverbob", "id": 23272429, "node_id": "MDQ6VXNlcjIzMjcyNDI5", "avatar_url": "https://avatars.githubusercontent.com/u/23272429?v=4", "gravatar_id": "", "url": "https://api.github.com/users/oliverbob", "html_url": "https://github.com/oliverbob", "followers_url": "https://api.github.com/users/...
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
null
[]
null
4
2024-01-10T10:02:39
2024-05-10T00:16:11
2024-05-10T00:16:11
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Steps to reproduce: Download a new Bible Dataset from [KJV Markdown .md](https://github.com/arleym/kjv-markdown/tree/master ) ``` #!/bin/bash sudo rm joined.md # Prepend content to the joined.md file echo "FROM dolphin-phi" >> ./joined.md echo "# set the temperature to 1 [higher is more creative, lower is...
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1889/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1889/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/939
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/939/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/939/comments
https://api.github.com/repos/ollama/ollama/issues/939/events
https://github.com/ollama/ollama/issues/939
1,966,246,147
I_kwDOJ0Z1Ps51MokD
939
Low memory systems with a lot of VRAM hit a memory issue
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2023-10-27T22:06:04
2024-01-10T15:08:21
2024-01-10T15:08:21
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
When creating a small instance with <4GB of RAM, `ollama` hits an error when loading the memory into VRAM
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmor...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/939/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/939/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/372
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/372/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/372/comments
https://api.github.com/repos/ollama/ollama/issues/372/events
https://github.com/ollama/ollama/pull/372
1,855,522,492
PR_kwDOJ0Z1Ps5YL5qy
372
model and file type as strings
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
0
2023-08-17T18:41:58
2023-08-17T22:10:59
2023-08-17T22:10:59
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/372", "html_url": "https://github.com/ollama/ollama/pull/372", "diff_url": "https://github.com/ollama/ollama/pull/372.diff", "patch_url": "https://github.com/ollama/ollama/pull/372.patch", "merged_at": "2023-08-17T22:10:59" }
instead of representing model and file type as their native int values in manifest config, represent them as user-friendly strings
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/372/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/372/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/270
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/270/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/270/comments
https://api.github.com/repos/ollama/ollama/issues/270/events
https://github.com/ollama/ollama/pull/270
1,835,562,166
PR_kwDOJ0Z1Ps5XIuTO
270
update llama.cpp
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
[]
closed
false
null
[]
null
0
2023-08-03T18:50:35
2023-08-03T19:09:02
2023-08-03T19:09:01
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/270", "html_url": "https://github.com/ollama/ollama/pull/270", "diff_url": "https://github.com/ollama/ollama/pull/270.diff", "patch_url": "https://github.com/ollama/ollama/pull/270.patch", "merged_at": "2023-08-03T19:09:01" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/follower...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/270/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/ollama/ollama/issues/270/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1759
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1759/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1759/comments
https://api.github.com/repos/ollama/ollama/issues/1759/events
https://github.com/ollama/ollama/issues/1759
2,062,122,844
I_kwDOJ0Z1Ps566X9c
1,759
Please add TinyGPT-V model support
{ "login": "yangyang0507", "id": 5666807, "node_id": "MDQ6VXNlcjU2NjY4MDc=", "avatar_url": "https://avatars.githubusercontent.com/u/5666807?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yangyang0507", "html_url": "https://github.com/yangyang0507", "followers_url": "https://api.github.com...
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
0
2024-01-02T09:03:45
2024-01-02T11:34:34
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
TinyGPT-V: Efficient Multimodal Large Language Model via Small Backbones Github: https://github.com/DLYuanGod/TinyGPT-V HuggingFace: https://huggingface.co/Tyrannosaurus/TinyGPT-V It stands out because it only requires a 24G GPU for training, and just an 8G GPU or CPU for inference. TinyGPT-V is based on Phi-2, ...
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1759/reactions", "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1759/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/483
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/483/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/483/comments
https://api.github.com/repos/ollama/ollama/issues/483/events
https://github.com/ollama/ollama/issues/483
1,885,300,022
I_kwDOJ0Z1Ps5wX2U2
483
No response from model with giant request
{ "login": "FairyTail2000", "id": 22645621, "node_id": "MDQ6VXNlcjIyNjQ1NjIx", "avatar_url": "https://avatars.githubusercontent.com/u/22645621?v=4", "gravatar_id": "", "url": "https://api.github.com/users/FairyTail2000", "html_url": "https://github.com/FairyTail2000", "followers_url": "https://api.githu...
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2023-09-07T07:41:24
2023-12-04T19:24:58
2023-12-04T19:24:57
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Using my own personal frontend with the model codellama:34b-code-q4_0 I send a giant block of code ~10kB. The model then runs for 5 - 6 minutes but only a single token comes out of the model. This is the http response: >{"model":"codellama:34b-code-q4_0","created_at":"2023-09-07T07:34:32.574995065Z","response":"\n"...
{ "login": "technovangelist", "id": 633681, "node_id": "MDQ6VXNlcjYzMzY4MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4", "gravatar_id": "", "url": "https://api.github.com/users/technovangelist", "html_url": "https://github.com/technovangelist", "followers_url": "https://api.git...
{ "url": "https://api.github.com/repos/ollama/ollama/issues/483/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/483/timeline
null
completed
false