url stringlengths 51 54 | repository_url stringclasses 1
value | labels_url stringlengths 65 68 | comments_url stringlengths 60 63 | events_url stringlengths 58 61 | html_url stringlengths 39 44 | id int64 1.78B 2.82B | node_id stringlengths 18 19 | number int64 1 8.69k | title stringlengths 1 382 | user dict | labels listlengths 0 5 | state stringclasses 2
values | locked bool 1
class | assignee dict | assignees listlengths 0 2 | milestone null | comments int64 0 323 | created_at timestamp[s] | updated_at timestamp[s] | closed_at timestamp[s] | author_association stringclasses 4
values | sub_issues_summary dict | active_lock_reason null | draft bool 2
classes | pull_request dict | body stringlengths 2 118k ⌀ | closed_by dict | reactions dict | timeline_url stringlengths 60 63 | performed_via_github_app null | state_reason stringclasses 4
values | is_pull_request bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/ollama/ollama/issues/7049 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7049/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7049/comments | https://api.github.com/repos/ollama/ollama/issues/7049/events | https://github.com/ollama/ollama/issues/7049 | 2,557,134,462 | I_kwDOJ0Z1Ps6YasZ- | 7,049 | ollama does not detect Quadro RTX 4000 - cuda driver library failed to get device context 801 | {
"login": "mfzhsn",
"id": 5251972,
"node_id": "MDQ6VXNlcjUyNTE5NzI=",
"avatar_url": "https://avatars.githubusercontent.com/u/5251972?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mfzhsn",
"html_url": "https://github.com/mfzhsn",
"followers_url": "https://api.github.com/users/mfzhsn/foll... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5755339642,
"node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg... | open | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 10 | 2024-09-30T16:24:28 | 2024-11-25T19:04:01 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Hi All,
I installed ollama both (on machine/docker) both with same behaviour of not detecting the GPU. Have LM Studio on the same machine which picks up GPU without any issues.
```
root@d50a3f8d8474:/# ollama run phi3.5:3.8b-mini-instruct-q2_K ""
root@d50a3f8d8474:/# ollama ps
NAME ... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7049/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7049/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/2914 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2914/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2914/comments | https://api.github.com/repos/ollama/ollama/issues/2914/events | https://github.com/ollama/ollama/issues/2914 | 2,167,049,192 | I_kwDOJ0Z1Ps6BKovo | 2,914 | ollama run starcoder2:15b | {
"login": "limaolin2017",
"id": 28923721,
"node_id": "MDQ6VXNlcjI4OTIzNzIx",
"avatar_url": "https://avatars.githubusercontent.com/u/28923721?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/limaolin2017",
"html_url": "https://github.com/limaolin2017",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | 2 | 2024-03-04T14:57:44 | 2024-03-04T15:18:58 | 2024-03-04T15:12:38 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I have encountered an error with Apple silicon M1 pro:
ollama run starcoder2:15b
Error: Post "http://127.0.0.1:11434/api/chat": EOF | {
"login": "limaolin2017",
"id": 28923721,
"node_id": "MDQ6VXNlcjI4OTIzNzIx",
"avatar_url": "https://avatars.githubusercontent.com/u/28923721?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/limaolin2017",
"html_url": "https://github.com/limaolin2017",
"followers_url": "https://api.github.c... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2914/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2914/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5251 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5251/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5251/comments | https://api.github.com/repos/ollama/ollama/issues/5251/events | https://github.com/ollama/ollama/issues/5251 | 2,369,624,310 | I_kwDOJ0Z1Ps6NPZj2 | 5,251 | how to install this in my steam deck? | {
"login": "olumolu",
"id": 162728301,
"node_id": "U_kgDOCbMJbQ",
"avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/olumolu",
"html_url": "https://github.com/olumolu",
"followers_url": "https://api.github.com/users/olumolu/foll... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 1 | 2024-06-24T08:46:38 | 2024-06-25T16:19:14 | 2024-06-25T16:18:58 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
cant install this in steam os 3
i think this is the issue for fedora silverblue like os opensuse aeon also
### OS
Linux
### GPU
AMD
### CPU
AMD
### Ollama version
_No response_ | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5251/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5251/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7377 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7377/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7377/comments | https://api.github.com/repos/ollama/ollama/issues/7377/events | https://github.com/ollama/ollama/pull/7377 | 2,616,184,675 | PR_kwDOJ0Z1Ps5__7QO | 7,377 | readme: add TextCraft to community integrations | {
"login": "suncloudsmoon",
"id": 34616349,
"node_id": "MDQ6VXNlcjM0NjE2MzQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/34616349?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/suncloudsmoon",
"html_url": "https://github.com/suncloudsmoon",
"followers_url": "https://api.githu... | [] | closed | false | null | [] | null | 0 | 2024-10-26T22:34:06 | 2024-11-04T00:53:51 | 2024-11-04T00:53:51 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7377",
"html_url": "https://github.com/ollama/ollama/pull/7377",
"diff_url": "https://github.com/ollama/ollama/pull/7377.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7377.patch",
"merged_at": "2024-11-04T00:53:51"
} | Hey everyone! I've recently been working on an extension for Word that aims to be a local, privacy-friendly alternative to Microsoft 365 Copilot by utilizing Ollama as the backend. I would like to introduce TextCraft, which is an add-in for Word that seamlessly integrates essential AI tools, including text generation, ... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7377/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7377/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6713 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6713/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6713/comments | https://api.github.com/repos/ollama/ollama/issues/6713/events | https://github.com/ollama/ollama/issues/6713 | 2,514,609,406 | I_kwDOJ0Z1Ps6V4eT- | 6,713 | Talking to Mistral-Nemo via OpenAI tool calling - fails | {
"login": "ChristianWeyer",
"id": 888718,
"node_id": "MDQ6VXNlcjg4ODcxOA==",
"avatar_url": "https://avatars.githubusercontent.com/u/888718?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ChristianWeyer",
"html_url": "https://github.com/ChristianWeyer",
"followers_url": "https://api.github... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q... | open | false | {
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/... | [
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "htt... | null | 10 | 2024-09-09T18:12:02 | 2025-01-16T14:54:23 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
With this curl command:
```
curl http://localhost:11434/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{"model":"mistral-nemo:12b-instruct-2407-fp16",
"messages": [
{
"role": "user",
"content": "What is the weather like in Boston?"
}
],
"to... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6713/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6713/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/4253 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4253/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4253/comments | https://api.github.com/repos/ollama/ollama/issues/4253/events | https://github.com/ollama/ollama/issues/4253 | 2,284,818,276 | I_kwDOJ0Z1Ps6IL49k | 4,253 | A repeatable hang issue on Linux - dual radeon | {
"login": "eliranwong",
"id": 25262722,
"node_id": "MDQ6VXNlcjI1MjYyNzIy",
"avatar_url": "https://avatars.githubusercontent.com/u/25262722?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/eliranwong",
"html_url": "https://github.com/eliranwong",
"followers_url": "https://api.github.com/use... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6433346500,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf3UTxA... | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 5 | 2024-05-08T06:49:58 | 2024-05-09T22:30:37 | 2024-05-09T22:08:58 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Experience a hang issue consistently.
Device information:
OS: Ubuntu, CPU: AMD Threadripper [AMD Ryzen Threadripper 7960X, 24 Cores, 48 Threads, 4.2GHz Base, 5.3GHz Turbo], Memory: 256GB RAM, Two GPUs: AMD RX 7900XTX + AMD RX 7900XTX
To reproduce the hang issue:
1. ollama run command... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4253/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4253/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3000 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3000/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3000/comments | https://api.github.com/repos/ollama/ollama/issues/3000/events | https://github.com/ollama/ollama/issues/3000 | 2,175,452,464 | I_kwDOJ0Z1Ps6BqsUw | 3,000 | Server hangs with no responsewhen running `gemma` | {
"login": "songsh",
"id": 2272252,
"node_id": "MDQ6VXNlcjIyNzIyNTI=",
"avatar_url": "https://avatars.githubusercontent.com/u/2272252?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/songsh",
"html_url": "https://github.com/songsh",
"followers_url": "https://api.github.com/users/songsh/foll... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q... | closed | false | null | [] | null | 7 | 2024-03-08T07:29:25 | 2024-05-02T22:33:21 | 2024-05-02T22:33:21 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | serve dead, how i to check the problem ,where is logs | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3000/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3000/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7798 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7798/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7798/comments | https://api.github.com/repos/ollama/ollama/issues/7798/events | https://github.com/ollama/ollama/issues/7798 | 2,683,215,299 | I_kwDOJ0Z1Ps6f7p3D | 7,798 | Is this a bug? (2GB model -> up to 20GB pagefile) | {
"login": "sebkont",
"id": 189359503,
"node_id": "U_kgDOC0lljw",
"avatar_url": "https://avatars.githubusercontent.com/u/189359503?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sebkont",
"html_url": "https://github.com/sebkont",
"followers_url": "https://api.github.com/users/sebkont/foll... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 8 | 2024-11-22T13:22:13 | 2024-12-02T15:36:27 | 2024-12-02T15:36:27 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
My GPU is old (GTX 1070) with 8GB, but should still be enough for running a model based on Phi 3 Mini? [This one ](https://huggingface.co/v8karlo/UNCENSORED-Phi-3-mini-4k-geminified-Q4_K_M-GGUF)
Unfortunately what happens is 'ollama ps' says 20 GB 63%/37% CPU/GPU + C:/ drive instantly gets ... | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7798/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7798/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5635 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5635/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5635/comments | https://api.github.com/repos/ollama/ollama/issues/5635/events | https://github.com/ollama/ollama/issues/5635 | 2,403,793,783 | I_kwDOJ0Z1Ps6PRvt3 | 5,635 | ollama not use all GPUs | {
"login": "mavershang",
"id": 8919917,
"node_id": "MDQ6VXNlcjg5MTk5MTc=",
"avatar_url": "https://avatars.githubusercontent.com/u/8919917?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mavershang",
"html_url": "https://github.com/mavershang",
"followers_url": "https://api.github.com/users... | [
{
"id": 5667396191,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aXw",
"url": "https://api.github.com/repos/ollama/ollama/labels/documentation",
"name": "documentation",
"color": "0075ca",
"default": true,
"description": "Improvements or additions to documentation"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 2 | 2024-07-11T18:12:48 | 2024-07-29T21:25:42 | 2024-07-29T21:25:42 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I ran ollama on a server with 4x A100. It only uses 1 of them. Is there some setting need to be changed? Thanks

### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5635/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5635/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/49 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/49/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/49/comments | https://api.github.com/repos/ollama/ollama/issues/49/events | https://github.com/ollama/ollama/pull/49 | 1,792,432,386 | PR_kwDOJ0Z1Ps5U26Op | 49 | Go run | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 1 | 2023-07-06T23:03:39 | 2023-07-07T00:19:03 | 2023-07-07T00:18:58 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/49",
"html_url": "https://github.com/ollama/ollama/pull/49",
"diff_url": "https://github.com/ollama/ollama/pull/49.diff",
"patch_url": "https://github.com/ollama/ollama/pull/49.patch",
"merged_at": "2023-07-07T00:18:58"
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/49/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/49/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5098 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5098/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5098/comments | https://api.github.com/repos/ollama/ollama/issues/5098/events | https://github.com/ollama/ollama/pull/5098 | 2,357,217,804 | PR_kwDOJ0Z1Ps5yr-W8 | 5,098 | feat: support setting the KV cache quant type | {
"login": "sammcj",
"id": 862951,
"node_id": "MDQ6VXNlcjg2Mjk1MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sammcj",
"html_url": "https://github.com/sammcj",
"followers_url": "https://api.github.com/users/sammcj/follow... | [] | closed | false | null | [] | null | 3 | 2024-06-17T12:24:22 | 2024-06-29T01:17:07 | 2024-06-28T21:50:53 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | true | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5098",
"html_url": "https://github.com/ollama/ollama/pull/5098",
"diff_url": "https://github.com/ollama/ollama/pull/5098.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5098.patch",
"merged_at": null
} | WIP
Testing adding configuration to allow setting the KV cache type re: #5091
---
- Allow setting the KV cache type in the env and params.
- Allow setting flashattention in params (as well as the existing env). | {
"login": "sammcj",
"id": 862951,
"node_id": "MDQ6VXNlcjg2Mjk1MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sammcj",
"html_url": "https://github.com/sammcj",
"followers_url": "https://api.github.com/users/sammcj/follow... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5098/reactions",
"total_count": 2,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 2,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5098/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/637 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/637/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/637/comments | https://api.github.com/repos/ollama/ollama/issues/637/events | https://github.com/ollama/ollama/pull/637 | 1,918,203,306 | PR_kwDOJ0Z1Ps5bex-l | 637 | windows runner fixes | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 0 | 2023-09-28T20:13:13 | 2023-09-29T15:47:56 | 2023-09-29T15:47:55 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/637",
"html_url": "https://github.com/ollama/ollama/pull/637",
"diff_url": "https://github.com/ollama/ollama/pull/637.diff",
"patch_url": "https://github.com/ollama/ollama/pull/637.patch",
"merged_at": "2023-09-29T15:47:55"
} | - use filepath for runner files
- get embedded files with unix filepath
- runner is only available is embedded directories have files | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/637/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/637/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2490 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2490/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2490/comments | https://api.github.com/repos/ollama/ollama/issues/2490/events | https://github.com/ollama/ollama/issues/2490 | 2,134,150,180 | I_kwDOJ0Z1Ps5_NIwk | 2,490 | [Question] Do not offload to CPU RAM | {
"login": "freQuensy23-coder",
"id": 64750224,
"node_id": "MDQ6VXNlcjY0NzUwMjI0",
"avatar_url": "https://avatars.githubusercontent.com/u/64750224?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/freQuensy23-coder",
"html_url": "https://github.com/freQuensy23-coder",
"followers_url": "https... | [] | closed | false | null | [] | null | 3 | 2024-02-14T11:32:17 | 2024-03-16T19:39:26 | 2024-03-11T18:28:34 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | By default, after some time of inactivity, ollama will automatically be offloaded from GPU memory, that caused some latency, especially to large models) | {
"login": "hoyyeva",
"id": 63033505,
"node_id": "MDQ6VXNlcjYzMDMzNTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoyyeva",
"html_url": "https://github.com/hoyyeva",
"followers_url": "https://api.github.com/users/hoyyev... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2490/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2490/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/238 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/238/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/238/comments | https://api.github.com/repos/ollama/ollama/issues/238/events | https://github.com/ollama/ollama/issues/238 | 1,827,345,140 | I_kwDOJ0Z1Ps5s6xL0 | 238 | Ability to download LLAMA2 7b 32k context | {
"login": "jlarmstrongiv",
"id": 20903247,
"node_id": "MDQ6VXNlcjIwOTAzMjQ3",
"avatar_url": "https://avatars.githubusercontent.com/u/20903247?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jlarmstrongiv",
"html_url": "https://github.com/jlarmstrongiv",
"followers_url": "https://api.githu... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 5789807732,
"node_id": ... | closed | false | null | [] | null | 2 | 2023-07-29T06:23:31 | 2023-12-04T19:02:13 | 2023-12-04T19:02:12 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | - https://together.ai/blog/llama-2-7b-32k
- https://github.com/togethercomputer/OpenChatKit
- https://huggingface.co/togethercomputer/LLaMA-2-7B-32K | {
"login": "technovangelist",
"id": 633681,
"node_id": "MDQ6VXNlcjYzMzY4MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/technovangelist",
"html_url": "https://github.com/technovangelist",
"followers_url": "https://api.git... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/238/reactions",
"total_count": 4,
"+1": 4,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/238/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1365 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1365/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1365/comments | https://api.github.com/repos/ollama/ollama/issues/1365/events | https://github.com/ollama/ollama/issues/1365 | 2,022,633,560 | I_kwDOJ0Z1Ps54jvBY | 1,365 | llama_print_timings have disappeared from the logs. | {
"login": "madsamjp",
"id": 49611363,
"node_id": "MDQ6VXNlcjQ5NjExMzYz",
"avatar_url": "https://avatars.githubusercontent.com/u/49611363?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/madsamjp",
"html_url": "https://github.com/madsamjp",
"followers_url": "https://api.github.com/users/mad... | [] | closed | false | null | [] | null | 3 | 2023-12-03T17:26:16 | 2024-01-20T00:18:22 | 2024-01-20T00:18:21 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | In a previous version of Ollama, following the logs (on Linux using `journalctl -t ollama -f`) would give helpful information after the model has finished with its response (such as tokens per second).
e.g. this:
```
Dec 03 14:58:42 osm-server ollama[20658]: llama server listening at http://127.0.0.1:54457
Dec 03... | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1365/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1365/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6356 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6356/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6356/comments | https://api.github.com/repos/ollama/ollama/issues/6356/events | https://github.com/ollama/ollama/issues/6356 | 2,465,338,496 | I_kwDOJ0Z1Ps6S8hSA | 6,356 | AMD Multiple GPU support | {
"login": "VitalickS",
"id": 10177561,
"node_id": "MDQ6VXNlcjEwMTc3NTYx",
"avatar_url": "https://avatars.githubusercontent.com/u/10177561?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/VitalickS",
"html_url": "https://github.com/VitalickS",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg... | open | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 6 | 2024-08-14T09:25:57 | 2024-10-16T00:15:13 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### Hi,
I think the current AMD ROCm doesn’t work well with multiple video cards. I have an XTX 7900 (24GB) and an XT 7900 (20GB). My processor also has a small integrated GPU, but that shouldn’t be a problem.
When I try to load the model llama3.1:70b (39GB):
1. It doesn’t crash, but it has an infinite load ti... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6356/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6356/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/8661 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8661/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8661/comments | https://api.github.com/repos/ollama/ollama/issues/8661/events | https://github.com/ollama/ollama/issues/8661 | 2,818,282,626 | I_kwDOJ0Z1Ps6n-5SC | 8,661 | Will Ollama run on the NPU(ANE) of Apple M silicon? | {
"login": "imJack6",
"id": 58357771,
"node_id": "MDQ6VXNlcjU4MzU3Nzcx",
"avatar_url": "https://avatars.githubusercontent.com/u/58357771?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/imJack6",
"html_url": "https://github.com/imJack6",
"followers_url": "https://api.github.com/users/imJack... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2025-01-29T13:50:08 | 2025-01-29T13:50:08 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | RT | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8661/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8661/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/3119 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3119/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3119/comments | https://api.github.com/repos/ollama/ollama/issues/3119/events | https://github.com/ollama/ollama/issues/3119 | 2,184,596,868 | I_kwDOJ0Z1Ps6CNk2E | 3,119 | Tensor `token_embed.weight` has wrong shape | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 4 | 2024-03-13T18:01:27 | 2024-03-13T20:30:47 | 2024-03-13T18:21:34 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | 
split from #2753
| {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3119/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3119/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4932 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4932/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4932/comments | https://api.github.com/repos/ollama/ollama/issues/4932/events | https://github.com/ollama/ollama/issues/4932 | 2,341,666,209 | I_kwDOJ0Z1Ps6Lkv2h | 4,932 | Cant see installed models | {
"login": "ahgsql",
"id": 35695543,
"node_id": "MDQ6VXNlcjM1Njk1NTQz",
"avatar_url": "https://avatars.githubusercontent.com/u/35695543?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ahgsql",
"html_url": "https://github.com/ahgsql",
"followers_url": "https://api.github.com/users/ahgsql/fo... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q... | closed | false | null | [] | null | 4 | 2024-06-08T13:51:57 | 2024-08-10T05:41:42 | 2024-08-09T23:51:33 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I have 7 models installed and was using them till yesterday.
But now it re-tries to download them, even i have all manifests files and my blobs folder is over 18 GB.
After shutdown and restart of WSL, ollama is not running and i m trying with ollama serve cmd
### OS
Windows
### ... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4932/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4932/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5323 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5323/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5323/comments | https://api.github.com/repos/ollama/ollama/issues/5323/events | https://github.com/ollama/ollama/issues/5323 | 2,377,959,722 | I_kwDOJ0Z1Ps6NvMkq | 5,323 | Weird output with any typos in accepted commands | {
"login": "yoshimario",
"id": 8993080,
"node_id": "MDQ6VXNlcjg5OTMwODA=",
"avatar_url": "https://avatars.githubusercontent.com/u/8993080?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yoshimario",
"html_url": "https://github.com/yoshimario",
"followers_url": "https://api.github.com/users... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 3 | 2024-06-27T11:59:02 | 2024-06-27T22:31:00 | 2024-06-27T22:29:45 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Problem is program outputs weird text if the program experiences commands that are not in the program list. This should result in an error message instead of a never ending loop of erraneous output. This should be handled better instead of forcing the process to close using ctr + c.
`>>> e... | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjha... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5323/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5323/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4153 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4153/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4153/comments | https://api.github.com/repos/ollama/ollama/issues/4153/events | https://github.com/ollama/ollama/pull/4153 | 2,279,118,097 | PR_kwDOJ0Z1Ps5ujE7R | 4,153 | Add GPU usage | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 2 | 2024-05-04T17:10:08 | 2024-05-08T23:39:14 | 2024-05-08T23:39:11 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4153",
"html_url": "https://github.com/ollama/ollama/pull/4153",
"diff_url": "https://github.com/ollama/ollama/pull/4153.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4153.patch",
"merged_at": "2024-05-08T23:39:11"
} | Help users understand how much of the model fit into their GPU without having to resort to inspecting the server log
A few examples from different systems and models
```
eval rate: 4.40 tokens/s
gpu usage: 1 GPU (14/27 layers) 3.2 GB (2.0 GB GPU)
eval rate: 6.64 tokens/s
gpu... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4153/reactions",
"total_count": 3,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4153/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7815 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7815/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7815/comments | https://api.github.com/repos/ollama/ollama/issues/7815/events | https://github.com/ollama/ollama/issues/7815 | 2,687,656,860 | I_kwDOJ0Z1Ps6gMmOc | 7,815 | Any fine-tuning ways? | {
"login": "Niifuji",
"id": 111742025,
"node_id": "U_kgDOBqkMSQ",
"avatar_url": "https://avatars.githubusercontent.com/u/111742025?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Niifuji",
"html_url": "https://github.com/Niifuji",
"followers_url": "https://api.github.com/users/Niifuji/foll... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 1 | 2024-11-24T12:14:14 | 2024-12-23T07:57:42 | 2024-12-23T07:57:42 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | As I mentioned in the title, I want to "continuously" fine-tune a pre-trained model with my custom dataset and explore adding some "emotion" to it (not sure why this idea came to mind). If you have any features or suggestions for this, I would appreciate your input. | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7815/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7815/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5187 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5187/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5187/comments | https://api.github.com/repos/ollama/ollama/issues/5187/events | https://github.com/ollama/ollama/pull/5187 | 2,364,767,566 | PR_kwDOJ0Z1Ps5zF1G4 | 5,187 | fix: skip os.removeAll() in assets.go if no PID | {
"login": "joshyan1",
"id": 76125168,
"node_id": "MDQ6VXNlcjc2MTI1MTY4",
"avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joshyan1",
"html_url": "https://github.com/joshyan1",
"followers_url": "https://api.github.com/users/jos... | [] | closed | false | null | [] | null | 1 | 2024-06-20T15:48:16 | 2024-06-20T15:53:26 | 2024-06-20T15:49:39 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5187",
"html_url": "https://github.com/ollama/ollama/pull/5187",
"diff_url": "https://github.com/ollama/ollama/pull/5187.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5187.patch",
"merged_at": null
} | we accidentally deleted every directory in $TMPDIR that was in the form "ollama*". Used errorcheck with PID to ensure directory is ours before deleting.
Resolves: https://github.com/ollama/ollama/issues/5129 | {
"login": "joshyan1",
"id": 76125168,
"node_id": "MDQ6VXNlcjc2MTI1MTY4",
"avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joshyan1",
"html_url": "https://github.com/joshyan1",
"followers_url": "https://api.github.com/users/jos... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5187/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5187/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4131 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4131/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4131/comments | https://api.github.com/repos/ollama/ollama/issues/4131/events | https://github.com/ollama/ollama/issues/4131 | 2,278,060,992 | I_kwDOJ0Z1Ps6HyHPA | 4,131 | Error "timed out waiting for llama runner to start: " on larger models. | {
"login": "CalvesGEH",
"id": 42101564,
"node_id": "MDQ6VXNlcjQyMTAxNTY0",
"avatar_url": "https://avatars.githubusercontent.com/u/42101564?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/CalvesGEH",
"html_url": "https://github.com/CalvesGEH",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 45 | 2024-05-03T16:45:32 | 2024-12-18T05:50:45 | 2024-07-03T23:28:05 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I just setup Ollama on a fresh machine and am running into an issue starting Ollama on larger models.
I am running Ubuntu 22.04.4 LTS with 2 Nvidia Tesla P40 GPUs with Driver Version: 535.161.08 and CUDA Version: 12.2.
Small 8b models work great and have no issues but when I try something ... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4131/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4131/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5454 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5454/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5454/comments | https://api.github.com/repos/ollama/ollama/issues/5454/events | https://github.com/ollama/ollama/issues/5454 | 2,387,738,883 | I_kwDOJ0Z1Ps6OUgED | 5,454 | When can we perform function calls like OpenAI? | {
"login": "qq1005894049",
"id": 48113255,
"node_id": "MDQ6VXNlcjQ4MTEzMjU1",
"avatar_url": "https://avatars.githubusercontent.com/u/48113255?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/qq1005894049",
"html_url": "https://github.com/qq1005894049",
"followers_url": "https://api.github.c... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 2 | 2024-07-03T05:57:43 | 2024-07-30T17:25:18 | 2024-07-30T17:25:18 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | <img width="1341" alt="image" src="https://github.com/ollama/ollama/assets/48113255/70ca615f-aae2-4b48-bd8f-c913b2ede23e">
| {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjha... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5454/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5454/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8110 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8110/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8110/comments | https://api.github.com/repos/ollama/ollama/issues/8110/events | https://github.com/ollama/ollama/issues/8110 | 2,741,302,826 | I_kwDOJ0Z1Ps6jZPYq | 8,110 | Support llama.cpp's Control Vector Functionality | {
"login": "amyb-asu",
"id": 156008468,
"node_id": "U_kgDOCUyAFA",
"avatar_url": "https://avatars.githubusercontent.com/u/156008468?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/amyb-asu",
"html_url": "https://github.com/amyb-asu",
"followers_url": "https://api.github.com/users/amyb-asu/... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 3 | 2024-12-16T04:30:35 | 2024-12-18T00:30:05 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | llama.cpp added support for control vectors a while ago https://github.com/ggerganov/llama.cpp/pull/5970
They can be loaded via `llama_control_vector_load` and `llama_control_vector_apply` which can take a vector in the form of a `.gguf`
https://github.com/ollama/ollama/blob/main/llama/common.h#L645
https://gith... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8110/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8110/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/64 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/64/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/64/comments | https://api.github.com/repos/ollama/ollama/issues/64/events | https://github.com/ollama/ollama/pull/64 | 1,796,593,917 | PR_kwDOJ0Z1Ps5VE8wr | 64 | Do not seg fault on client disconnect | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 0 | 2023-07-10T11:46:30 | 2023-07-11T14:19:33 | 2023-07-10T15:00:45 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/64",
"html_url": "https://github.com/ollama/ollama/pull/64",
"diff_url": "https://github.com/ollama/ollama/pull/64.diff",
"patch_url": "https://github.com/ollama/ollama/pull/64.patch",
"merged_at": "2023-07-10T15:00:44"
} | This was nicer to fix on the revised `b2` branch, so this is a pull request into that simplified change | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/64/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/64/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3494 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3494/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3494/comments | https://api.github.com/repos/ollama/ollama/issues/3494/events | https://github.com/ollama/ollama/pull/3494 | 2,226,051,077 | PR_kwDOJ0Z1Ps5rvMkF | 3,494 | Fail fast if mingw missing on windows | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 1 | 2024-04-04T16:52:06 | 2024-04-04T17:15:44 | 2024-04-04T17:15:40 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3494",
"html_url": "https://github.com/ollama/ollama/pull/3494",
"diff_url": "https://github.com/ollama/ollama/pull/3494.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3494.patch",
"merged_at": "2024-04-04T17:15:40"
} | null | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3494/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3494/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4974 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4974/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4974/comments | https://api.github.com/repos/ollama/ollama/issues/4974/events | https://github.com/ollama/ollama/issues/4974 | 2,345,961,529 | I_kwDOJ0Z1Ps6L1Ig5 | 4,974 | panic: runtime error: invalid memory address or nil pointer dereference | {
"login": "wywself",
"id": 8843053,
"node_id": "MDQ6VXNlcjg4NDMwNTM=",
"avatar_url": "https://avatars.githubusercontent.com/u/8843053?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/wywself",
"html_url": "https://github.com/wywself",
"followers_url": "https://api.github.com/users/wywself/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2024-06-11T09:52:18 | 2024-06-12T02:04:31 | 2024-06-12T02:04:31 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I am using Tesla M60, which is on the GPU card support list. However, when I execute the following command to start the model, an error is reported as follows.
```
# ollama run qwen:7b
Error: Post "http://127.0.0.1:11434/api/chat": EOF
```
The log is as follows:

| {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/860/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/860/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7010 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7010/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7010/comments | https://api.github.com/repos/ollama/ollama/issues/7010/events | https://github.com/ollama/ollama/pull/7010 | 2,553,850,810 | PR_kwDOJ0Z1Ps58-L8j | 7,010 | llama: Fix directory for conditional flash attention patch | {
"login": "jessegross",
"id": 6468499,
"node_id": "MDQ6VXNlcjY0Njg0OTk=",
"avatar_url": "https://avatars.githubusercontent.com/u/6468499?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jessegross",
"html_url": "https://github.com/jessegross",
"followers_url": "https://api.github.com/users... | [] | closed | false | null | [] | null | 1 | 2024-09-27T22:52:40 | 2024-10-10T21:38:26 | 2024-09-30T19:41:33 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7010",
"html_url": "https://github.com/ollama/ollama/pull/7010",
"diff_url": "https://github.com/ollama/ollama/pull/7010.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7010.patch",
"merged_at": null
} | Patches are against the llama.cpp directory structure, otherwise sync.sh can't apply them. | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7010/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7010/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/523 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/523/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/523/comments | https://api.github.com/repos/ollama/ollama/issues/523/events | https://github.com/ollama/ollama/issues/523 | 1,894,010,611 | I_kwDOJ0Z1Ps5w5E7z | 523 | LLM falcon:text infinity loop | {
"login": "dcasota",
"id": 14890243,
"node_id": "MDQ6VXNlcjE0ODkwMjQz",
"avatar_url": "https://avatars.githubusercontent.com/u/14890243?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dcasota",
"html_url": "https://github.com/dcasota",
"followers_url": "https://api.github.com/users/dcasot... | [] | closed | false | null | [] | null | 3 | 2023-09-13T08:26:09 | 2023-09-13T14:32:53 | 2023-09-13T13:56:46 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi,
I was trying to run falcon, but it responds ... weired.
Setup recipe.
```
git clone https://github.com/jmorganca/ollama
cd .\ollama
mkdir ..\.ollama
go generate .\...
go build .
```
Then, start the server component of ollama.
`start "Ollama server component" ollama.exe serve`
Download the select... | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/523/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/523/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3441 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3441/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3441/comments | https://api.github.com/repos/ollama/ollama/issues/3441/events | https://github.com/ollama/ollama/issues/3441 | 2,218,582,612 | I_kwDOJ0Z1Ps6EPOJU | 3,441 | Download/Archive and move models offline | {
"login": "Solomin0",
"id": 37559666,
"node_id": "MDQ6VXNlcjM3NTU5NjY2",
"avatar_url": "https://avatars.githubusercontent.com/u/37559666?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Solomin0",
"html_url": "https://github.com/Solomin0",
"followers_url": "https://api.github.com/users/Sol... | [
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] | closed | false | null | [] | null | 10 | 2024-04-01T16:09:15 | 2024-10-21T08:22:19 | 2024-05-10T20:18:11 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What are you trying to do?
I would like to be able to move ollama models between environments that are offline. There does not seem to be a supported official way to do this.
### How should we solve this?
An ollama archive command would be great! Then the user could just ollama pull from the path the arc... | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3441/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3441/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2051 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2051/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2051/comments | https://api.github.com/repos/ollama/ollama/issues/2051/events | https://github.com/ollama/ollama/issues/2051 | 2,088,450,843 | I_kwDOJ0Z1Ps58ezsb | 2,051 | Mixtral : How to connect to the Web | {
"login": "ymoymo",
"id": 10183941,
"node_id": "MDQ6VXNlcjEwMTgzOTQx",
"avatar_url": "https://avatars.githubusercontent.com/u/10183941?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ymoymo",
"html_url": "https://github.com/ymoymo",
"followers_url": "https://api.github.com/users/ymoymo/fo... | [] | closed | false | null | [] | null | 2 | 2024-01-18T14:50:38 | 2024-03-11T18:13:47 | 2024-03-11T18:13:47 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi,
I want to modify scipt to get this service, but I can't find the docker id or name that run Mixtral instance.
sudo docker ps return nothing while Mixtral is running.
Is there somthing I don't understand ?
Thx for any help.
Linux Pop Os | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2051/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2051/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3651 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3651/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3651/comments | https://api.github.com/repos/ollama/ollama/issues/3651/events | https://github.com/ollama/ollama/pull/3651 | 2,243,436,782 | PR_kwDOJ0Z1Ps5sqeHc | 3,651 | If OLLAMA_CONTAINER_MANAGER is set, only install NVIDIA drivers | {
"login": "ericcurtin",
"id": 1694275,
"node_id": "MDQ6VXNlcjE2OTQyNzU=",
"avatar_url": "https://avatars.githubusercontent.com/u/1694275?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ericcurtin",
"html_url": "https://github.com/ericcurtin",
"followers_url": "https://api.github.com/users... | [] | open | false | null | [] | null | 2 | 2024-04-15T11:33:41 | 2024-04-16T08:22:58 | null | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3651",
"html_url": "https://github.com/ollama/ollama/pull/3651",
"diff_url": "https://github.com/ollama/ollama/pull/3651.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3651.patch",
"merged_at": null
} | If installing for containerized environment, we should not have to install ollama binary, configure systemd, install rocm, etc.
Intended to be run like this:
curl -fsSL https://ollama.com/install.sh | OLLAMA_CONTAINER_MANAGER=podman sh | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3651/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3651/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3142 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3142/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3142/comments | https://api.github.com/repos/ollama/ollama/issues/3142/events | https://github.com/ollama/ollama/pull/3142 | 2,186,567,894 | PR_kwDOJ0Z1Ps5ppF8X | 3,142 | doc: faq gpu compatibility | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 3 | 2024-03-14T14:47:11 | 2024-03-21T09:21:35 | 2024-03-21T09:21:34 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3142",
"html_url": "https://github.com/ollama/ollama/pull/3142",
"diff_url": "https://github.com/ollama/ollama/pull/3142.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3142.patch",
"merged_at": "2024-03-21T09:21:34"
} | Add some information about GPU compatibility to the FAQs. | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3142/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3142/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/13 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/13/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/13/comments | https://api.github.com/repos/ollama/ollama/issues/13/events | https://github.com/ollama/ollama/pull/13 | 1,779,612,155 | PR_kwDOJ0Z1Ps5ULXaq | 13 | update development.md | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-06-28T19:30:01 | 2023-06-28T19:44:59 | 2023-06-28T19:44:56 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/13",
"html_url": "https://github.com/ollama/ollama/pull/13",
"diff_url": "https://github.com/ollama/ollama/pull/13.diff",
"patch_url": "https://github.com/ollama/ollama/pull/13.patch",
"merged_at": "2023-06-28T19:44:56"
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/13/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/13/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1838 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1838/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1838/comments | https://api.github.com/repos/ollama/ollama/issues/1838/events | https://github.com/ollama/ollama/issues/1838 | 2,069,059,120 | I_kwDOJ0Z1Ps57U1Yw | 1,838 | Cuda Error with 2GB VRAM: `Error: Post "http://127.0.0.1:11434/api/generate": EOF` | {
"login": "falaimo",
"id": 29931008,
"node_id": "MDQ6VXNlcjI5OTMxMDA4",
"avatar_url": "https://avatars.githubusercontent.com/u/29931008?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/falaimo",
"html_url": "https://github.com/falaimo",
"followers_url": "https://api.github.com/users/falaim... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 8 | 2024-01-07T09:38:40 | 2024-01-08T21:42:01 | 2024-01-08T21:42:01 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hello everyone, in Ollama version 0.1.18, I'm encountering the error "Error: Post "http://127.0.0.1:11434/api/generate": EOF" when starting Ollama with any model. I think it depends of cuda...
[logs_ollama.txt](https://github.com/jmorganca/ollama/files/13852832/logs_ollama.txt)
| {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1838/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1838/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6061 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6061/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6061/comments | https://api.github.com/repos/ollama/ollama/issues/6061/events | https://github.com/ollama/ollama/issues/6061 | 2,436,319,840 | I_kwDOJ0Z1Ps6RN0pg | 6,061 | [Feature Request] Force function calling for a model | {
"login": "mak448a",
"id": 94062293,
"node_id": "U_kgDOBZtG1Q",
"avatar_url": "https://avatars.githubusercontent.com/u/94062293?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mak448a",
"html_url": "https://github.com/mak448a",
"followers_url": "https://api.github.com/users/mak448a/follow... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 4 | 2024-07-29T20:43:00 | 2025-01-06T07:17:25 | 2025-01-06T07:17:24 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Not sure if this belongs in ollama-python or here, but I'll open it here. Could you add a way to use function calling on any model, or is this something that the model itself has to support? | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6061/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6061/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2096 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2096/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2096/comments | https://api.github.com/repos/ollama/ollama/issues/2096/events | https://github.com/ollama/ollama/issues/2096 | 2,090,734,716 | I_kwDOJ0Z1Ps58nhR8 | 2,096 | How is Tinyllama on Ollama trained? | {
"login": "oliverbob",
"id": 23272429,
"node_id": "MDQ6VXNlcjIzMjcyNDI5",
"avatar_url": "https://avatars.githubusercontent.com/u/23272429?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/oliverbob",
"html_url": "https://github.com/oliverbob",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 9 | 2024-01-19T14:58:44 | 2024-03-18T20:47:59 | 2024-02-20T22:51:52 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi everyone, as always, thank you for the great work you have done with this project for the good of humanity. I have tried importing gguf file using tintyllama on huggingface, but when I chat with it using ollama, it returns gibberish talk. But when I download the one from Ollama with ollama pull/run tinyllama, it wor... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2096/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2096/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3765 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3765/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3765/comments | https://api.github.com/repos/ollama/ollama/issues/3765/events | https://github.com/ollama/ollama/issues/3765 | 2,253,970,292 | I_kwDOJ0Z1Ps6GWNt0 | 3,765 | CUDA error: out of memory - other VRAM consumers not detected in available memory | {
"login": "martinus",
"id": 14386,
"node_id": "MDQ6VXNlcjE0Mzg2",
"avatar_url": "https://avatars.githubusercontent.com/u/14386?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/martinus",
"html_url": "https://github.com/martinus",
"followers_url": "https://api.github.com/users/martinus/foll... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5755339642,
"node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg... | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 18 | 2024-04-19T20:47:24 | 2024-06-14T22:35:02 | 2024-06-14T22:35:02 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
When I try the llama3 model I get out of memory errors. I have 64GB of RAM and 24GB on the GPU.
```
❯ ollama run llama3:70b-instruct-q2_K --verbose "write a constexpr GCD that is not recursive in C++17"
Error: an unknown error was encountered while running the model CUDA error: out of memo... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3765/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3765/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8319 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8319/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8319/comments | https://api.github.com/repos/ollama/ollama/issues/8319/events | https://github.com/ollama/ollama/pull/8319 | 2,770,591,352 | PR_kwDOJ0Z1Ps6G0bib | 8,319 | Add Safetensor Conversion for Granite Models | {
"login": "alex-jw-brooks",
"id": 10740300,
"node_id": "MDQ6VXNlcjEwNzQwMzAw",
"avatar_url": "https://avatars.githubusercontent.com/u/10740300?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/alex-jw-brooks",
"html_url": "https://github.com/alex-jw-brooks",
"followers_url": "https://api.gi... | [] | open | false | null | [] | null | 0 | 2025-01-06T12:48:04 | 2025-01-16T04:20:40 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8319",
"html_url": "https://github.com/ollama/ollama/pull/8319",
"diff_url": "https://github.com/ollama/ollama/pull/8319.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8319.patch",
"merged_at": null
} | This PR fixes the unrecognized architecture for converting Granite Models (`GraniteForCausalLM`) for use from safetensors. | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8319/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8319/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1430 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1430/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1430/comments | https://api.github.com/repos/ollama/ollama/issues/1430/events | https://github.com/ollama/ollama/issues/1430 | 2,031,816,063 | I_kwDOJ0Z1Ps55Gw1_ | 1,430 | coda error 222 after building | {
"login": "rhettg",
"id": 50074,
"node_id": "MDQ6VXNlcjUwMDc0",
"avatar_url": "https://avatars.githubusercontent.com/u/50074?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rhettg",
"html_url": "https://github.com/rhettg",
"followers_url": "https://api.github.com/users/rhettg/followers",
... | [
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg",
"url": "https://api.github.com/repos/ollama/ollama/labels/nvidia",
"name": "nvidia",
"color": "8CDB00",
"default": false,
"description": "Issues relating to Nvidia GPUs and CUDA"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 3 | 2023-12-08T02:14:32 | 2024-02-01T23:15:41 | 2024-02-01T23:15:41 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | This might be a llama.cpp question, but I'm struggling to get Ollama to work when I build it myself.
The release builds work fine for me:
```console
$ sudo -u ollama /usr/bin/ollama serve
2023/12/07 17:52:41 images.go:779: total blobs: 10
2023/12/07 17:52:41 images.go:786: total unused blobs removed: 0
2023/... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1430/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1430/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8662 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8662/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8662/comments | https://api.github.com/repos/ollama/ollama/issues/8662/events | https://github.com/ollama/ollama/pull/8662 | 2,818,376,666 | PR_kwDOJ0Z1Ps6JXtvk | 8,662 | Update README.md Adding DeepSeek to the table of models | {
"login": "teymuur",
"id": 64795612,
"node_id": "MDQ6VXNlcjY0Nzk1NjEy",
"avatar_url": "https://avatars.githubusercontent.com/u/64795612?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/teymuur",
"html_url": "https://github.com/teymuur",
"followers_url": "https://api.github.com/users/teymuu... | [] | closed | false | null | [] | null | 0 | 2025-01-29T14:25:25 | 2025-01-29T14:33:35 | 2025-01-29T14:33:28 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8662",
"html_url": "https://github.com/ollama/ollama/pull/8662",
"diff_url": "https://github.com/ollama/ollama/pull/8662.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8662.patch",
"merged_at": null
} | This is just a minor change, I added DeepSeek R1 to the model library table. Only changed `README.md` | {
"login": "teymuur",
"id": 64795612,
"node_id": "MDQ6VXNlcjY0Nzk1NjEy",
"avatar_url": "https://avatars.githubusercontent.com/u/64795612?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/teymuur",
"html_url": "https://github.com/teymuur",
"followers_url": "https://api.github.com/users/teymuu... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8662/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8662/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5631 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5631/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5631/comments | https://api.github.com/repos/ollama/ollama/issues/5631/events | https://github.com/ollama/ollama/pull/5631 | 2,403,403,309 | PR_kwDOJ0Z1Ps51G6eZ | 5,631 | Refactor linux packaging | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 3 | 2024-07-11T14:56:45 | 2024-08-17T17:16:53 | 2024-08-17T17:16:45 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5631",
"html_url": "https://github.com/ollama/ollama/pull/5631",
"diff_url": "https://github.com/ollama/ollama/pull/5631.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5631.patch",
"merged_at": null
} | This adjusts linux to follow a similar model to windows with a discrete archive (zip/tgz) to cary the primary executable, and dependent libraries. Runners are still carried as payloads inside the main binary.
As Darwin has no significant dependent libraries, it still functions as a discrete stand-alone executable ca... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5631/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5631/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1112 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1112/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1112/comments | https://api.github.com/repos/ollama/ollama/issues/1112/events | https://github.com/ollama/ollama/issues/1112 | 1,991,164,963 | I_kwDOJ0Z1Ps52rsQj | 1,112 | Support `ollama create` with PyTorch | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 1 | 2023-11-13T17:53:57 | 2024-05-06T23:26:01 | 2024-05-06T23:26:01 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Currently, create a model via a `Modelfile` supports importing GGUF format model binaries. Ollama should also support importing PyTorch models directly via `ollama create`
Related:
* https://github.com/jmorganca/ollama/issues/1037
* https://github.com/jmorganca/ollama/issues/1097
| {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1112/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1112/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/112 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/112/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/112/comments | https://api.github.com/repos/ollama/ollama/issues/112/events | https://github.com/ollama/ollama/pull/112 | 1,811,022,006 | PR_kwDOJ0Z1Ps5V18xY | 112 | resolve modelfile before passing to server | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-07-19T02:34:32 | 2023-07-19T02:36:27 | 2023-07-19T02:36:25 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/112",
"html_url": "https://github.com/ollama/ollama/pull/112",
"diff_url": "https://github.com/ollama/ollama/pull/112.diff",
"patch_url": "https://github.com/ollama/ollama/pull/112.patch",
"merged_at": "2023-07-19T02:36:25"
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/112/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/112/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4095 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4095/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4095/comments | https://api.github.com/repos/ollama/ollama/issues/4095/events | https://github.com/ollama/ollama/issues/4095 | 2,274,734,176 | I_kwDOJ0Z1Ps6HlbBg | 4,095 | Is there a problem with the document? | {
"login": "ggjk616",
"id": 168710680,
"node_id": "U_kgDOCg5SGA",
"avatar_url": "https://avatars.githubusercontent.com/u/168710680?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ggjk616",
"html_url": "https://github.com/ggjk616",
"followers_url": "https://api.github.com/users/ggjk616/foll... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 0 | 2024-05-02T06:37:54 | 2024-05-02T10:16:12 | 2024-05-02T10:16:12 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Can you help me,In the documentation, I noticed the following statement: “You can set OLLAMA_LLM_LIBRARY to any of the available LLM libraries to bypass autodetection, so for example, if you have a CUDA card, but want to force the CPU LLM library with AVX2 vector support, use:
OLLAMA_LLM_LIBRAR... | {
"login": "ggjk616",
"id": 168710680,
"node_id": "U_kgDOCg5SGA",
"avatar_url": "https://avatars.githubusercontent.com/u/168710680?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ggjk616",
"html_url": "https://github.com/ggjk616",
"followers_url": "https://api.github.com/users/ggjk616/foll... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4095/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4095/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4155 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4155/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4155/comments | https://api.github.com/repos/ollama/ollama/issues/4155/events | https://github.com/ollama/ollama/issues/4155 | 2,279,169,147 | I_kwDOJ0Z1Ps6H2Vx7 | 4,155 | Add option in the install scripts to auto set OLLAMA_HOST environment variable | {
"login": "centopw",
"id": 30675552,
"node_id": "MDQ6VXNlcjMwNjc1NTUy",
"avatar_url": "https://avatars.githubusercontent.com/u/30675552?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/centopw",
"html_url": "https://github.com/centopw",
"followers_url": "https://api.github.com/users/centop... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 1 | 2024-05-04T19:37:42 | 2024-05-09T21:16:30 | 2024-05-09T21:16:30 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | In the installer scripts, add a option that ask if user want to allow other machine on the same network to connect. Base on this docs: [faq](https://github.com/ollama/ollama/blob/main/docs/faq.md#setting-environment-variables-on-mac)
I can create a PR if needed | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4155/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4155/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3030 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3030/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3030/comments | https://api.github.com/repos/ollama/ollama/issues/3030/events | https://github.com/ollama/ollama/pull/3030 | 2,177,439,108 | PR_kwDOJ0Z1Ps5pJ9Nn | 3,030 | Update llama.cpp submodule to `77d1ac7` | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 0 | 2024-03-09T23:10:52 | 2024-03-09T23:55:35 | 2024-03-09T23:55:34 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3030",
"html_url": "https://github.com/ollama/ollama/pull/3030",
"diff_url": "https://github.com/ollama/ollama/pull/3030.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3030.patch",
"merged_at": "2024-03-09T23:55:34"
} | Note we use `-DLLAMA_METAL_EMBED_LIBRARY=on` on arm64 darwin to embed ggml-metal.metal. This change also required us to prepend ggml-common.h to the top of ggml-metal.metal to avoid a runtime lookup error. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3030/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3030/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1157 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1157/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1157/comments | https://api.github.com/repos/ollama/ollama/issues/1157/events | https://github.com/ollama/ollama/issues/1157 | 1,997,826,755 | I_kwDOJ0Z1Ps53FGrD | 1,157 | [Linux] - Instructions for exposing Ollama doesn't work | {
"login": "SoloBSD",
"id": 17459633,
"node_id": "MDQ6VXNlcjE3NDU5NjMz",
"avatar_url": "https://avatars.githubusercontent.com/u/17459633?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/SoloBSD",
"html_url": "https://github.com/SoloBSD",
"followers_url": "https://api.github.com/users/SoloBS... | [] | closed | false | null | [] | null | 8 | 2023-11-16T21:33:43 | 2023-11-17T01:09:35 | 2023-11-17T00:55:21 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Instructions for Linux on how to expose ollama doesn't work.
https://github.com/jmorganca/ollama/blob/main/docs/faq.md#how-can-i-expose-ollama-on-my-network
For some reason when Ollama gets installed on Linux it creates:
/etc/systemd/system/ollama.service
So it seems it never processes
/etc/systemd/system/o... | {
"login": "SoloBSD",
"id": 17459633,
"node_id": "MDQ6VXNlcjE3NDU5NjMz",
"avatar_url": "https://avatars.githubusercontent.com/u/17459633?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/SoloBSD",
"html_url": "https://github.com/SoloBSD",
"followers_url": "https://api.github.com/users/SoloBS... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1157/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1157/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2466 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2466/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2466/comments | https://api.github.com/repos/ollama/ollama/issues/2466/events | https://github.com/ollama/ollama/pull/2466 | 2,130,496,639 | PR_kwDOJ0Z1Ps5mp7JN | 2,466 | Added NextJS web interface for Ollama models to readme.md | {
"login": "jakobhoeg",
"id": 114422072,
"node_id": "U_kgDOBtHxOA",
"avatar_url": "https://avatars.githubusercontent.com/u/114422072?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jakobhoeg",
"html_url": "https://github.com/jakobhoeg",
"followers_url": "https://api.github.com/users/jakobh... | [] | closed | false | null | [] | null | 1 | 2024-02-12T16:26:42 | 2024-02-20T02:57:36 | 2024-02-20T02:57:36 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2466",
"html_url": "https://github.com/ollama/ollama/pull/2466",
"diff_url": "https://github.com/ollama/ollama/pull/2466.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2466.patch",
"merged_at": "2024-02-20T02:57:36"
} | Added [nextjs-ollama-llm-ui](https://github.com/jakobhoeg/nextjs-ollama-llm-ui) to the readme file. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2466/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2466/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2659 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2659/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2659/comments | https://api.github.com/repos/ollama/ollama/issues/2659/events | https://github.com/ollama/ollama/issues/2659 | 2,148,066,317 | I_kwDOJ0Z1Ps6ACOQN | 2,659 | Add phixtral | {
"login": "vprelovac",
"id": 4319401,
"node_id": "MDQ6VXNlcjQzMTk0MDE=",
"avatar_url": "https://avatars.githubusercontent.com/u/4319401?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/vprelovac",
"html_url": "https://github.com/vprelovac",
"followers_url": "https://api.github.com/users/vp... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | open | false | null | [] | null | 0 | 2024-02-22T02:39:04 | 2024-03-12T02:02:47 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Currently the best 2B model
https://huggingface.co/shadowml/phixtral-4x2_8odd | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2659/reactions",
"total_count": 3,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2659/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/2850 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2850/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2850/comments | https://api.github.com/repos/ollama/ollama/issues/2850/events | https://github.com/ollama/ollama/issues/2850 | 2,162,455,308 | I_kwDOJ0Z1Ps6A5HMM | 2,850 | `ollama push` and `ollama pull` are slow or hang on windows | {
"login": "ewebgh33",
"id": 123797054,
"node_id": "U_kgDOB2D-Pg",
"avatar_url": "https://avatars.githubusercontent.com/u/123797054?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ewebgh33",
"html_url": "https://github.com/ewebgh33",
"followers_url": "https://api.github.com/users/ewebgh33/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677370291,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCVsw... | closed | false | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/... | null | 10 | 2024-03-01T02:21:35 | 2024-08-06T18:10:32 | 2024-08-06T18:10:32 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Can't download ANY models.
What is happening? Not my internet, speed test blasts.
Your servers OK?
Happening on Windows version buggy still? Using latest, 0.1.27 (Win11).
As per docs, set Windows environment variable to:
OLLAMA_MODELS = D:\AI\text\ollama-models
I am familiar with environment variables and ... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2850/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2850/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5474 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5474/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5474/comments | https://api.github.com/repos/ollama/ollama/issues/5474/events | https://github.com/ollama/ollama/issues/5474 | 2,389,647,879 | I_kwDOJ0Z1Ps6ObyIH | 5,474 | InternLM2.5 - hallucinations - lot of repetitions etc | {
"login": "Qualzz",
"id": 35169816,
"node_id": "MDQ6VXNlcjM1MTY5ODE2",
"avatar_url": "https://avatars.githubusercontent.com/u/35169816?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Qualzz",
"html_url": "https://github.com/Qualzz",
"followers_url": "https://api.github.com/users/Qualzz/fo... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.g... | null | 5 | 2024-07-03T23:18:06 | 2024-10-04T17:07:56 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Seems like something is wrong with InternLM2.5, I can't get any meaningful out of it. (tried with 32k context)
### OS
Linux
### GPU
Nvidia
### CPU
AMD
### Ollama version
v0.1.48 | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5474/reactions",
"total_count": 2,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 2
} | https://api.github.com/repos/ollama/ollama/issues/5474/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/1169 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1169/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1169/comments | https://api.github.com/repos/ollama/ollama/issues/1169/events | https://github.com/ollama/ollama/issues/1169 | 1,998,422,439 | I_kwDOJ0Z1Ps53HYGn | 1,169 | Update the model name in the api doc | {
"login": "shenli",
"id": 1192573,
"node_id": "MDQ6VXNlcjExOTI1NzM=",
"avatar_url": "https://avatars.githubusercontent.com/u/1192573?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/shenli",
"html_url": "https://github.com/shenli",
"followers_url": "https://api.github.com/users/shenli/foll... | [] | closed | false | null | [] | null | 1 | 2023-11-17T07:10:44 | 2023-11-17T12:18:09 | 2023-11-17T12:18:09 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi, I am new to Ollama.
I followed the [Quickstart](https://github.com/jmorganca/ollama/tree/main#quickstart) to try Ollama with model Llama2. It is very easy to run and a very interesting project.
When I explored further in the [API doc](https://github.com/jmorganca/ollama/blob/main/docs/api.md), I found that t... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1169/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1169/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2517 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2517/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2517/comments | https://api.github.com/repos/ollama/ollama/issues/2517/events | https://github.com/ollama/ollama/issues/2517 | 2,137,087,723 | I_kwDOJ0Z1Ps5_YV7r | 2,517 | parser/parser.go:9:2: package log/slog is not in GOROOT (/usr/local/go120/src/log/slog) | {
"login": "yurivict",
"id": 271906,
"node_id": "MDQ6VXNlcjI3MTkwNg==",
"avatar_url": "https://avatars.githubusercontent.com/u/271906?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yurivict",
"html_url": "https://github.com/yurivict",
"followers_url": "https://api.github.com/users/yurivic... | [] | closed | false | null | [] | null | 1 | 2024-02-15T17:36:41 | 2024-02-15T19:51:08 | 2024-02-15T19:51:08 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Build fails:
```
===> Building for ollama-0.1.25
(cd /usr/ports/misc/ollama/work/github.com/ollama/ollama@v0.1.25; for t in ./cmd; do out=$(/usr/bin/basename $(echo ${t} | /usr/bin/sed -Ee 's/^[^:]*:([^:]+).*$/\1/' -e 's/^\.$/ollama/')); pkg=$(echo ${t} | /usr/bin/sed -Ee 's/^([^:]*).*$/\1/' -e 's/^ollama$/./'... | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2517/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2517/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7287 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7287/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7287/comments | https://api.github.com/repos/ollama/ollama/issues/7287/events | https://github.com/ollama/ollama/issues/7287 | 2,601,723,323 | I_kwDOJ0Z1Ps6bEyW7 | 7,287 | Version v0.3.14 impacted CPU inference performance | {
"login": "closesim",
"id": 9018799,
"node_id": "MDQ6VXNlcjkwMTg3OTk=",
"avatar_url": "https://avatars.githubusercontent.com/u/9018799?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/closesim",
"html_url": "https://github.com/closesim",
"followers_url": "https://api.github.com/users/close... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677677816,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgVG-A... | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 9 | 2024-10-21T08:10:06 | 2024-10-30T22:05:47 | 2024-10-30T22:05:47 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Hi, I just updated my docker container where I run my small models to the latest version, as I use to every 15 days or so. I'm using a Quad Core CPU (no GPU) and with this new version I noticed that LLama 3.1 8b performance was very slow. I Initially thought it was a hardware issue, like overhea... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7287/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7287/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1088 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1088/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1088/comments | https://api.github.com/repos/ollama/ollama/issues/1088/events | https://github.com/ollama/ollama/issues/1088 | 1,989,023,621 | I_kwDOJ0Z1Ps52jheF | 1,088 | Problems installing the docker image. | {
"login": "pdavis68",
"id": 2781885,
"node_id": "MDQ6VXNlcjI3ODE4ODU=",
"avatar_url": "https://avatars.githubusercontent.com/u/2781885?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdavis68",
"html_url": "https://github.com/pdavis68",
"followers_url": "https://api.github.com/users/pdavi... | [] | closed | false | null | [] | null | 0 | 2023-11-11T16:17:04 | 2023-11-11T16:19:32 | 2023-11-11T16:19:32 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | null | {
"login": "pdavis68",
"id": 2781885,
"node_id": "MDQ6VXNlcjI3ODE4ODU=",
"avatar_url": "https://avatars.githubusercontent.com/u/2781885?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdavis68",
"html_url": "https://github.com/pdavis68",
"followers_url": "https://api.github.com/users/pdavi... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1088/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1088/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3262 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3262/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3262/comments | https://api.github.com/repos/ollama/ollama/issues/3262/events | https://github.com/ollama/ollama/issues/3262 | 2,196,585,763 | I_kwDOJ0Z1Ps6C7T0j | 3,262 | Ollama can support windows 7? | {
"login": "zhaosd",
"id": 5444416,
"node_id": "MDQ6VXNlcjU0NDQ0MTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/5444416?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/zhaosd",
"html_url": "https://github.com/zhaosd",
"followers_url": "https://api.github.com/users/zhaosd/foll... | [] | closed | false | null | [] | null | 6 | 2024-03-20T03:12:53 | 2024-11-18T23:47:12 | 2024-03-20T07:41:50 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | null | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3262/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3262/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7124 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7124/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7124/comments | https://api.github.com/repos/ollama/ollama/issues/7124/events | https://github.com/ollama/ollama/pull/7124 | 2,571,787,113 | PR_kwDOJ0Z1Ps594dhh | 7,124 | llama: Decouple patching script from submodule | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-10-08T00:29:54 | 2024-10-08T16:21:35 | 2024-10-08T15:54:00 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7124",
"html_url": "https://github.com/ollama/ollama/pull/7124",
"diff_url": "https://github.com/ollama/ollama/pull/7124.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7124.patch",
"merged_at": null
} | Replaced by #7139 on main | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7124/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7124/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/862 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/862/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/862/comments | https://api.github.com/repos/ollama/ollama/issues/862/events | https://github.com/ollama/ollama/pull/862 | 1,955,129,810 | PR_kwDOJ0Z1Ps5dbU1e | 862 | fix/Predict: A prediction should use the options sent with the request | {
"login": "CyrilPeponnet",
"id": 2277387,
"node_id": "MDQ6VXNlcjIyNzczODc=",
"avatar_url": "https://avatars.githubusercontent.com/u/2277387?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/CyrilPeponnet",
"html_url": "https://github.com/CyrilPeponnet",
"followers_url": "https://api.github.... | [] | closed | false | null | [] | null | 2 | 2023-10-20T23:26:14 | 2023-10-26T15:07:42 | 2023-10-26T15:07:42 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/862",
"html_url": "https://github.com/ollama/ollama/pull/862",
"diff_url": "https://github.com/ollama/ollama/pull/862.diff",
"patch_url": "https://github.com/ollama/ollama/pull/862.patch",
"merged_at": null
} | Consecutive query to the same running model should use the client request parameters instead of the one set during the model loading. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/862/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/862/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6799 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6799/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6799/comments | https://api.github.com/repos/ollama/ollama/issues/6799/events | https://github.com/ollama/ollama/issues/6799 | 2,526,152,209 | I_kwDOJ0Z1Ps6WkgYR | 6,799 | Is it possible to configure ollama deployed in docker? | {
"login": "wizounovziki",
"id": 42036658,
"node_id": "MDQ6VXNlcjQyMDM2NjU4",
"avatar_url": "https://avatars.githubusercontent.com/u/42036658?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/wizounovziki",
"html_url": "https://github.com/wizounovziki",
"followers_url": "https://api.github.c... | [
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
},
{
"id": 6677677816,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjg... | closed | false | null | [] | null | 1 | 2024-09-14T09:29:47 | 2024-09-25T21:23:15 | 2024-09-25T21:23:15 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I pull docker image from dockerhub and launched a few models and then found the num of user requests was limited.
In the documentation it shows that this could be solved by set up OLLAMA_NUM_PARALLEL by systemctl commands.
How can I do this since systemctl is not included in the docker container?
| {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6799/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6799/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1499 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1499/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1499/comments | https://api.github.com/repos/ollama/ollama/issues/1499/events | https://github.com/ollama/ollama/issues/1499 | 2,039,445,848 | I_kwDOJ0Z1Ps55j3lY | 1,499 | Add mistral's new 7B-instruct-v0.2 | {
"login": "tarek-ayed",
"id": 45576986,
"node_id": "MDQ6VXNlcjQ1NTc2OTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/45576986?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tarek-ayed",
"html_url": "https://github.com/tarek-ayed",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 4 | 2023-12-13T10:51:12 | 2023-12-14T03:11:15 | 2023-12-13T16:17:43 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Along with many releases, Mistral vastly improved their existing 7B model with a version named `v0.2`.
It has 32k context instead of 8k and better benchmark scores: https://x.com/dchaplot/status/1734198245067243629?s=20
More can be found here: https://docs.mistral.ai/platform/endpoints (see "Mistral Tiny")
The w... | {
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/us... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1499/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1499/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5399 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5399/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5399/comments | https://api.github.com/repos/ollama/ollama/issues/5399/events | https://github.com/ollama/ollama/issues/5399 | 2,382,949,738 | I_kwDOJ0Z1Ps6OCO1q | 5,399 | Please support models of rerank type | {
"login": "yushengliao",
"id": 29765903,
"node_id": "MDQ6VXNlcjI5NzY1OTAz",
"avatar_url": "https://avatars.githubusercontent.com/u/29765903?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yushengliao",
"html_url": "https://github.com/yushengliao",
"followers_url": "https://api.github.com/... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 10 | 2024-07-01T06:23:13 | 2024-09-02T20:51:51 | 2024-09-02T20:51:50 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | There are so many users for the Ollama project, why hasn't it been so long to support Renanker
Similar software such as Localai、xinreference already supports rerank
https://localai.io/features/reranker/
https://inference.readthedocs.io/en/latest/models/builtin/rerank/index.html
 anytime soon? | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1709/reactions",
"total_count": 22,
"+1": 7,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 3,
"eyes": 12
} | https://api.github.com/repos/ollama/ollama/issues/1709/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/6481 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6481/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6481/comments | https://api.github.com/repos/ollama/ollama/issues/6481/events | https://github.com/ollama/ollama/issues/6481 | 2,483,787,305 | I_kwDOJ0Z1Ps6UC5Yp | 6,481 | gork2 | {
"login": "olumolu",
"id": 162728301,
"node_id": "U_kgDOCbMJbQ",
"avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/olumolu",
"html_url": "https://github.com/olumolu",
"followers_url": "https://api.github.com/users/olumolu/foll... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 2 | 2024-08-23T19:46:43 | 2024-08-24T04:31:25 | 2024-08-23T20:35:17 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | New gork2 has already published can we have support for that thanks. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6481/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6481/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3844 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3844/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3844/comments | https://api.github.com/repos/ollama/ollama/issues/3844/events | https://github.com/ollama/ollama/issues/3844 | 2,258,843,905 | I_kwDOJ0Z1Ps6GozkB | 3,844 | api error occurred after some times request | {
"login": "Shiyaoa",
"id": 48488459,
"node_id": "MDQ6VXNlcjQ4NDg4NDU5",
"avatar_url": "https://avatars.githubusercontent.com/u/48488459?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Shiyaoa",
"html_url": "https://github.com/Shiyaoa",
"followers_url": "https://api.github.com/users/Shiyao... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 7 | 2024-04-23T13:09:20 | 2025-01-06T03:55:27 | 2024-04-25T11:50:41 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
i try to post request using the url http://localhost:11434/v1 and model "llama3:8b-instruct-q8_0", it works successfully at the initially first time, but then failed with these information:
Error occurred: Error code: 400 - {'error': {'message': 'unexpected server status: 1', 'type': 'api_erro... | {
"login": "Shiyaoa",
"id": 48488459,
"node_id": "MDQ6VXNlcjQ4NDg4NDU5",
"avatar_url": "https://avatars.githubusercontent.com/u/48488459?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Shiyaoa",
"html_url": "https://github.com/Shiyaoa",
"followers_url": "https://api.github.com/users/Shiyao... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3844/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3844/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4275 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4275/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4275/comments | https://api.github.com/repos/ollama/ollama/issues/4275/events | https://github.com/ollama/ollama/issues/4275 | 2,286,923,121 | I_kwDOJ0Z1Ps6IT61x | 4,275 | Degraded accuracy when using the nomic-embed-text (v1.5) model with Ollama versions 0.1.32 and 0.1.33 | {
"login": "Ganesh1030",
"id": 48667223,
"node_id": "MDQ6VXNlcjQ4NjY3MjIz",
"avatar_url": "https://avatars.githubusercontent.com/u/48667223?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Ganesh1030",
"html_url": "https://github.com/Ganesh1030",
"followers_url": "https://api.github.com/use... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2024-05-09T05:29:55 | 2024-06-26T05:47:57 | 2024-06-25T16:46:35 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
We have an application where we are training the CSV file and using the following things for it:.
- 'nomic-embed-text(v1.5) model
- chromadb
- ollama(0.1.31)
At runtime, we are using'similarity_search' and getting good accuracy with ollama version 0.1.31, but when we upgrade ollama vers... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4275/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4275/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/633 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/633/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/633/comments | https://api.github.com/repos/ollama/ollama/issues/633/events | https://github.com/ollama/ollama/pull/633 | 1,917,657,126 | PR_kwDOJ0Z1Ps5bc6m4 | 633 | do not download updates multiple times | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 0 | 2023-09-28T14:20:30 | 2023-09-28T19:29:18 | 2023-09-28T19:29:18 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/633",
"html_url": "https://github.com/ollama/ollama/pull/633",
"diff_url": "https://github.com/ollama/ollama/pull/633.diff",
"patch_url": "https://github.com/ollama/ollama/pull/633.patch",
"merged_at": "2023-09-28T19:29:18"
} | We've hit a bug in the Electron auto-updater that prevents the toolbar app from restarting after update when `autoUpdater.checkForUpdates()` is called more than once. The root cause of this is not clear, it may be related to [this Electron issue](https://github.com/electron-userland/electron-builder/issues/7800). In an... | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/633/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/633/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4395 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4395/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4395/comments | https://api.github.com/repos/ollama/ollama/issues/4395/events | https://github.com/ollama/ollama/issues/4395 | 2,292,283,708 | I_kwDOJ0Z1Ps6IoXk8 | 4,395 | Cannot Use GPU properly | {
"login": "applepieiris",
"id": 36785462,
"node_id": "MDQ6VXNlcjM2Nzg1NDYy",
"avatar_url": "https://avatars.githubusercontent.com/u/36785462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/applepieiris",
"html_url": "https://github.com/applepieiris",
"followers_url": "https://api.github.c... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg... | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 7 | 2024-05-13T09:25:56 | 2024-06-02T00:29:42 | 2024-06-02T00:29:42 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I installed the Ollama in my linux server according to the official documents:
`curl -fsSL https://ollama.com/install.sh | sh`
Installation is ok and it returns:
`
>>> Downloading ollama...
######################################################################## 100.0%-#O#- # #
>>> Ins... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4395/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4395/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8654 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8654/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8654/comments | https://api.github.com/repos/ollama/ollama/issues/8654/events | https://github.com/ollama/ollama/issues/8654 | 2,817,986,286 | I_kwDOJ0Z1Ps6n9w7u | 8,654 | Available memory check should be disabled when mmap is in use | {
"login": "outis151",
"id": 11805613,
"node_id": "MDQ6VXNlcjExODA1NjEz",
"avatar_url": "https://avatars.githubusercontent.com/u/11805613?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/outis151",
"html_url": "https://github.com/outis151",
"followers_url": "https://api.github.com/users/out... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 1 | 2025-01-29T11:48:38 | 2025-01-29T13:07:03 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
With mmap enabled, a model does not need to fit in the system RAM. Therefore the associated check should be disabled in this case.
### OS
Linux
### GPU
_No response_
### CPU
Intel
### Ollama version
0.5.7 | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8654/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8654/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/3396 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3396/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3396/comments | https://api.github.com/repos/ollama/ollama/issues/3396/events | https://github.com/ollama/ollama/issues/3396 | 2,214,189,195 | I_kwDOJ0Z1Ps6D-diL | 3,396 | exec format error when Running Ollama Container on AMD64 Architecture | {
"login": "joshyorko",
"id": 54248591,
"node_id": "MDQ6VXNlcjU0MjQ4NTkx",
"avatar_url": "https://avatars.githubusercontent.com/u/54248591?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joshyorko",
"html_url": "https://github.com/joshyorko",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2024-03-28T21:26:11 | 2024-03-29T00:11:02 | 2024-03-29T00:11:01 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
When attempting to run the Ollama container, I continuously encounter an exec format error. The container fails to start properly and keeps trying to restart, logging the same error multiple times.
What did you expect to happen? I expected the Ollama container to start successfully without any ... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3396/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3396/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7553 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7553/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7553/comments | https://api.github.com/repos/ollama/ollama/issues/7553/events | https://github.com/ollama/ollama/issues/7553 | 2,640,556,716 | I_kwDOJ0Z1Ps6dY7Ks | 7,553 | Unable to load images from network fileshares on Windows | {
"login": "Antsiscool",
"id": 4112838,
"node_id": "MDQ6VXNlcjQxMTI4Mzg=",
"avatar_url": "https://avatars.githubusercontent.com/u/4112838?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Antsiscool",
"html_url": "https://github.com/Antsiscool",
"followers_url": "https://api.github.com/users... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg... | open | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 4 | 2024-11-07T10:24:54 | 2024-11-17T19:50:17 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Using Ollama on Windows via the terminal, if you ask a question and reference an image on a network fileshare, it will give a response about it not been able to see the photo. If you copy the image locally and then reference the local image, it has no problem with analysing the image.
Paths s... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7553/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7553/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/8002 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8002/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8002/comments | https://api.github.com/repos/ollama/ollama/issues/8002/events | https://github.com/ollama/ollama/pull/8002 | 2,725,706,888 | PR_kwDOJ0Z1Ps6Edzde | 8,002 | llama: preserve field order in user-defined JSON schemas | {
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers"... | [] | closed | false | null | [] | null | 3 | 2024-12-09T01:14:51 | 2024-12-11T22:07:32 | 2024-12-11T22:07:30 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8002",
"html_url": "https://github.com/ollama/ollama/pull/8002",
"diff_url": "https://github.com/ollama/ollama/pull/8002.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8002.patch",
"merged_at": "2024-12-11T22:07:30"
} | llama: preserve field order in user-defined JSON schemas
Previously we decoded and re-encoded JSON schemas during validation,
which served no purpose since json.RawMessage already validates JSON
syntax. Worse, the re-encoding lost field ordering from the original
schema, which affects inference quality during ste... | {
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers"... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8002/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8002/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5985 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5985/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5985/comments | https://api.github.com/repos/ollama/ollama/issues/5985/events | https://github.com/ollama/ollama/pull/5985 | 2,432,324,831 | PR_kwDOJ0Z1Ps52leCz | 5,985 | Use llama3.1 in tools example | {
"login": "rgbkrk",
"id": 836375,
"node_id": "MDQ6VXNlcjgzNjM3NQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/836375?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rgbkrk",
"html_url": "https://github.com/rgbkrk",
"followers_url": "https://api.github.com/users/rgbkrk/follow... | [] | closed | false | null | [] | null | 0 | 2024-07-26T14:07:11 | 2024-08-08T01:31:50 | 2024-08-07T21:20:51 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5985",
"html_url": "https://github.com/ollama/ollama/pull/5985",
"diff_url": "https://github.com/ollama/ollama/pull/5985.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5985.patch",
"merged_at": "2024-08-07T21:20:50"
} | Running this example with `mistral` produces the error "mistral does not support tools". What wasn't obvious to me until I made this PR was that my copy of mistral needed upgrading for tools (`ollama pull mistral`). Making the example be `llama3.1` will lead to more success for other long time ollama users. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5985/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5985/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6319 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6319/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6319/comments | https://api.github.com/repos/ollama/ollama/issues/6319/events | https://github.com/ollama/ollama/issues/6319 | 2,460,663,129 | I_kwDOJ0Z1Ps6Sqr1Z | 6,319 | Models RuGPT3, RuBERT | {
"login": "DewiarQR",
"id": 64423698,
"node_id": "MDQ6VXNlcjY0NDIzNjk4",
"avatar_url": "https://avatars.githubusercontent.com/u/64423698?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/DewiarQR",
"html_url": "https://github.com/DewiarQR",
"followers_url": "https://api.github.com/users/Dew... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | open | false | null | [] | null | 2 | 2024-08-12T10:51:26 | 2024-12-21T10:38:07 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | All models currently have pretty poor Russian language support. Is it possible to add RuGPT3, RuBERT models? | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6319/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6319/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/766 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/766/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/766/comments | https://api.github.com/repos/ollama/ollama/issues/766/events | https://github.com/ollama/ollama/issues/766 | 1,939,890,658 | I_kwDOJ0Z1Ps5zoGHi | 766 | Release mac and linux binaries alongside the desktop packages | {
"login": "Clivern",
"id": 1634427,
"node_id": "MDQ6VXNlcjE2MzQ0Mjc=",
"avatar_url": "https://avatars.githubusercontent.com/u/1634427?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Clivern",
"html_url": "https://github.com/Clivern",
"followers_url": "https://api.github.com/users/Clivern/... | [] | closed | false | null | [] | null | 5 | 2023-10-12T12:30:39 | 2023-11-11T22:56:39 | 2023-10-12T16:07:48 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Maybe ollama is intended to be a desktop app but I believe a lot are using it as an API service.
Honestly i couldn't get it to work as desktop app on Intel Mac but works as API service. Assuming the 500% spike in cpu usage is expected with each prompt and model pulling on a decent mac. I guess i need to give it a t... | {
"login": "Clivern",
"id": 1634427,
"node_id": "MDQ6VXNlcjE2MzQ0Mjc=",
"avatar_url": "https://avatars.githubusercontent.com/u/1634427?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Clivern",
"html_url": "https://github.com/Clivern",
"followers_url": "https://api.github.com/users/Clivern/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/766/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/766/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6293 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6293/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6293/comments | https://api.github.com/repos/ollama/ollama/issues/6293/events | https://github.com/ollama/ollama/issues/6293 | 2,458,713,929 | I_kwDOJ0Z1Ps6SjP9J | 6,293 | "The model you are attempting to pull requires a newer version of Ollama" when Ollama is built from the latest source | {
"login": "sammcj",
"id": 862951,
"node_id": "MDQ6VXNlcjg2Mjk1MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sammcj",
"html_url": "https://github.com/sammcj",
"followers_url": "https://api.github.com/users/sammcj/follow... | [
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info",
"name": "needs more info",
"color": "BA8041",
"default": false,
"description": "More information is needed to assist"
}
] | closed | false | null | [] | null | 8 | 2024-08-09T21:59:06 | 2024-08-15T02:32:17 | 2024-08-09T22:31:01 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
When trying to pull models from the official Ollama registry - if you're building Ollama from source it now seems to fail with an error that your Ollama version is too old.
```
ollama pull llama3.1:8b-instruct-q8_0
pulling manifest
Error: pull model manifest: 412:
The model you are atte... | {
"login": "sammcj",
"id": 862951,
"node_id": "MDQ6VXNlcjg2Mjk1MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sammcj",
"html_url": "https://github.com/sammcj",
"followers_url": "https://api.github.com/users/sammcj/follow... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6293/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6293/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/269 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/269/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/269/comments | https://api.github.com/repos/ollama/ollama/issues/269/events | https://github.com/ollama/ollama/issues/269 | 1,835,339,821 | I_kwDOJ0Z1Ps5tZRAt | 269 | Pressing enter during `ollama pull` causes newlines to be printed repeatedly | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2023-08-03T16:02:52 | 2023-12-24T21:39:30 | 2023-12-24T21:39:30 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | <img width="1531" alt="Screenshot 2023-08-03 at 11 59 21 AM" src="https://github.com/jmorganca/ollama/assets/251292/1e782cfa-75f2-4bc3-84da-567c685ef36c">
| {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/269/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/269/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1251 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1251/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1251/comments | https://api.github.com/repos/ollama/ollama/issues/1251/events | https://github.com/ollama/ollama/issues/1251 | 2,007,405,923 | I_kwDOJ0Z1Ps53ppVj | 1,251 | How can I disable automatic model offloading from GPU memory | {
"login": "anan-dad",
"id": 30836142,
"node_id": "MDQ6VXNlcjMwODM2MTQy",
"avatar_url": "https://avatars.githubusercontent.com/u/30836142?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/anan-dad",
"html_url": "https://github.com/anan-dad",
"followers_url": "https://api.github.com/users/ana... | [] | closed | false | null | [] | null | 1 | 2023-11-23T02:58:48 | 2023-11-23T03:05:50 | 2023-11-23T03:05:18 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | First of all, thank you for your great work with ollama!
I found that ollama will automatically offload models from GPU memory (very frequently, even after 2-minute inactive use).
But the loading process takes too much time, how can I forge ollama keep the model loading in GPU memory?
Thanks | {
"login": "anan-dad",
"id": 30836142,
"node_id": "MDQ6VXNlcjMwODM2MTQy",
"avatar_url": "https://avatars.githubusercontent.com/u/30836142?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/anan-dad",
"html_url": "https://github.com/anan-dad",
"followers_url": "https://api.github.com/users/ana... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1251/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1251/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2175 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2175/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2175/comments | https://api.github.com/repos/ollama/ollama/issues/2175/events | https://github.com/ollama/ollama/pull/2175 | 2,098,909,748 | PR_kwDOJ0Z1Ps5k_KNU | 2,175 | refactor tensor read | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2024-01-24T19:10:03 | 2024-01-25T17:22:43 | 2024-01-25T17:22:42 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2175",
"html_url": "https://github.com/ollama/ollama/pull/2175",
"diff_url": "https://github.com/ollama/ollama/pull/2175.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2175.patch",
"merged_at": "2024-01-25T17:22:42"
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2175/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2175/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8486 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8486/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8486/comments | https://api.github.com/repos/ollama/ollama/issues/8486/events | https://github.com/ollama/ollama/issues/8486 | 2,797,681,012 | I_kwDOJ0Z1Ps6mwTl0 | 8,486 | Add Tool Calling to the Generate Function | {
"login": "twright-0x1",
"id": 13889385,
"node_id": "MDQ6VXNlcjEzODg5Mzg1",
"avatar_url": "https://avatars.githubusercontent.com/u/13889385?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/twright-0x1",
"html_url": "https://github.com/twright-0x1",
"followers_url": "https://api.github.com/... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2025-01-19T15:21:05 | 2025-01-19T15:21:05 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | It appears from the API documentation and code examples available that tool calling is only possible with chat(). If this capability is feasible to add to generate() it would be much appreciated!
| null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8486/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8486/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/2908 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2908/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2908/comments | https://api.github.com/repos/ollama/ollama/issues/2908/events | https://github.com/ollama/ollama/issues/2908 | 2,166,302,209 | I_kwDOJ0Z1Ps6BHyYB | 2,908 | How to specify the installation directory | {
"login": "yuanjie-ai",
"id": 20265321,
"node_id": "MDQ6VXNlcjIwMjY1MzIx",
"avatar_url": "https://avatars.githubusercontent.com/u/20265321?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yuanjie-ai",
"html_url": "https://github.com/yuanjie-ai",
"followers_url": "https://api.github.com/use... | [] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 3 | 2024-03-04T08:56:56 | 2024-05-26T09:21:51 | 2024-03-21T11:36:12 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | How to specify the installation directory
| {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2908/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2908/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4135 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4135/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4135/comments | https://api.github.com/repos/ollama/ollama/issues/4135/events | https://github.com/ollama/ollama/pull/4135 | 2,278,248,414 | PR_kwDOJ0Z1Ps5ugQRD | 4,135 | Skip PhysX cudart library | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 1 | 2024-05-03T18:56:55 | 2024-05-06T20:34:03 | 2024-05-06T20:34:00 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4135",
"html_url": "https://github.com/ollama/ollama/pull/4135",
"diff_url": "https://github.com/ollama/ollama/pull/4135.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4135.patch",
"merged_at": "2024-05-06T20:34:00"
} | For some reason this library gives incorrect GPU information, so skip it
I'm not convinced yet this is the optimal fix, but queuing this up in case we get ready to cut a new release and haven't found a better solution yet.
Fixes #4008 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4135/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4135/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3089 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3089/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3089/comments | https://api.github.com/repos/ollama/ollama/issues/3089/events | https://github.com/ollama/ollama/issues/3089 | 2,182,976,031 | I_kwDOJ0Z1Ps6CHZIf | 3,089 | Error when requesting ollama api from another pc (windows) | {
"login": "insooneelife",
"id": 8437769,
"node_id": "MDQ6VXNlcjg0Mzc3Njk=",
"avatar_url": "https://avatars.githubusercontent.com/u/8437769?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/insooneelife",
"html_url": "https://github.com/insooneelife",
"followers_url": "https://api.github.com... | [
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] | closed | false | null | [] | null | 16 | 2024-03-13T02:06:05 | 2024-05-13T21:15:11 | 2024-03-15T13:36:22 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I plan to set up ollama on another PC and proceed with the work on the current PC.
However, when sending a request to ollama from a PC, I entered the IP address of the PC and sent it, but there is no reply.
Can you tell me what the problem is?
request url
http://localhost:11434/api/chat -> http://172.168.10.1... | {
"login": "insooneelife",
"id": 8437769,
"node_id": "MDQ6VXNlcjg0Mzc3Njk=",
"avatar_url": "https://avatars.githubusercontent.com/u/8437769?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/insooneelife",
"html_url": "https://github.com/insooneelife",
"followers_url": "https://api.github.com... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3089/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3089/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5452 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5452/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5452/comments | https://api.github.com/repos/ollama/ollama/issues/5452/events | https://github.com/ollama/ollama/issues/5452 | 2,387,458,446 | I_kwDOJ0Z1Ps6OTbmO | 5,452 | MARKDOWN!! | {
"login": "ashercn97",
"id": 131724380,
"node_id": "U_kgDOB9n0XA",
"avatar_url": "https://avatars.githubusercontent.com/u/131724380?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ashercn97",
"html_url": "https://github.com/ashercn97",
"followers_url": "https://api.github.com/users/asherc... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 5 | 2024-07-03T01:42:42 | 2024-10-17T17:32:45 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I think it would be so cool if this could render markdown in the terminal. It is kind of hard to read some of the stuff, and I would love if it could use something like glow or mdcat. Thanks! | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5452/reactions",
"total_count": 10,
"+1": 10,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5452/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/2168 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2168/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2168/comments | https://api.github.com/repos/ollama/ollama/issues/2168/events | https://github.com/ollama/ollama/issues/2168 | 2,097,815,632 | I_kwDOJ0Z1Ps59CiBQ | 2,168 | Issues Running Ollama Container Behind Proxy - No Error Logs Found | {
"login": "OM-EL",
"id": 36996895,
"node_id": "MDQ6VXNlcjM2OTk2ODk1",
"avatar_url": "https://avatars.githubusercontent.com/u/36996895?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/OM-EL",
"html_url": "https://github.com/OM-EL",
"followers_url": "https://api.github.com/users/OM-EL/follow... | [] | closed | false | null | [] | null | 10 | 2024-01-24T09:26:08 | 2024-10-17T07:06:36 | 2024-03-11T19:02:35 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I'm encountering issues while trying to run an Ollama container behind a proxy. Here are the steps I've taken and the issues I've faced:
1. **Creating an Image with Certificate**:
```
cat Dockerfile
FROM ollama/ollama
COPY my-ca.pem /usr/local/share/ca-certificates/my-ca.crt
RUN update-ca-c... | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2168/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2168/timeline | null | not_planned | false |
https://api.github.com/repos/ollama/ollama/issues/5988 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5988/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5988/comments | https://api.github.com/repos/ollama/ollama/issues/5988/events | https://github.com/ollama/ollama/issues/5988 | 2,432,490,642 | I_kwDOJ0Z1Ps6Q_NyS | 5,988 | GPU with 12GB VRAM couldn't load 8B model under WSL2 | {
"login": "hoangminh1109",
"id": 20716428,
"node_id": "MDQ6VXNlcjIwNzE2NDI4",
"avatar_url": "https://avatars.githubusercontent.com/u/20716428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoangminh1109",
"html_url": "https://github.com/hoangminh1109",
"followers_url": "https://api.githu... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg... | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 6 | 2024-07-26T15:37:39 | 2024-08-03T10:55:00 | 2024-08-03T10:55:00 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I'm unable to run any of the small model (8B model) on my RTX 3060 12GB.
Ollama is installed in WSL2 under Win10.

Server log uploaded [ollama_log_error.txt](https://github.com/user-attachments/files/1639... | {
"login": "hoangminh1109",
"id": 20716428,
"node_id": "MDQ6VXNlcjIwNzE2NDI4",
"avatar_url": "https://avatars.githubusercontent.com/u/20716428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoangminh1109",
"html_url": "https://github.com/hoangminh1109",
"followers_url": "https://api.githu... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5988/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5988/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3742 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3742/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3742/comments | https://api.github.com/repos/ollama/ollama/issues/3742/events | https://github.com/ollama/ollama/issues/3742 | 2,251,904,090 | I_kwDOJ0Z1Ps6GOVRa | 3,742 | Slow Performance with Llama2 on a Dual-GPU System - Seeking Advice | {
"login": "AkiMatsushita",
"id": 5045321,
"node_id": "MDQ6VXNlcjUwNDUzMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/5045321?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/AkiMatsushita",
"html_url": "https://github.com/AkiMatsushita",
"followers_url": "https://api.github.... | [
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info",
"name": "needs more info",
"color": "BA8041",
"default": false,
"description": "More information is needed to assist"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 5 | 2024-04-19T01:41:32 | 2024-04-22T22:41:38 | 2024-04-22T22:39:29 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hello ollama Community,
I'm encountering extremely slow performance while running ollama on my PC, specifically with models like Llama2 13B. The issue isn't just the slow output speed (around 1 token/min), but I'm also concerned that my GPUs might not be utilized properly. Below are my PC specs:
- CPU: Intel Core... | {
"login": "AkiMatsushita",
"id": 5045321,
"node_id": "MDQ6VXNlcjUwNDUzMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/5045321?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/AkiMatsushita",
"html_url": "https://github.com/AkiMatsushita",
"followers_url": "https://api.github.... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3742/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3742/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5912 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5912/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5912/comments | https://api.github.com/repos/ollama/ollama/issues/5912/events | https://github.com/ollama/ollama/pull/5912 | 2,427,590,112 | PR_kwDOJ0Z1Ps52V55j | 5,912 | Server tls 3203 | {
"login": "gabe-l-hart",
"id": 1254484,
"node_id": "MDQ6VXNlcjEyNTQ0ODQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/1254484?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gabe-l-hart",
"html_url": "https://github.com/gabe-l-hart",
"followers_url": "https://api.github.com/us... | [] | closed | false | null | [] | null | 1 | 2024-07-24T13:25:29 | 2024-10-03T16:04:14 | 2024-10-03T16:04:14 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5912",
"html_url": "https://github.com/ollama/ollama/pull/5912",
"diff_url": "https://github.com/ollama/ollama/pull/5912.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5912.patch",
"merged_at": null
} | **Disclaimer!**
This PR started as a small feature addition and resulted in some significant scope creep when I added the unit tests. I'm certainly open to trying to remove some of that refactoring for `ServerNonBlocking` if that's preferred, but figured it was worth presenting as-is to start the discussion.
## I... | {
"login": "gabe-l-hart",
"id": 1254484,
"node_id": "MDQ6VXNlcjEyNTQ0ODQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/1254484?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gabe-l-hart",
"html_url": "https://github.com/gabe-l-hart",
"followers_url": "https://api.github.com/us... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5912/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5912/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/236 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/236/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/236/comments | https://api.github.com/repos/ollama/ollama/issues/236/events | https://github.com/ollama/ollama/pull/236 | 1,826,971,486 | PR_kwDOJ0Z1Ps5WryWE | 236 | check os.Walk err | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-07-28T19:15:46 | 2023-07-28T21:14:22 | 2023-07-28T21:14:21 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/236",
"html_url": "https://github.com/ollama/ollama/pull/236",
"diff_url": "https://github.com/ollama/ollama/pull/236.diff",
"patch_url": "https://github.com/ollama/ollama/pull/236.patch",
"merged_at": "2023-07-28T21:14:21"
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/236/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/236/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1541 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1541/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1541/comments | https://api.github.com/repos/ollama/ollama/issues/1541/events | https://github.com/ollama/ollama/pull/1541 | 2,042,976,921 | PR_kwDOJ0Z1Ps5iEfYP | 1,541 | add API create/copy handlers | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | [] | closed | false | null | [] | null | 0 | 2023-12-15T06:20:11 | 2023-12-15T19:59:19 | 2023-12-15T19:59:18 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1541",
"html_url": "https://github.com/ollama/ollama/pull/1541",
"diff_url": "https://github.com/ollama/ollama/pull/1541.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1541.patch",
"merged_at": "2023-12-15T19:59:18"
} | This change adds a test for calling `POST /api/create` which creates a new model. | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1541/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1541/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1889 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1889/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1889/comments | https://api.github.com/repos/ollama/ollama/issues/1889/events | https://github.com/ollama/ollama/issues/1889 | 2,074,013,731 | I_kwDOJ0Z1Ps57nvAj | 1,889 | Phi2/dolphin-phi Disobedient on system prompt Biblical topics: | {
"login": "oliverbob",
"id": 23272429,
"node_id": "MDQ6VXNlcjIzMjcyNDI5",
"avatar_url": "https://avatars.githubusercontent.com/u/23272429?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/oliverbob",
"html_url": "https://github.com/oliverbob",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] | closed | false | null | [] | null | 4 | 2024-01-10T10:02:39 | 2024-05-10T00:16:11 | 2024-05-10T00:16:11 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Steps to reproduce:
Download a new Bible Dataset from [KJV Markdown .md](https://github.com/arleym/kjv-markdown/tree/master
)
```
#!/bin/bash
sudo rm joined.md
# Prepend content to the joined.md file
echo "FROM dolphin-phi" >> ./joined.md
echo "# set the temperature to 1 [higher is more creative, lower is... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1889/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1889/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/939 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/939/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/939/comments | https://api.github.com/repos/ollama/ollama/issues/939/events | https://github.com/ollama/ollama/issues/939 | 1,966,246,147 | I_kwDOJ0Z1Ps51MokD | 939 | Low memory systems with a lot of VRAM hit a memory issue | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2023-10-27T22:06:04 | 2024-01-10T15:08:21 | 2024-01-10T15:08:21 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | When creating a small instance with <4GB of RAM, `ollama` hits an error when loading the memory into VRAM | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/939/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/939/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/372 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/372/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/372/comments | https://api.github.com/repos/ollama/ollama/issues/372/events | https://github.com/ollama/ollama/pull/372 | 1,855,522,492 | PR_kwDOJ0Z1Ps5YL5qy | 372 | model and file type as strings | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-08-17T18:41:58 | 2023-08-17T22:10:59 | 2023-08-17T22:10:59 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/372",
"html_url": "https://github.com/ollama/ollama/pull/372",
"diff_url": "https://github.com/ollama/ollama/pull/372.diff",
"patch_url": "https://github.com/ollama/ollama/pull/372.patch",
"merged_at": "2023-08-17T22:10:59"
} | instead of representing model and file type as their native int values in manifest config, represent them as user-friendly strings | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/372/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/372/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/270 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/270/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/270/comments | https://api.github.com/repos/ollama/ollama/issues/270/events | https://github.com/ollama/ollama/pull/270 | 1,835,562,166 | PR_kwDOJ0Z1Ps5XIuTO | 270 | update llama.cpp | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-08-03T18:50:35 | 2023-08-03T19:09:02 | 2023-08-03T19:09:01 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/270",
"html_url": "https://github.com/ollama/ollama/pull/270",
"diff_url": "https://github.com/ollama/ollama/pull/270.diff",
"patch_url": "https://github.com/ollama/ollama/pull/270.patch",
"merged_at": "2023-08-03T19:09:01"
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/270/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/ollama/ollama/issues/270/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1759 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1759/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1759/comments | https://api.github.com/repos/ollama/ollama/issues/1759/events | https://github.com/ollama/ollama/issues/1759 | 2,062,122,844 | I_kwDOJ0Z1Ps566X9c | 1,759 | Please add TinyGPT-V model support | {
"login": "yangyang0507",
"id": 5666807,
"node_id": "MDQ6VXNlcjU2NjY4MDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/5666807?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yangyang0507",
"html_url": "https://github.com/yangyang0507",
"followers_url": "https://api.github.com... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | open | false | null | [] | null | 0 | 2024-01-02T09:03:45 | 2024-01-02T11:34:34 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | TinyGPT-V: Efficient Multimodal Large Language Model via Small Backbones
Github: https://github.com/DLYuanGod/TinyGPT-V
HuggingFace: https://huggingface.co/Tyrannosaurus/TinyGPT-V
It stands out because it only requires a 24G GPU for training, and just an 8G GPU or CPU for inference. TinyGPT-V is based on Phi-2, ... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1759/reactions",
"total_count": 4,
"+1": 4,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1759/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/483 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/483/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/483/comments | https://api.github.com/repos/ollama/ollama/issues/483/events | https://github.com/ollama/ollama/issues/483 | 1,885,300,022 | I_kwDOJ0Z1Ps5wX2U2 | 483 | No response from model with giant request | {
"login": "FairyTail2000",
"id": 22645621,
"node_id": "MDQ6VXNlcjIyNjQ1NjIx",
"avatar_url": "https://avatars.githubusercontent.com/u/22645621?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/FairyTail2000",
"html_url": "https://github.com/FairyTail2000",
"followers_url": "https://api.githu... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2023-09-07T07:41:24 | 2023-12-04T19:24:58 | 2023-12-04T19:24:57 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Using my own personal frontend with the model codellama:34b-code-q4_0 I send a giant block of code ~10kB. The model then runs for 5 - 6 minutes but only a single token comes out of the model.
This is the http response:
>{"model":"codellama:34b-code-q4_0","created_at":"2023-09-07T07:34:32.574995065Z","response":"\n"... | {
"login": "technovangelist",
"id": 633681,
"node_id": "MDQ6VXNlcjYzMzY4MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/technovangelist",
"html_url": "https://github.com/technovangelist",
"followers_url": "https://api.git... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/483/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/483/timeline | null | completed | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.