url stringlengths 51 54 | repository_url stringclasses 1
value | labels_url stringlengths 65 68 | comments_url stringlengths 60 63 | events_url stringlengths 58 61 | html_url stringlengths 39 44 | id int64 1.78B 2.82B | node_id stringlengths 18 19 | number int64 1 8.69k | title stringlengths 1 382 | user dict | labels listlengths 0 5 | state stringclasses 2
values | locked bool 1
class | assignee dict | assignees listlengths 0 2 | milestone null | comments int64 0 323 | created_at timestamp[s] | updated_at timestamp[s] | closed_at timestamp[s] | author_association stringclasses 4
values | sub_issues_summary dict | active_lock_reason null | draft bool 2
classes | pull_request dict | body stringlengths 2 118k ⌀ | closed_by dict | reactions dict | timeline_url stringlengths 60 63 | performed_via_github_app null | state_reason stringclasses 4
values | is_pull_request bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/ollama/ollama/issues/1498 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1498/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1498/comments | https://api.github.com/repos/ollama/ollama/issues/1498/events | https://github.com/ollama/ollama/issues/1498 | 2,039,154,511 | I_kwDOJ0Z1Ps55iwdP | 1,498 | /set system no longer works | {
"login": "iplayfast",
"id": 751306,
"node_id": "MDQ6VXNlcjc1MTMwNg==",
"avatar_url": "https://avatars.githubusercontent.com/u/751306?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/iplayfast",
"html_url": "https://github.com/iplayfast",
"followers_url": "https://api.github.com/users/ipla... | [] | closed | false | null | [] | null | 2 | 2023-12-13T07:55:30 | 2023-12-15T18:56:26 | 2023-12-15T18:56:26 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I created a model
```
/show modelfile
# Modelfile generated by "ollama show"
# To build a new Modelfile based on this one, replace the FROM line with:
# FROM Sally:latest
FROM wizard-vicuna-uncensored:7b
TEMPLATE """{{ .System }}
USER: {{ .Prompt }}
ASSISTANT:
"""
SYSTEM """ You are a female named Sally,... | {
"login": "iplayfast",
"id": 751306,
"node_id": "MDQ6VXNlcjc1MTMwNg==",
"avatar_url": "https://avatars.githubusercontent.com/u/751306?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/iplayfast",
"html_url": "https://github.com/iplayfast",
"followers_url": "https://api.github.com/users/ipla... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1498/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1498/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7664 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7664/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7664/comments | https://api.github.com/repos/ollama/ollama/issues/7664/events | https://github.com/ollama/ollama/issues/7664 | 2,657,985,347 | I_kwDOJ0Z1Ps6ebaND | 7,664 | llama3.2-vision:90b is very slow at generating text on 4x V100 64G GPUs | {
"login": "YUHSINCHENG1230",
"id": 160841042,
"node_id": "U_kgDOCZY9Ug",
"avatar_url": "https://avatars.githubusercontent.com/u/160841042?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/YUHSINCHENG1230",
"html_url": "https://github.com/YUHSINCHENG1230",
"followers_url": "https://api.githu... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg... | closed | false | null | [] | null | 6 | 2024-11-14T08:15:45 | 2024-12-13T11:44:03 | 2024-12-13T11:44:03 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
when I use llama3.2-vision:90b ,the model always response so slow. how can I do?
And the GPU is not fully used. CPU used very high.
it's run on V100*4 = 64G GPU
Is anyone can help?
### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.4.1 | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7664/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7664/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8437 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8437/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8437/comments | https://api.github.com/repos/ollama/ollama/issues/8437/events | https://github.com/ollama/ollama/issues/8437 | 2,789,116,076 | I_kwDOJ0Z1Ps6mPois | 8,437 | run phi4 report timed out waiting for llama runner to start | {
"login": "Jeffhop",
"id": 34859341,
"node_id": "MDQ6VXNlcjM0ODU5MzQx",
"avatar_url": "https://avatars.githubusercontent.com/u/34859341?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Jeffhop",
"html_url": "https://github.com/Jeffhop",
"followers_url": "https://api.github.com/users/Jeffho... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 4 | 2025-01-15T08:14:14 | 2025-01-15T09:09:58 | 2025-01-15T09:09:57 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I used the docker ollama:latest image, entered the container, and got an error when running ollama run phi4 inside the container.

 | {
"login": "asmith26",
"id": 6988036,
"node_id": "MDQ6VXNlcjY5ODgwMzY=",
"avatar_url": "https://avatars.githubusercontent.com/u/6988036?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/asmith26",
"html_url": "https://github.com/asmith26",
"followers_url": "https://api.github.com/users/asmit... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 3 | 2024-05-02T20:07:48 | 2024-05-03T16:08:52 | 2024-05-02T21:07:21 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Hi,
I'm running the example from https://ollama.com/blog/embedding-models, and when I look at the value of `data = results['documents'][0][0]` every time I get:
```
Llamas are vegetarians and have very efficient digestive systems
```
The blog implies the value it gets for `data` (b... | {
"login": "asmith26",
"id": 6988036,
"node_id": "MDQ6VXNlcjY5ODgwMzY=",
"avatar_url": "https://avatars.githubusercontent.com/u/6988036?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/asmith26",
"html_url": "https://github.com/asmith26",
"followers_url": "https://api.github.com/users/asmit... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4104/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4104/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7532 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7532/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7532/comments | https://api.github.com/repos/ollama/ollama/issues/7532/events | https://github.com/ollama/ollama/pull/7532 | 2,639,246,118 | PR_kwDOJ0Z1Ps6BHAQd | 7,532 | Debug logging for nvcuda init | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 1 | 2024-11-06T21:15:30 | 2024-11-07T22:25:56 | 2024-11-07T22:25:54 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7532",
"html_url": "https://github.com/ollama/ollama/pull/7532",
"diff_url": "https://github.com/ollama/ollama/pull/7532.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7532.patch",
"merged_at": "2024-11-07T22:25:54"
} | Some users are reporting crashes during nvcuda.dll initialization on windows. This should help narrow down where things are going bad.
This will help isolate what the problem is with #7524 and #5625 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7532/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7532/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6361 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6361/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6361/comments | https://api.github.com/repos/ollama/ollama/issues/6361/events | https://github.com/ollama/ollama/pull/6361 | 2,466,050,712 | PR_kwDOJ0Z1Ps54XuGL | 6,361 | Add ollama-php library to README.MD | {
"login": "ArdaGnsrn",
"id": 52893305,
"node_id": "MDQ6VXNlcjUyODkzMzA1",
"avatar_url": "https://avatars.githubusercontent.com/u/52893305?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ArdaGnsrn",
"html_url": "https://github.com/ArdaGnsrn",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 3 | 2024-08-14T14:39:52 | 2024-09-05T05:23:03 | 2024-09-05T05:01:14 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/6361",
"html_url": "https://github.com/ollama/ollama/pull/6361",
"diff_url": "https://github.com/ollama/ollama/pull/6361.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6361.patch",
"merged_at": "2024-09-05T05:01:14"
} | ollama-php is a library that provides php support for ollama. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6361/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6361/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6128 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6128/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6128/comments | https://api.github.com/repos/ollama/ollama/issues/6128/events | https://github.com/ollama/ollama/pull/6128 | 2,443,546,265 | PR_kwDOJ0Z1Ps53Lh46 | 6,128 | enable gofmt/gofumpt/goimports/tenv | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2024-08-01T21:52:39 | 2024-08-02T21:58:42 | 2024-08-02T21:58:40 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/6128",
"html_url": "https://github.com/ollama/ollama/pull/6128",
"diff_url": "https://github.com/ollama/ollama/pull/6128.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6128.patch",
"merged_at": "2024-08-02T21:58:40"
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6128/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6128/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2485 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2485/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2485/comments | https://api.github.com/repos/ollama/ollama/issues/2485/events | https://github.com/ollama/ollama/pull/2485 | 2,133,414,047 | PR_kwDOJ0Z1Ps5mz1fR | 2,485 | Revert "Revert "bump submodule to `6c00a06` (#2479)"" | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 0 | 2024-02-14T01:52:39 | 2024-02-14T02:18:42 | 2024-02-14T02:18:41 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2485",
"html_url": "https://github.com/ollama/ollama/pull/2485",
"diff_url": "https://github.com/ollama/ollama/pull/2485.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2485.patch",
"merged_at": "2024-02-14T02:18:41"
} | This reverts commit 6920964b87971c8201097130bfdedbf56aaa13a7. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2485/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2485/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/664 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/664/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/664/comments | https://api.github.com/repos/ollama/ollama/issues/664/events | https://github.com/ollama/ollama/pull/664 | 1,920,766,718 | PR_kwDOJ0Z1Ps5bnRSq | 664 | add some missing code directives in docs | {
"login": "jimexist",
"id": 622789,
"node_id": "MDQ6VXNlcjYyMjc4OQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/622789?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jimexist",
"html_url": "https://github.com/jimexist",
"followers_url": "https://api.github.com/users/jimexis... | [] | closed | false | null | [] | null | 1 | 2023-10-01T13:24:48 | 2023-10-01T18:51:01 | 2023-10-01T18:51:01 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/664",
"html_url": "https://github.com/ollama/ollama/pull/664",
"diff_url": "https://github.com/ollama/ollama/pull/664.diff",
"patch_url": "https://github.com/ollama/ollama/pull/664.patch",
"merged_at": "2023-10-01T18:51:01"
} | - mostly just adding bash to shell scripts | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/664/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/664/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6520 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6520/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6520/comments | https://api.github.com/repos/ollama/ollama/issues/6520/events | https://github.com/ollama/ollama/issues/6520 | 2,487,486,714 | I_kwDOJ0Z1Ps6URAj6 | 6,520 | environmental variable not passed to service | {
"login": "N4S4",
"id": 33936751,
"node_id": "MDQ6VXNlcjMzOTM2NzUx",
"avatar_url": "https://avatars.githubusercontent.com/u/33936751?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/N4S4",
"html_url": "https://github.com/N4S4",
"followers_url": "https://api.github.com/users/N4S4/followers"... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 4 | 2024-08-26T18:22:47 | 2024-08-26T20:47:34 | 2024-08-26T20:47:34 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
when trying to set Environmental variable using systemctrl edit ollama.service, restart daemon and restarting ollama.service, the override is not passed to the service. trying to override HSA_OVERRIDE_GFX_VERSION to make igpu 780M ryzen 9 7940hs
### OS
Linux
### GPU
AMD
### CPU
AMD
### Ol... | {
"login": "N4S4",
"id": 33936751,
"node_id": "MDQ6VXNlcjMzOTM2NzUx",
"avatar_url": "https://avatars.githubusercontent.com/u/33936751?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/N4S4",
"html_url": "https://github.com/N4S4",
"followers_url": "https://api.github.com/users/N4S4/followers"... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6520/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6520/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2041 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2041/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2041/comments | https://api.github.com/repos/ollama/ollama/issues/2041/events | https://github.com/ollama/ollama/issues/2041 | 2,087,433,309 | I_kwDOJ0Z1Ps58a7Rd | 2,041 | Show or check the model of equipment minimum requirements | {
"login": "ChingWeiChan",
"id": 14084937,
"node_id": "MDQ6VXNlcjE0MDg0OTM3",
"avatar_url": "https://avatars.githubusercontent.com/u/14084937?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ChingWeiChan",
"html_url": "https://github.com/ChingWeiChan",
"followers_url": "https://api.github.c... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2024-01-18T03:52:17 | 2024-03-11T18:23:15 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | When I want to try a model, I need to check the minimum GPU vram of model.
* There are many parameters version like 8b, 13b 70b.
* If it's GGUF format, there will have many bits version like q3_K_S, q4_K_L.
* For example, I have 4070ti in my computer, Running llama2 70b model will get a poor performance because la... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2041/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2041/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/6103 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6103/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6103/comments | https://api.github.com/repos/ollama/ollama/issues/6103/events | https://github.com/ollama/ollama/issues/6103 | 2,440,845,387 | I_kwDOJ0Z1Ps6RfFhL | 6,103 | show --modelfile doesn't properly quote MESSAGE statements | {
"login": "Maltz42",
"id": 20978744,
"node_id": "MDQ6VXNlcjIwOTc4NzQ0",
"avatar_url": "https://avatars.githubusercontent.com/u/20978744?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Maltz42",
"html_url": "https://github.com/Maltz42",
"followers_url": "https://api.github.com/users/Maltz4... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 4 | 2024-07-31T20:01:18 | 2024-08-05T01:25:00 | 2024-08-01T00:05:45 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Exporting a model file ("ollama show --modelfile MODEL") doesn't properly quote MESSAGE statements. As a result, the model file generated cannot be re-imported ("ollama create MODEL -f <modelfile>) without modification. For example:
What is currently generated:
MESSAGE user write a humor... | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6103/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6103/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5428 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5428/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5428/comments | https://api.github.com/repos/ollama/ollama/issues/5428/events | https://github.com/ollama/ollama/issues/5428 | 2,385,409,151 | I_kwDOJ0Z1Ps6OLnR_ | 5,428 | bump llama.cpp for gemma2 fixes | {
"login": "ki-manufaktur",
"id": 160486566,
"node_id": "U_kgDOCZDUpg",
"avatar_url": "https://avatars.githubusercontent.com/u/160486566?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ki-manufaktur",
"html_url": "https://github.com/ki-manufaktur",
"followers_url": "https://api.github.com/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 7 | 2024-07-02T07:02:50 | 2024-07-10T11:47:00 | 2024-07-10T11:47:00 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Hi,
there are at least 3 merged PRs in lamma.cpp that fix performance issues for gemma2 models
https://github.com/ggerganov/llama.cpp/pull/8197
https://github.com/ggerganov/llama.cpp/pull/8227
https://github.com/ggerganov/llama.cpp/pull/8244
### OS
_No response_
### GPU
_No... | {
"login": "ki-manufaktur",
"id": 160486566,
"node_id": "U_kgDOCZDUpg",
"avatar_url": "https://avatars.githubusercontent.com/u/160486566?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ki-manufaktur",
"html_url": "https://github.com/ki-manufaktur",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5428/reactions",
"total_count": 11,
"+1": 11,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5428/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3135 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3135/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3135/comments | https://api.github.com/repos/ollama/ollama/issues/3135/events | https://github.com/ollama/ollama/issues/3135 | 2,185,681,739 | I_kwDOJ0Z1Ps6CRttL | 3,135 | LLava returns non-unique responses | {
"login": "jeffreylm",
"id": 13600081,
"node_id": "MDQ6VXNlcjEzNjAwMDgx",
"avatar_url": "https://avatars.githubusercontent.com/u/13600081?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jeffreylm",
"html_url": "https://github.com/jeffreylm",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 1 | 2024-03-14T07:58:03 | 2024-03-15T03:43:09 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Dear Ollama,
I am using LLava with WebUI, and have already spoken to WebUi github about this, and as here is no errors in the docker error logs, they advised me to contact the Ollama team's Github.
When I type : Describe Image
LLava give me a cloned answer no matter the images I give it. The first image/response... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3135/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3135/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/4629 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4629/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4629/comments | https://api.github.com/repos/ollama/ollama/issues/4629/events | https://github.com/ollama/ollama/issues/4629 | 2,316,801,691 | I_kwDOJ0Z1Ps6KF5ab | 4,629 | granite-code:20b-instruct-q8_0 error loading model vocabulary: unknown pre-tokenizer type: 'refact' | {
"login": "ekolawole",
"id": 79321648,
"node_id": "MDQ6VXNlcjc5MzIxNjQ4",
"avatar_url": "https://avatars.githubusercontent.com/u/79321648?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ekolawole",
"html_url": "https://github.com/ekolawole",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api... | null | 4 | 2024-05-25T07:35:56 | 2024-05-30T21:37:13 | 2024-05-30T21:37:12 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
llama_model_loader: - type f32: 419 tensors
llama_model_loader: - type q8_0: 209 tensors
llama_model_load: error loading model: error loading model vocabulary: unknown pre-tokenizer type: 'refact'
llama_load_model_from_file: exception loading model
libc++abi: terminating due to uncaught e... | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4629/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4629/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2303 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2303/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2303/comments | https://api.github.com/repos/ollama/ollama/issues/2303/events | https://github.com/ollama/ollama/issues/2303 | 2,111,468,658 | I_kwDOJ0Z1Ps592nRy | 2,303 | Large number of images causes hanging/error | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 0 | 2024-02-01T03:49:17 | 2024-02-20T04:04:36 | 2024-02-20T04:04:36 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | If submitting too many images via the `images` API parameter in `/api/generate` or` /api/chat`
```
[1706759134] slot 0 - loaded image
[1706759135] slot 0 - loaded image
[1706759135] slot 0 - loaded image
[1706759135] slot 0 - loaded image
[1706759135] slot 0 - loaded image
[1706759135] slot 0 is processing [ta... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2303/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/ollama/ollama/issues/2303/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4677 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4677/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4677/comments | https://api.github.com/repos/ollama/ollama/issues/4677/events | https://github.com/ollama/ollama/issues/4677 | 2,320,518,922 | I_kwDOJ0Z1Ps6KUE8K | 4,677 | Can't download Complete in 0.1.39 rc | {
"login": "Leroy-X",
"id": 13515498,
"node_id": "MDQ6VXNlcjEzNTE1NDk4",
"avatar_url": "https://avatars.githubusercontent.com/u/13515498?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Leroy-X",
"html_url": "https://github.com/Leroy-X",
"followers_url": "https://api.github.com/users/Leroy-... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2024-05-28T09:01:47 | 2024-05-28T20:22:39 | 2024-05-28T20:22:38 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
In the latest 0.1.39 version, I am almost unable to complete the model download. It often breaks off and prompts "Error: max retries exceeded: EOF". I have been trying for a whole day but still haven't finished downloading a 4Gb model.
I hope this issue can be fixed soon. Thanks!
### OS
W... | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4677/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4677/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8327 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8327/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8327/comments | https://api.github.com/repos/ollama/ollama/issues/8327/events | https://github.com/ollama/ollama/issues/8327 | 2,771,710,951 | I_kwDOJ0Z1Ps6lNPPn | 8,327 | I/O error on POST request for "http://localhost:11434/v1/chat/completions\ | {
"login": "OnceCrazyer",
"id": 16172911,
"node_id": "MDQ6VXNlcjE2MTcyOTEx",
"avatar_url": "https://avatars.githubusercontent.com/u/16172911?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/OnceCrazyer",
"html_url": "https://github.com/OnceCrazyer",
"followers_url": "https://api.github.com/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q... | open | false | null | [] | null | 6 | 2025-01-07T01:21:50 | 2025-01-15T07:57:23 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
error infomation:
{
"timestamp": "2025-01-07T01:02:47.506+00:00",
"status": 500,
"error": "Internal Server Error",
"trace": "org.springframework.web.client.ResourceAccessException: I/O error on POST request for \"http://localhost:11434/v1/chat/completions": Read timed out\",
"m... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8327/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8327/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/536 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/536/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/536/comments | https://api.github.com/repos/ollama/ollama/issues/536/events | https://github.com/ollama/ollama/pull/536 | 1,899,123,749 | PR_kwDOJ0Z1Ps5aeloe | 536 | explicitly follow upload redirects | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-09-15T21:58:10 | 2023-09-20T18:27:04 | 2023-09-20T18:27:03 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/536",
"html_url": "https://github.com/ollama/ollama/pull/536",
"diff_url": "https://github.com/ollama/ollama/pull/536.diff",
"patch_url": "https://github.com/ollama/ollama/pull/536.patch",
"merged_at": "2023-09-20T18:27:03"
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/536/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/536/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8605 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8605/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8605/comments | https://api.github.com/repos/ollama/ollama/issues/8605/events | https://github.com/ollama/ollama/issues/8605 | 2,812,486,439 | I_kwDOJ0Z1Ps6noyMn | 8,605 | Error fetching ANY model locally | {
"login": "devroopsaha744",
"id": 130696540,
"node_id": "U_kgDOB8pFXA",
"avatar_url": "https://avatars.githubusercontent.com/u/130696540?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/devroopsaha744",
"html_url": "https://github.com/devroopsaha744",
"followers_url": "https://api.github.c... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677370291,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCVsw... | open | false | null | [] | null | 4 | 2025-01-27T09:25:12 | 2025-01-28T17:14:43 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
This is the Error message that I am getting:
pulling manifest
pulling 6e9f90f02bb3... 0% ▕ ▏ 0 B/9.0 GB
Error: max retries exceeded: Get "https://dd20bb891979d25aebc8bec07b2b3bbc.r2.cloudflarestorage.com/ollama/docker/registry/v2/blobs... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8605/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8605/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/5286 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5286/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5286/comments | https://api.github.com/repos/ollama/ollama/issues/5286/events | https://github.com/ollama/ollama/issues/5286 | 2,373,833,987 | I_kwDOJ0Z1Ps6NfdUD | 5,286 | [feature request] Easy clean-up of large `ollama` files | {
"login": "hamirmahal",
"id": 43425812,
"node_id": "MDQ6VXNlcjQzNDI1ODEy",
"avatar_url": "https://avatars.githubusercontent.com/u/43425812?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hamirmahal",
"html_url": "https://github.com/hamirmahal",
"followers_url": "https://api.github.com/use... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 4 | 2024-06-25T23:14:16 | 2024-06-26T08:43:16 | 2024-06-26T08:35:11 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | # Problem
It isn't immediately clear how to delete large models from `ollama`.

```
/usr/share/ollama/.ollama/models/blobs $ du -sh *
4.0K sha256:2e0493f67d0c8c9c68a8aeacdf6a38a2151cb3c4c1d42accf296e19810527988
4... | {
"login": "hamirmahal",
"id": 43425812,
"node_id": "MDQ6VXNlcjQzNDI1ODEy",
"avatar_url": "https://avatars.githubusercontent.com/u/43425812?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hamirmahal",
"html_url": "https://github.com/hamirmahal",
"followers_url": "https://api.github.com/use... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5286/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5286/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1051 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1051/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1051/comments | https://api.github.com/repos/ollama/ollama/issues/1051/events | https://github.com/ollama/ollama/pull/1051 | 1,984,681,273 | PR_kwDOJ0Z1Ps5e_A3_ | 1,051 | JSON mode: add `"format": "json"` as an api parameter | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 1 | 2023-11-09T01:58:51 | 2023-11-10T16:25:33 | 2023-11-10T00:44:02 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1051",
"html_url": "https://github.com/ollama/ollama/pull/1051",
"diff_url": "https://github.com/ollama/ollama/pull/1051.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1051.patch",
"merged_at": "2023-11-10T00:44:02"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1051/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1051/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5446 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5446/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5446/comments | https://api.github.com/repos/ollama/ollama/issues/5446/events | https://github.com/ollama/ollama/pull/5446 | 2,387,237,769 | PR_kwDOJ0Z1Ps50QKEe | 5,446 | update faq | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2024-07-02T22:02:41 | 2024-08-23T21:06:00 | 2024-08-23T21:05:59 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5446",
"html_url": "https://github.com/ollama/ollama/pull/5446",
"diff_url": "https://github.com/ollama/ollama/pull/5446.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5446.patch",
"merged_at": "2024-08-23T21:05:59"
} | clarify https proxy faq
related #4834
related #1337 | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5446/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5446/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5521 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5521/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5521/comments | https://api.github.com/repos/ollama/ollama/issues/5521/events | https://github.com/ollama/ollama/pull/5521 | 2,393,782,707 | PR_kwDOJ0Z1Ps50mW4B | 5,521 | gpu: report system free memory instead of 0 | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 0 | 2024-07-06T23:29:03 | 2024-07-06T23:35:06 | 2024-07-06T23:35:05 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5521",
"html_url": "https://github.com/ollama/ollama/pull/5521",
"diff_url": "https://github.com/ollama/ollama/pull/5521.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5521.patch",
"merged_at": "2024-07-06T23:35:04"
} | After https://github.com/ollama/ollama/pull/5469, no model will run on CPU on macOS because we always report system available memory as 0. This PR adds system/cpu memory detection to avoid erroring. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5521/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5521/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/305 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/305/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/305/comments | https://api.github.com/repos/ollama/ollama/issues/305/events | https://github.com/ollama/ollama/issues/305 | 1,840,296,346 | I_kwDOJ0Z1Ps5tsLGa | 305 | OpenAI API compatibility | {
"login": "handrew",
"id": 1425219,
"node_id": "MDQ6VXNlcjE0MjUyMTk=",
"avatar_url": "https://avatars.githubusercontent.com/u/1425219?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/handrew",
"html_url": "https://github.com/handrew",
"followers_url": "https://api.github.com/users/handrew/... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.g... | null | 61 | 2023-08-07T22:36:36 | 2024-10-13T12:17:48 | 2024-02-07T22:24:30 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Any chance you would consider mirroring OpenAI's API specs and output? e.g., /completions and /chat/completions. That way, it could be a drop-in replacement for the Python openai package by changing out the url. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/305/reactions",
"total_count": 169,
"+1": 151,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 11,
"rocket": 0,
"eyes": 7
} | https://api.github.com/repos/ollama/ollama/issues/305/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8421 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8421/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8421/comments | https://api.github.com/repos/ollama/ollama/issues/8421/events | https://github.com/ollama/ollama/issues/8421 | 2,787,133,701 | I_kwDOJ0Z1Ps6mIEkF | 8,421 | `tool_choice` argument is being ignored | {
"login": "pminervini",
"id": 227357,
"node_id": "MDQ6VXNlcjIyNzM1Nw==",
"avatar_url": "https://avatars.githubusercontent.com/u/227357?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pminervini",
"html_url": "https://github.com/pminervini",
"followers_url": "https://api.github.com/users/p... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2025-01-14T13:30:16 | 2025-01-15T08:01:34 | 2025-01-15T08:01:33 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Consider the following snippet:
```python
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
from openai import OpenAI
client_kwargs = {
'base_url': "http://localhost:11434/v1"
}
client = OpenAI(**client_kwargs)
# Messages from the log
messages = [{
'role': 'system',
'content': 'Yo... | {
"login": "pminervini",
"id": 227357,
"node_id": "MDQ6VXNlcjIyNzM1Nw==",
"avatar_url": "https://avatars.githubusercontent.com/u/227357?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pminervini",
"html_url": "https://github.com/pminervini",
"followers_url": "https://api.github.com/users/p... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8421/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8421/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/455 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/455/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/455/comments | https://api.github.com/repos/ollama/ollama/issues/455/events | https://github.com/ollama/ollama/pull/455 | 1,878,030,462 | PR_kwDOJ0Z1Ps5ZXyMz | 455 | generate in build script | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 1 | 2023-09-01T20:50:01 | 2023-09-08T15:13:13 | 2023-09-05T14:45:05 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/455",
"html_url": "https://github.com/ollama/ollama/pull/455",
"diff_url": "https://github.com/ollama/ollama/pull/455.diff",
"patch_url": "https://github.com/ollama/ollama/pull/455.patch",
"merged_at": null
} | Add llama.cpp exe generation to the build script. Still need to figure out a way to pack both amd64 and arm64 to make this binary truly universal. | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/455/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/455/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6967 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6967/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6967/comments | https://api.github.com/repos/ollama/ollama/issues/6967/events | https://github.com/ollama/ollama/pull/6967 | 2,549,137,037 | PR_kwDOJ0Z1Ps58uBTK | 6,967 | windows: Support alt install paths, fit and finish | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 3 | 2024-09-25T22:59:26 | 2024-11-02T09:16:53 | 2024-10-30T16:24:31 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/6967",
"html_url": "https://github.com/ollama/ollama/pull/6967",
"diff_url": "https://github.com/ollama/ollama/pull/6967.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6967.patch",
"merged_at": "2024-10-30T16:24:31"
} | - Document alternate install location instructions for binaries and model, and fix path resolution relative to executable
- Pop up progress UI for upgrades (automatic, with cancel button).
- Expose non-default port in menu to disambiguate multiple instances.
Fixes #6725
Fixes #2776
Fixes #6609 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6967/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6967/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7538 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7538/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7538/comments | https://api.github.com/repos/ollama/ollama/issues/7538/events | https://github.com/ollama/ollama/issues/7538 | 2,639,626,515 | I_kwDOJ0Z1Ps6dVYET | 7,538 | Qwen2.5-Coder-7B-Instruct 2024/11/06 update | {
"login": "vYLQs6",
"id": 143073604,
"node_id": "U_kgDOCIchRA",
"avatar_url": "https://avatars.githubusercontent.com/u/143073604?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/vYLQs6",
"html_url": "https://github.com/vYLQs6",
"followers_url": "https://api.github.com/users/vYLQs6/follower... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 2 | 2024-11-07T01:45:59 | 2024-11-13T21:37:58 | 2024-11-13T21:37:58 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | https://huggingface.co/Qwen/Qwen2.5-Coder-7B-Instruct/commit/9092a8ae57da39f15b76b309b4f71ff11b6ef01a
They just updated the weights of Qwen2.5-Coder-7B-Instruct without any announcement, so I don't know what exactly what the update is, but it's an update
 | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7538/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7538/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2308 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2308/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2308/comments | https://api.github.com/repos/ollama/ollama/issues/2308/events | https://github.com/ollama/ollama/issues/2308 | 2,112,075,161 | I_kwDOJ0Z1Ps5947WZ | 2,308 | Expose Ollama Service to use it in Chrome Browser Extension | {
"login": "DevChrisRoth",
"id": 58110317,
"node_id": "MDQ6VXNlcjU4MTEwMzE3",
"avatar_url": "https://avatars.githubusercontent.com/u/58110317?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/DevChrisRoth",
"html_url": "https://github.com/DevChrisRoth",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | 34 | 2024-02-01T10:02:42 | 2024-10-20T09:30:02 | 2024-02-02T11:36:48 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hey guys,
can you show us, how we can actually use the provided Ollama API's in an chrome extension.
I tried it before, but I get an 403 - Forbidden error. I already looked in the ollama documentation but did not found anything useful.
Hopefully someone can show us how to use it properly. I opened a [stackoverflow... | {
"login": "DevChrisRoth",
"id": 58110317,
"node_id": "MDQ6VXNlcjU4MTEwMzE3",
"avatar_url": "https://avatars.githubusercontent.com/u/58110317?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/DevChrisRoth",
"html_url": "https://github.com/DevChrisRoth",
"followers_url": "https://api.github.c... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2308/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2308/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2857 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2857/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2857/comments | https://api.github.com/repos/ollama/ollama/issues/2857/events | https://github.com/ollama/ollama/issues/2857 | 2,162,750,832 | I_kwDOJ0Z1Ps6A6PVw | 2,857 | CNCF TAG Runtime and/or Working Group AI Discussion/Presentation | {
"login": "rajaskakodkar",
"id": 23639907,
"node_id": "MDQ6VXNlcjIzNjM5OTA3",
"avatar_url": "https://avatars.githubusercontent.com/u/23639907?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rajaskakodkar",
"html_url": "https://github.com/rajaskakodkar",
"followers_url": "https://api.githu... | [] | closed | false | null | [] | null | 3 | 2024-03-01T07:07:35 | 2024-05-23T05:44:13 | 2024-05-17T22:50:52 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hello Ollama maintainers,
We at CNCF TAG Runtime which is the Technical Advisory Group for workload execution on cloud native infrastructure in the Cloud Native Computing Foundation are interested to explore the space of intersection between AI and cloud native technologies.
As the Tech Lead of [CNCF TAG Runtime]... | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2857/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2857/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2978 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2978/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2978/comments | https://api.github.com/repos/ollama/ollama/issues/2978/events | https://github.com/ollama/ollama/issues/2978 | 2,173,509,806 | I_kwDOJ0Z1Ps6BjSCu | 2,978 | windows generate error | {
"login": "SilentEchoe",
"id": 33675262,
"node_id": "MDQ6VXNlcjMzNjc1MjYy",
"avatar_url": "https://avatars.githubusercontent.com/u/33675262?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/SilentEchoe",
"html_url": "https://github.com/SilentEchoe",
"followers_url": "https://api.github.com/... | [] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 8 | 2024-03-07T10:21:07 | 2024-05-22T01:16:49 | 2024-03-08T08:05:54 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi:
I'm getting a build error when building with Windows
The error message is as follows:
Building LCD CPU
generating config with: cmake -S ../llama.cpp -B ../llama.cpp/build/windows/amd64/cpu -DCMAKE_POSITION_INDEPENDENT_CODE=on -DLLAMA_AVX=off -DLLAMA_AVX2=off -DLLAMA_AVX512=off -DLLAMA_FMA=off -DLLAMA_F16C=of... | {
"login": "SilentEchoe",
"id": 33675262,
"node_id": "MDQ6VXNlcjMzNjc1MjYy",
"avatar_url": "https://avatars.githubusercontent.com/u/33675262?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/SilentEchoe",
"html_url": "https://github.com/SilentEchoe",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2978/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2978/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4419 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4419/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4419/comments | https://api.github.com/repos/ollama/ollama/issues/4419/events | https://github.com/ollama/ollama/pull/4419 | 2,294,188,313 | PR_kwDOJ0Z1Ps5vVVUX | 4,419 | add load timeout env | {
"login": "lengrongfu",
"id": 15009201,
"node_id": "MDQ6VXNlcjE1MDA5MjAx",
"avatar_url": "https://avatars.githubusercontent.com/u/15009201?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lengrongfu",
"html_url": "https://github.com/lengrongfu",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 3 | 2024-05-14T01:58:06 | 2024-05-23T21:11:32 | 2024-05-23T21:11:31 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4419",
"html_url": "https://github.com/ollama/ollama/pull/4419",
"diff_url": "https://github.com/ollama/ollama/pull/4419.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4419.patch",
"merged_at": null
} | Fixes: https://github.com/ollama/ollama/issues/4350 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4419/reactions",
"total_count": 3,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4419/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5413 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5413/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5413/comments | https://api.github.com/repos/ollama/ollama/issues/5413/events | https://github.com/ollama/ollama/issues/5413 | 2,384,450,337 | I_kwDOJ0Z1Ps6OH9Mh | 5,413 | CodeSage Base, Large, and Small | {
"login": "unclemusclez",
"id": 8789242,
"node_id": "MDQ6VXNlcjg3ODkyNDI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8789242?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/unclemusclez",
"html_url": "https://github.com/unclemusclez",
"followers_url": "https://api.github.com... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 1 | 2024-07-01T18:11:59 | 2024-12-04T07:42:37 | 2024-12-04T07:42:37 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | https://github.com/amazon-science/CodeSage
https://huggingface.co/codesage/codesage-base
https://huggingface.co/codesage/codesage-large
https://huggingface.co/codesage/codesage-small | {
"login": "unclemusclez",
"id": 8789242,
"node_id": "MDQ6VXNlcjg3ODkyNDI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8789242?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/unclemusclez",
"html_url": "https://github.com/unclemusclez",
"followers_url": "https://api.github.com... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5413/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5413/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7877 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7877/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7877/comments | https://api.github.com/repos/ollama/ollama/issues/7877/events | https://github.com/ollama/ollama/issues/7877 | 2,703,672,753 | I_kwDOJ0Z1Ps6hJsWx | 7,877 | Add a ShowUI model | {
"login": "barinov274",
"id": 54442153,
"node_id": "MDQ6VXNlcjU0NDQyMTUz",
"avatar_url": "https://avatars.githubusercontent.com/u/54442153?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/barinov274",
"html_url": "https://github.com/barinov274",
"followers_url": "https://api.github.com/use... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | open | false | null | [] | null | 1 | 2024-11-29T02:36:15 | 2024-11-29T02:46:51 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | https://huggingface.co/showlab/ShowUI-2B | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7877/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7877/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/2423 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2423/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2423/comments | https://api.github.com/repos/ollama/ollama/issues/2423/events | https://github.com/ollama/ollama/pull/2423 | 2,126,561,400 | PR_kwDOJ0Z1Ps5mc1Tk | 2,423 | Install: don't overwrite systemd config if it exists | {
"login": "rvagg",
"id": 495647,
"node_id": "MDQ6VXNlcjQ5NTY0Nw==",
"avatar_url": "https://avatars.githubusercontent.com/u/495647?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rvagg",
"html_url": "https://github.com/rvagg",
"followers_url": "https://api.github.com/users/rvagg/followers"... | [] | closed | false | null | [] | null | 1 | 2024-02-09T06:37:12 | 2024-02-12T00:50:23 | 2024-02-12T00:50:17 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2423",
"html_url": "https://github.com/ollama/ollama/pull/2423",
"diff_url": "https://github.com/ollama/ollama/pull/2423.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2423.patch",
"merged_at": null
} | A little annoying if you have a custom config of any kind, like an `OLLAMA_HOST` variable; but perhaps also you've set it up custom in other ways, like a custom user account.
(I've used this locally and it works great) | {
"login": "rvagg",
"id": 495647,
"node_id": "MDQ6VXNlcjQ5NTY0Nw==",
"avatar_url": "https://avatars.githubusercontent.com/u/495647?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rvagg",
"html_url": "https://github.com/rvagg",
"followers_url": "https://api.github.com/users/rvagg/followers"... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2423/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2423/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7068 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7068/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7068/comments | https://api.github.com/repos/ollama/ollama/issues/7068/events | https://github.com/ollama/ollama/pull/7068 | 2,560,243,583 | PR_kwDOJ0Z1Ps59TwgR | 7,068 | runner.go: Enable llamafile (all platforms) and BLAS (Mac OS) | {
"login": "jessegross",
"id": 6468499,
"node_id": "MDQ6VXNlcjY0Njg0OTk=",
"avatar_url": "https://avatars.githubusercontent.com/u/6468499?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jessegross",
"html_url": "https://github.com/jessegross",
"followers_url": "https://api.github.com/users... | [] | closed | false | null | [] | null | 0 | 2024-10-01T21:26:52 | 2024-10-01T22:58:52 | 2024-10-01T22:58:50 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7068",
"html_url": "https://github.com/ollama/ollama/pull/7068",
"diff_url": "https://github.com/ollama/ollama/pull/7068.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7068.patch",
"merged_at": "2024-10-01T22:58:50"
} | These are two features that are shown on llama.cpp's system info that are currently different between the two runners. On my test systems the performance difference is very small to negligible but it is probably still good to equalize the features. | {
"login": "jessegross",
"id": 6468499,
"node_id": "MDQ6VXNlcjY0Njg0OTk=",
"avatar_url": "https://avatars.githubusercontent.com/u/6468499?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jessegross",
"html_url": "https://github.com/jessegross",
"followers_url": "https://api.github.com/users... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7068/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7068/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2664 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2664/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2664/comments | https://api.github.com/repos/ollama/ollama/issues/2664/events | https://github.com/ollama/ollama/issues/2664 | 2,148,224,628 | I_kwDOJ0Z1Ps6AC050 | 2,664 | Ollama windows installer fails due to virus/trojan detection from Windows Defender!! Help | {
"login": "ewebgh33",
"id": 123797054,
"node_id": "U_kgDOB2D-Pg",
"avatar_url": "https://avatars.githubusercontent.com/u/123797054?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ewebgh33",
"html_url": "https://github.com/ewebgh33",
"followers_url": "https://api.github.com/users/ewebgh33/... | [] | closed | false | null | [] | null | 2 | 2024-02-22T05:21:00 | 2024-02-22T06:40:40 | 2024-02-22T06:40:39 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | As title says
Downloaded the windows installer literally a few mins ago.
Clicked install... window pops up, progress bar counts up... then womp! It Disappears.
Little notification in the corner of windows, I ignore.
Right, where did it go? Hmm. Weird. Can't see Ollama anywhere. Maybe it didn't install?
Try agai... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2664/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2664/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/767 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/767/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/767/comments | https://api.github.com/repos/ollama/ollama/issues/767/events | https://github.com/ollama/ollama/issues/767 | 1,940,344,125 | I_kwDOJ0Z1Ps5zp009 | 767 | Install ollama binary to user home directory? | {
"login": "pneumic",
"id": 80791807,
"node_id": "MDQ6VXNlcjgwNzkxODA3",
"avatar_url": "https://avatars.githubusercontent.com/u/80791807?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pneumic",
"html_url": "https://github.com/pneumic",
"followers_url": "https://api.github.com/users/pneumi... | [] | closed | false | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api... | null | 4 | 2023-10-12T16:20:18 | 2023-10-16T03:11:29 | 2023-10-13T15:14:59 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Context: https://github.com/jmorganca/ollama/issues/429#issuecomment-1694642023
Are there plans to provide an option for the user to install the binary to their home directory? In my particular case my work machine limits custom software installs to my home directory, for example. | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/767/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/767/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7494 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7494/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7494/comments | https://api.github.com/repos/ollama/ollama/issues/7494/events | https://github.com/ollama/ollama/issues/7494 | 2,633,528,418 | I_kwDOJ0Z1Ps6c-HRi | 7,494 | Molmo by Ai2 | {
"login": "haimat",
"id": 6633976,
"node_id": "MDQ6VXNlcjY2MzM5NzY=",
"avatar_url": "https://avatars.githubusercontent.com/u/6633976?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/haimat",
"html_url": "https://github.com/haimat",
"followers_url": "https://api.github.com/users/haimat/foll... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 1 | 2024-11-04T18:19:23 | 2024-11-04T18:56:05 | 2024-11-04T18:56:05 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | [Molmo-7B-D-0924](https://huggingface.co/allenai/Molmo-7B-D-0924) | {
"login": "haimat",
"id": 6633976,
"node_id": "MDQ6VXNlcjY2MzM5NzY=",
"avatar_url": "https://avatars.githubusercontent.com/u/6633976?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/haimat",
"html_url": "https://github.com/haimat",
"followers_url": "https://api.github.com/users/haimat/foll... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7494/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7494/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7515 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7515/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7515/comments | https://api.github.com/repos/ollama/ollama/issues/7515/events | https://github.com/ollama/ollama/pull/7515 | 2,636,183,637 | PR_kwDOJ0Z1Ps6A90DN | 7,515 | One corrupt manifest should not wedge model operations | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 2 | 2024-11-05T18:32:36 | 2024-11-05T22:21:48 | 2024-11-05T22:21:45 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7515",
"html_url": "https://github.com/ollama/ollama/pull/7515",
"diff_url": "https://github.com/ollama/ollama/pull/7515.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7515.patch",
"merged_at": "2024-11-05T22:21:45"
} | One potential failure mode is an empty file which bubbles up as an EOF error, leading to all attempts to pull anything failing, and listing operations failing. Instead, continue and warn about the corrupt manifest, and allow re-pulling the model to clean up the corruption.
Fixes #7393
Fixes #6920
| {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7515/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7515/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1404 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1404/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1404/comments | https://api.github.com/repos/ollama/ollama/issues/1404/events | https://github.com/ollama/ollama/issues/1404 | 2,029,283,543 | I_kwDOJ0Z1Ps549GjX | 1,404 | Add support for llamafile | {
"login": "rupurt",
"id": 680789,
"node_id": "MDQ6VXNlcjY4MDc4OQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/680789?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rupurt",
"html_url": "https://github.com/rupurt",
"followers_url": "https://api.github.com/users/rupurt/follow... | [] | closed | false | null | [] | null | 3 | 2023-12-06T19:55:44 | 2024-07-24T05:05:17 | 2023-12-06T20:22:33 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Mozilla has announced a new file format like the modelfile but compiled to a single executable. Are there any plans to support it?
https://github.com/Mozilla-Ocho/llamafile | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1404/reactions",
"total_count": 3,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 3
} | https://api.github.com/repos/ollama/ollama/issues/1404/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2957 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2957/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2957/comments | https://api.github.com/repos/ollama/ollama/issues/2957/events | https://github.com/ollama/ollama/pull/2957 | 2,172,195,005 | PR_kwDOJ0Z1Ps5o4GyJ | 2,957 | Update gpu.go with additional RocmLinuxGlob | {
"login": "sroecker",
"id": 362733,
"node_id": "MDQ6VXNlcjM2MjczMw==",
"avatar_url": "https://avatars.githubusercontent.com/u/362733?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sroecker",
"html_url": "https://github.com/sroecker",
"followers_url": "https://api.github.com/users/sroecke... | [] | closed | false | null | [] | null | 1 | 2024-03-06T18:48:39 | 2024-03-17T15:09:07 | 2024-03-17T15:09:07 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2957",
"html_url": "https://github.com/ollama/ollama/pull/2957",
"diff_url": "https://github.com/ollama/ollama/pull/2957.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2957.patch",
"merged_at": null
} | Add additional RocmLinuxGlob for e.g Fedora
This leads to a correctly discovered GPU library in the log.
However something else needs to be fixed as a cpu_avx2 server is loaded. | {
"login": "sroecker",
"id": 362733,
"node_id": "MDQ6VXNlcjM2MjczMw==",
"avatar_url": "https://avatars.githubusercontent.com/u/362733?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sroecker",
"html_url": "https://github.com/sroecker",
"followers_url": "https://api.github.com/users/sroecke... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2957/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2957/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4605 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4605/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4605/comments | https://api.github.com/repos/ollama/ollama/issues/4605/events | https://github.com/ollama/ollama/pull/4605 | 2,314,441,606 | PR_kwDOJ0Z1Ps5wacey | 4,605 | Set darwin CI timeout to longer | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 0 | 2024-05-24T05:46:18 | 2024-05-24T05:46:24 | 2024-05-24T05:46:23 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4605",
"html_url": "https://github.com/ollama/ollama/pull/4605",
"diff_url": "https://github.com/ollama/ollama/pull/4605.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4605.patch",
"merged_at": "2024-05-24T05:46:23"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4605/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4605/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8128 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8128/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8128/comments | https://api.github.com/repos/ollama/ollama/issues/8128/events | https://github.com/ollama/ollama/pull/8128 | 2,743,779,305 | PR_kwDOJ0Z1Ps6FbpAX | 8,128 | server: loosen json format constraints | {
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/... | [] | closed | false | null | [] | null | 2 | 2024-12-17T02:23:27 | 2024-12-17T05:53:26 | 2024-12-17T05:45:36 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8128",
"html_url": "https://github.com/ollama/ollama/pull/8128",
"diff_url": "https://github.com/ollama/ollama/pull/8128.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8128.patch",
"merged_at": null
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8128/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8128/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6996 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6996/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6996/comments | https://api.github.com/repos/ollama/ollama/issues/6996/events | https://github.com/ollama/ollama/issues/6996 | 2,552,118,768 | I_kwDOJ0Z1Ps6YHj3w | 6,996 | pulling manifest Error: 403: | {
"login": "JitaekJo",
"id": 144883428,
"node_id": "U_kgDOCKK-5A",
"avatar_url": "https://avatars.githubusercontent.com/u/144883428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/JitaekJo",
"html_url": "https://github.com/JitaekJo",
"followers_url": "https://api.github.com/users/JitaekJo/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg... | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 1 | 2024-09-27T06:34:25 | 2024-10-23T00:14:58 | 2024-10-23T00:14:58 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Dears
I faced an error when I pull models.
I don't have any idea. Could you please help?

### OS
Windows
### GPU
Intel
### CPU
Intel
### Ollama version
0.3.12 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6996/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6996/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2450 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2450/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2450/comments | https://api.github.com/repos/ollama/ollama/issues/2450/events | https://github.com/ollama/ollama/issues/2450 | 2,129,148,286 | I_kwDOJ0Z1Ps5-6Dl- | 2,450 | Add ollama-zig library to README.md | {
"login": "tr1ckydev",
"id": 86865279,
"node_id": "MDQ6VXNlcjg2ODY1Mjc5",
"avatar_url": "https://avatars.githubusercontent.com/u/86865279?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tr1ckydev",
"html_url": "https://github.com/tr1ckydev",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5667396191,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aXw",
"url": "https://api.github.com/repos/ollama/ollama/labels/documentation",
"name": "documentation",
"color": "0075ca",
"default": true,
"description": "Improvements or additions to documentation"
}
] | closed | false | null | [] | null | 1 | 2024-02-11T20:08:09 | 2024-05-11T00:30:04 | 2024-05-11T00:30:03 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I have created zig library for the Ollama REST API.
https://github.com/tr1ckydev/ollama-zig | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2450/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2450/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2548 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2548/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2548/comments | https://api.github.com/repos/ollama/ollama/issues/2548/events | https://github.com/ollama/ollama/issues/2548 | 2,139,186,955 | I_kwDOJ0Z1Ps5_gWcL | 2,548 | Where the models installed, I installed llama2 and I am not sure I want to keep it I dont have much space (windows)Help?. | {
"login": "APISeeker",
"id": 145594487,
"node_id": "U_kgDOCK2Ydw",
"avatar_url": "https://avatars.githubusercontent.com/u/145594487?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/APISeeker",
"html_url": "https://github.com/APISeeker",
"followers_url": "https://api.github.com/users/APISee... | [] | closed | false | null | [] | null | 3 | 2024-02-16T18:34:22 | 2024-02-18T06:12:07 | 2024-02-18T06:12:07 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | 
So I just installed ollama and wrote a comman,d to download llama2, but I dont see much, here is a screenshot and nothing indicating presence of models?
| {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2548/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2548/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1665 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1665/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1665/comments | https://api.github.com/repos/ollama/ollama/issues/1665/events | https://github.com/ollama/ollama/issues/1665 | 2,053,105,447 | I_kwDOJ0Z1Ps56X-cn | 1,665 | I refreshed my local repo and now get this error, Not finding libcublas.so | {
"login": "phalexo",
"id": 4603365,
"node_id": "MDQ6VXNlcjQ2MDMzNjU=",
"avatar_url": "https://avatars.githubusercontent.com/u/4603365?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/phalexo",
"html_url": "https://github.com/phalexo",
"followers_url": "https://api.github.com/users/phalexo/... | [] | closed | false | null | [] | null | 1 | 2023-12-21T21:59:17 | 2023-12-21T22:12:03 | 2023-12-21T22:12:03 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | llm/llama.go:58:12: pattern llama.cpp/*/build/*/bin/*: cannot embed file llama.cpp/ggml/build/cuda/bin/libcublas.so.: invalid name libcublas.so | {
"login": "phalexo",
"id": 4603365,
"node_id": "MDQ6VXNlcjQ2MDMzNjU=",
"avatar_url": "https://avatars.githubusercontent.com/u/4603365?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/phalexo",
"html_url": "https://github.com/phalexo",
"followers_url": "https://api.github.com/users/phalexo/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1665/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1665/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3886 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3886/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3886/comments | https://api.github.com/repos/ollama/ollama/issues/3886/events | https://github.com/ollama/ollama/pull/3886 | 2,261,934,869 | PR_kwDOJ0Z1Ps5toyR6 | 3,886 | types/model: make ParseName use default without question | {
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers"... | [] | closed | false | null | [] | null | 0 | 2024-04-24T18:35:27 | 2024-04-24T18:52:55 | 2024-04-24T18:52:55 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3886",
"html_url": "https://github.com/ollama/ollama/pull/3886",
"diff_url": "https://github.com/ollama/ollama/pull/3886.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3886.patch",
"merged_at": "2024-04-24T18:52:55"
} | null | {
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers"... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3886/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3886/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1436 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1436/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1436/comments | https://api.github.com/repos/ollama/ollama/issues/1436/events | https://github.com/ollama/ollama/issues/1436 | 2,032,983,003 | I_kwDOJ0Z1Ps55LNvb | 1,436 | Windows binary race condition | {
"login": "csaben",
"id": 76020733,
"node_id": "MDQ6VXNlcjc2MDIwNzMz",
"avatar_url": "https://avatars.githubusercontent.com/u/76020733?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/csaben",
"html_url": "https://github.com/csaben",
"followers_url": "https://api.github.com/users/csaben/fo... | [] | closed | false | null | [] | null | 2 | 2023-12-08T16:36:04 | 2024-01-27T01:48:34 | 2024-01-27T01:48:34 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Cross compiled a working ollama.exe for windows but meet the following error when running a model.
```zsh
./ollama.exe run llama2
```
Errors with
```
Error: Post "http://127.0.0.1:11434/api/generate": read tcp 127.0.0.1:52248->127.0.0.1:11434: wsarecv: An existing connection was forcibly closed by the remote ho... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1436/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1436/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1196 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1196/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1196/comments | https://api.github.com/repos/ollama/ollama/issues/1196/events | https://github.com/ollama/ollama/issues/1196 | 2,000,793,782 | I_kwDOJ0Z1Ps53QbC2 | 1,196 | Model only working from cli but not API | {
"login": "skadefro",
"id": 4155937,
"node_id": "MDQ6VXNlcjQxNTU5Mzc=",
"avatar_url": "https://avatars.githubusercontent.com/u/4155937?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/skadefro",
"html_url": "https://github.com/skadefro",
"followers_url": "https://api.github.com/users/skade... | [] | closed | false | null | [] | null | 1 | 2023-11-19T10:39:01 | 2023-11-19T10:58:59 | 2023-11-19T10:58:58 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I downloaded TheBloke/deepseek-coder-6.7B-instruct-GGUF
I create a Modelfile and add
```
FROM ./deepseek-coder-6.7b-instruct.Q6_K.gguf
```
and run
```
ollama create skadefro/deepseek -f Modelfile
ollama run skadefro/deepseek
```
No issues.
I then try and run it from code ( using [litellm](https://github.com/... | {
"login": "skadefro",
"id": 4155937,
"node_id": "MDQ6VXNlcjQxNTU5Mzc=",
"avatar_url": "https://avatars.githubusercontent.com/u/4155937?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/skadefro",
"html_url": "https://github.com/skadefro",
"followers_url": "https://api.github.com/users/skade... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1196/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1196/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6623 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6623/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6623/comments | https://api.github.com/repos/ollama/ollama/issues/6623/events | https://github.com/ollama/ollama/issues/6623 | 2,504,178,908 | I_kwDOJ0Z1Ps6VQrzc | 6,623 | nvidia/NV-Embed-v2 support | {
"login": "youxiaoxing",
"id": 44542133,
"node_id": "MDQ6VXNlcjQ0NTQyMTMz",
"avatar_url": "https://avatars.githubusercontent.com/u/44542133?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/youxiaoxing",
"html_url": "https://github.com/youxiaoxing",
"followers_url": "https://api.github.com/... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | open | false | null | [] | null | 14 | 2024-09-04T02:48:55 | 2024-11-15T12:54:38 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Can you support the NVIDIA/NV-Embed-v2 model?
https://huggingface.co/nvidia/NV-Embed-v2 | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6623/reactions",
"total_count": 39,
"+1": 39,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6623/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/2367 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2367/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2367/comments | https://api.github.com/repos/ollama/ollama/issues/2367/events | https://github.com/ollama/ollama/issues/2367 | 2,119,799,833 | I_kwDOJ0Z1Ps5-WZQZ | 2,367 | `/api/chat` and `/api/generate` hang if image cannot be decoded | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.g... | null | 4 | 2024-02-06T02:14:44 | 2024-02-20T04:00:09 | 2024-02-20T04:00:09 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ```
llama_new_context_with_model: graph splits (measure): 3
2024/02/05 21:13:46 dyn_ext_server.go:156: INFO Starting llama main loop
2024/02/05 21:13:46 dyn_ext_server.go:170: INFO loaded 1 images
clip_image_load_from_bytes: failed to decode image bytes
``` | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2367/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2367/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5623 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5623/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5623/comments | https://api.github.com/repos/ollama/ollama/issues/5623/events | https://github.com/ollama/ollama/pull/5623 | 2,402,016,303 | PR_kwDOJ0Z1Ps51CNCB | 5,623 | Update README.md with new terminal tool ParLlama | {
"login": "paulrobello",
"id": 955011,
"node_id": "MDQ6VXNlcjk1NTAxMQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/955011?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/paulrobello",
"html_url": "https://github.com/paulrobello",
"followers_url": "https://api.github.com/user... | [] | closed | false | null | [] | null | 1 | 2024-07-11T01:20:08 | 2024-11-21T10:55:35 | 2024-11-21T10:55:35 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5623",
"html_url": "https://github.com/ollama/ollama/pull/5623",
"diff_url": "https://github.com/ollama/ollama/pull/5623.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5623.patch",
"merged_at": "2024-11-21T10:55:35"
} | null | {
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/us... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5623/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5623/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3306 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3306/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3306/comments | https://api.github.com/repos/ollama/ollama/issues/3306/events | https://github.com/ollama/ollama/issues/3306 | 2,203,720,292 | I_kwDOJ0Z1Ps6DWhpk | 3,306 | No GPU found! | {
"login": "InfoOfInfo",
"id": 164606679,
"node_id": "U_kgDOCc-y1w",
"avatar_url": "https://avatars.githubusercontent.com/u/164606679?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/InfoOfInfo",
"html_url": "https://github.com/InfoOfInfo",
"followers_url": "https://api.github.com/users/Inf... | [] | closed | false | null | [] | null | 2 | 2024-03-23T06:14:10 | 2024-04-07T00:07:45 | 2024-03-23T06:51:52 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What are you trying to do?
root@localhost:~# ollama serve
time=2024-03-23T11:28:06.617+05:30 level=INFO source=images.go:806 msg="total blobs: 0"
time=2024-03-23T11:28:06.682+05:30 level=INFO source=images.go:813 msg="total unused blobs removed: 0"
time=2024-03-23T11:28:06.685+05:30 level=INFO source=rout... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3306/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3306/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5514 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5514/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5514/comments | https://api.github.com/repos/ollama/ollama/issues/5514/events | https://github.com/ollama/ollama/pull/5514 | 2,393,433,448 | PR_kwDOJ0Z1Ps50lP-6 | 5,514 | Update community integration - Painting Droid | {
"login": "mateuszmigas",
"id": 54471371,
"node_id": "MDQ6VXNlcjU0NDcxMzcx",
"avatar_url": "https://avatars.githubusercontent.com/u/54471371?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mateuszmigas",
"html_url": "https://github.com/mateuszmigas",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | 1 | 2024-07-06T06:22:00 | 2024-09-03T20:15:55 | 2024-09-03T20:15:54 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5514",
"html_url": "https://github.com/ollama/ollama/pull/5514",
"diff_url": "https://github.com/ollama/ollama/pull/5514.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5514.patch",
"merged_at": "2024-09-03T20:15:54"
} | I've added integration with Ollama to [Painting Droid](https://github.com/mateuszmigas/painting-droid). Users can now ask Llava about stuff they are drawing. More integration is coming to allow Ollama to invoke drawing functions within the app.
` iter
follow up to #3715 | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3718/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3718/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1071 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1071/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1071/comments | https://api.github.com/repos/ollama/ollama/issues/1071/events | https://github.com/ollama/ollama/issues/1071 | 1,986,955,471 | I_kwDOJ0Z1Ps52bojP | 1,071 | Enhancement: Enable GPU Support for NVIDIA JetPack and the NVIDIA Jetson Lineup | {
"login": "bnodnarb",
"id": 97063458,
"node_id": "U_kgDOBckSIg",
"avatar_url": "https://avatars.githubusercontent.com/u/97063458?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bnodnarb",
"html_url": "https://github.com/bnodnarb",
"followers_url": "https://api.github.com/users/bnodnarb/fo... | [] | closed | false | null | [] | null | 8 | 2023-11-10T05:49:09 | 2023-11-12T13:49:14 | 2023-11-12T11:04:54 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi all,
I recently purchased an [NVIDIA Jetson Orin Developer Kit](https://store.nvidia.com/en-us/jetson/store/?page=1&limit=9&locale=en-us) and am hoping to get Ollama running on it.
These little powerhouses are specifically built for AI applications and they have a ton of capability crammed into a tiny form fac... | {
"login": "bnodnarb",
"id": 97063458,
"node_id": "U_kgDOBckSIg",
"avatar_url": "https://avatars.githubusercontent.com/u/97063458?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bnodnarb",
"html_url": "https://github.com/bnodnarb",
"followers_url": "https://api.github.com/users/bnodnarb/fo... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1071/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1071/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4938 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4938/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4938/comments | https://api.github.com/repos/ollama/ollama/issues/4938/events | https://github.com/ollama/ollama/pull/4938 | 2,341,804,102 | PR_kwDOJ0Z1Ps5x3lvZ | 4,938 | fix parsing big endian gguf | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2024-06-08T19:34:13 | 2024-06-10T16:38:13 | 2024-06-10T16:38:12 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4938",
"html_url": "https://github.com/ollama/ollama/pull/4938",
"diff_url": "https://github.com/ollama/ollama/pull/4938.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4938.patch",
"merged_at": "2024-06-10T16:38:12"
} | most gguf files will be little endian but a big endian file should not panic | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4938/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4938/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2185 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2185/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2185/comments | https://api.github.com/repos/ollama/ollama/issues/2185/events | https://github.com/ollama/ollama/issues/2185 | 2,099,973,484 | I_kwDOJ0Z1Ps59Kw1s | 2,185 | Python example does not work | {
"login": "claremont-awilson",
"id": 110598735,
"node_id": "U_kgDOBpeaTw",
"avatar_url": "https://avatars.githubusercontent.com/u/110598735?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/claremont-awilson",
"html_url": "https://github.com/claremont-awilson",
"followers_url": "https://api... | [] | closed | false | null | [] | null | 2 | 2024-01-25T09:44:19 | 2024-11-07T10:07:50 | 2024-01-25T10:15:51 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Reading this: https://ollama.ai/blog/python-javascript-libraries
Trying to perform the Python example errors on macOS and OEL9:
```python
import ollama
response = ollama.chat(
model="llama2",
messages=[
{
"role": "user",
"content": "Why is the sky blue?",
},... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2185/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2185/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/302 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/302/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/302/comments | https://api.github.com/repos/ollama/ollama/issues/302/events | https://github.com/ollama/ollama/pull/302 | 1,839,821,450 | PR_kwDOJ0Z1Ps5XWtGp | 302 | langchain JS integration | {
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/us... | [] | closed | false | null | [] | null | 0 | 2023-08-07T16:21:06 | 2023-08-07T16:21:37 | 2023-08-07T16:21:37 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/302",
"html_url": "https://github.com/ollama/ollama/pull/302",
"diff_url": "https://github.com/ollama/ollama/pull/302.diff",
"patch_url": "https://github.com/ollama/ollama/pull/302.patch",
"merged_at": "2023-08-07T16:21:36"
} | langchain JS integration | {
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/us... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/302/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/302/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3056 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3056/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3056/comments | https://api.github.com/repos/ollama/ollama/issues/3056/events | https://github.com/ollama/ollama/pull/3056 | 2,179,477,779 | PR_kwDOJ0Z1Ps5pQxPq | 3,056 | Avoid rocm runner and dependency clash | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-03-11T15:47:46 | 2024-03-11T16:48:51 | 2024-03-11T16:48:48 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3056",
"html_url": "https://github.com/ollama/ollama/pull/3056",
"diff_url": "https://github.com/ollama/ollama/pull/3056.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3056.patch",
"merged_at": "2024-03-11T16:48:48"
} | Putting the rocm symlink next to the runners is risky. This moves the payloads into a subdir to avoid potential clashes.
Fixes #3035
Untested at the moment... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3056/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3056/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8572 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8572/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8572/comments | https://api.github.com/repos/ollama/ollama/issues/8572/events | https://github.com/ollama/ollama/issues/8572 | 2,810,750,132 | I_kwDOJ0Z1Ps6niKS0 | 8,572 | Ollama cannot start because it try to create an existing directory | {
"login": "brianhuster",
"id": 111893501,
"node_id": "U_kgDOBqtb_Q",
"avatar_url": "https://avatars.githubusercontent.com/u/111893501?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/brianhuster",
"html_url": "https://github.com/brianhuster",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 1 | 2025-01-25T04:16:27 | 2025-01-25T08:37:23 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Since my main partition doesn't have enough memory, I change OLLAMA_MODELS to /media/brianhuster/E/ollama/models. I have created that directory, and `mv /usr/share/ollama/.ollama/models /media/brianhuster/E/ollama/models`. However, after that I cannot restart Ollama because, according to `sudo j... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8572/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8572/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/5762 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5762/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5762/comments | https://api.github.com/repos/ollama/ollama/issues/5762/events | https://github.com/ollama/ollama/issues/5762 | 2,415,254,514 | I_kwDOJ0Z1Ps6P9dvy | 5,762 | How to pass a custom file (txt\doc\pdf) to prompt through the API | {
"login": "liukesoft",
"id": 15062018,
"node_id": "MDQ6VXNlcjE1MDYyMDE4",
"avatar_url": "https://avatars.githubusercontent.com/u/15062018?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/liukesoft",
"html_url": "https://github.com/liukesoft",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] | closed | false | null | [] | null | 5 | 2024-07-18T05:14:02 | 2024-11-17T15:40:12 | 2024-11-17T15:40:12 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hello, I have deployed Olama on my local Linux and would like to call it through an API. The model is llama3, and prompt is my custom Python template, which allows the model to generate new Python code according to my Python template. How can I send my template (txt \ doc \ pdf) to the interface through API calls? How ... | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5762/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5762/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5209 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5209/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5209/comments | https://api.github.com/repos/ollama/ollama/issues/5209/events | https://github.com/ollama/ollama/pull/5209 | 2,367,395,231 | PR_kwDOJ0Z1Ps5zOwjY | 5,209 | OpenAI: v1/completions compatibility | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjha... | [] | closed | false | null | [] | null | 4 | 2024-06-22T00:46:53 | 2024-07-08T14:30:21 | 2024-07-02T23:01:45 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5209",
"html_url": "https://github.com/ollama/ollama/pull/5209",
"diff_url": "https://github.com/ollama/ollama/pull/5209.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5209.patch",
"merged_at": "2024-07-02T23:01:45"
} | `curl http://localhost:11434/v1/completions \
-H "Content-Type: application/json" \
-d '{
"model": "llama3",
"prompt": "Say this is a test",
"temperature": 0
}' | jq`
```
{
"id": "cmpl-210",
"object": "text_completion",
"created": 1719010238,
"model": "llama3",
... | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjha... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5209/reactions",
"total_count": 7,
"+1": 4,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 3,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5209/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1944 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1944/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1944/comments | https://api.github.com/repos/ollama/ollama/issues/1944/events | https://github.com/ollama/ollama/pull/1944 | 2,078,097,750 | PR_kwDOJ0Z1Ps5j4pwD | 1,944 | Add Dify.AI to community integrations | {
"login": "guchenhe",
"id": 13230914,
"node_id": "MDQ6VXNlcjEzMjMwOTE0",
"avatar_url": "https://avatars.githubusercontent.com/u/13230914?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/guchenhe",
"html_url": "https://github.com/guchenhe",
"followers_url": "https://api.github.com/users/guc... | [] | closed | false | null | [] | null | 1 | 2024-01-12T05:27:02 | 2024-03-25T19:06:39 | 2024-03-25T19:06:39 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1944",
"html_url": "https://github.com/ollama/ollama/pull/1944",
"diff_url": "https://github.com/ollama/ollama/pull/1944.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1944.patch",
"merged_at": "2024-03-25T19:06:39"
} | Dify.AI is a model-agnostic LLMOps platform for building and managing LLM applications. Ollama is now a supported model provider on the platform, as documented in https://docs.dify.ai/advanced/model-configuration/ollama | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1944/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1944/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/752 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/752/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/752/comments | https://api.github.com/repos/ollama/ollama/issues/752/events | https://github.com/ollama/ollama/pull/752 | 1,936,345,717 | PR_kwDOJ0Z1Ps5ccDDK | 752 | prevent waiting on exited command | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 0 | 2023-10-10T22:03:14 | 2023-10-11T16:32:15 | 2023-10-11T16:32:14 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/752",
"html_url": "https://github.com/ollama/ollama/pull/752",
"diff_url": "https://github.com/ollama/ollama/pull/752.diff",
"patch_url": "https://github.com/ollama/ollama/pull/752.patch",
"merged_at": "2023-10-11T16:32:14"
} | this fixes the issue in main from #724 that causes the server to wait on a subprocess which has already exited | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/752/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/752/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4488 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4488/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4488/comments | https://api.github.com/repos/ollama/ollama/issues/4488/events | https://github.com/ollama/ollama/issues/4488 | 2,301,744,058 | I_kwDOJ0Z1Ps6JMdO6 | 4,488 | can ollama support openai-whisper | {
"login": "gfkdliucheng",
"id": 24772003,
"node_id": "MDQ6VXNlcjI0NzcyMDAz",
"avatar_url": "https://avatars.githubusercontent.com/u/24772003?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gfkdliucheng",
"html_url": "https://github.com/gfkdliucheng",
"followers_url": "https://api.github.c... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 1 | 2024-05-17T03:36:41 | 2024-05-17T06:20:51 | 2024-05-17T06:20:50 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | can ollama support openai-whisper | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4488/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4488/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/840 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/840/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/840/comments | https://api.github.com/repos/ollama/ollama/issues/840/events | https://github.com/ollama/ollama/pull/840 | 1,950,357,682 | PR_kwDOJ0Z1Ps5dLJBR | 840 | do not reload the running llm when runtime params change | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 0 | 2023-10-18T18:10:23 | 2023-10-19T14:40:00 | 2023-10-19T14:39:59 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/840",
"html_url": "https://github.com/ollama/ollama/pull/840",
"diff_url": "https://github.com/ollama/ollama/pull/840.diff",
"patch_url": "https://github.com/ollama/ollama/pull/840.patch",
"merged_at": "2023-10-19T14:39:59"
} | - only reload the running llm if the model has changed, or the options for loading the running model have changed
- rename loaded llm to runner to differentiate from loaded model image
- remove logic which keeps the first system prompt in the generation context
Say I have 2 models, both are based on llama2, but th... | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/840/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/840/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4426 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4426/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4426/comments | https://api.github.com/repos/ollama/ollama/issues/4426/events | https://github.com/ollama/ollama/pull/4426 | 2,294,936,004 | PR_kwDOJ0Z1Ps5vX3N4 | 4,426 | Add qollama to list of Web & Desktop integrations | {
"login": "farleyrunkel",
"id": 162782461,
"node_id": "U_kgDOCbPc_Q",
"avatar_url": "https://avatars.githubusercontent.com/u/162782461?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/farleyrunkel",
"html_url": "https://github.com/farleyrunkel",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 1 | 2024-05-14T09:34:08 | 2024-06-06T07:34:20 | 2024-06-06T07:34:20 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4426",
"html_url": "https://github.com/ollama/ollama/pull/4426",
"diff_url": "https://github.com/ollama/ollama/pull/4426.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4426.patch",
"merged_at": null
} | QOllama is a Qt-based client for [ollama](https://github.com/ollama/ollama), providing a user-friendly interface for interacting with the model and managing chat history. It supports cross-platform functionality, ensuring a seamless experience on Windows, macOS, and Linux.
Goto QOllama: [https://github.com/farleyrun... | {
"login": "farleyrunkel",
"id": 162782461,
"node_id": "U_kgDOCbPc_Q",
"avatar_url": "https://avatars.githubusercontent.com/u/162782461?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/farleyrunkel",
"html_url": "https://github.com/farleyrunkel",
"followers_url": "https://api.github.com/use... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4426/reactions",
"total_count": 2,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 2,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4426/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6001 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6001/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6001/comments | https://api.github.com/repos/ollama/ollama/issues/6001/events | https://github.com/ollama/ollama/pull/6001 | 2,433,047,670 | PR_kwDOJ0Z1Ps52n60x | 6,001 | Increase Num Parallel for Embedding Models | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjha... | [] | closed | false | null | [] | null | 0 | 2024-07-26T22:18:57 | 2024-08-12T17:29:13 | 2024-08-12T17:29:13 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | true | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/6001",
"html_url": "https://github.com/ollama/ollama/pull/6001",
"diff_url": "https://github.com/ollama/ollama/pull/6001.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6001.patch",
"merged_at": null
} | negligible impact after num_parallel 4 | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjha... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6001/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/ollama/ollama/issues/6001/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2872 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2872/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2872/comments | https://api.github.com/repos/ollama/ollama/issues/2872/events | https://github.com/ollama/ollama/issues/2872 | 2,164,537,347 | I_kwDOJ0Z1Ps6BBDgD | 2,872 | I dont want the systemd service. | {
"login": "donuts-are-good",
"id": 96031819,
"node_id": "U_kgDOBblUSw",
"avatar_url": "https://avatars.githubusercontent.com/u/96031819?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/donuts-are-good",
"html_url": "https://github.com/donuts-are-good",
"followers_url": "https://api.github.... | [] | closed | false | null | [] | null | 3 | 2024-03-02T02:48:01 | 2024-04-09T02:33:49 | 2024-03-02T03:11:05 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Title says it. I dont want this installer to set up a systemd service without asking. I use `ollama serve`, so the systemd service isn't something i need or want. Upgrading is a pain as a result.
`curl -fsSL https://ollama.com/install.sh | sh` <-- need a better way to update | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2872/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2872/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5555 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5555/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5555/comments | https://api.github.com/repos/ollama/ollama/issues/5555/events | https://github.com/ollama/ollama/pull/5555 | 2,396,891,147 | PR_kwDOJ0Z1Ps50wz3d | 5,555 | Bundle missing CRT libraries | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-07-09T01:26:51 | 2024-07-10T19:50:05 | 2024-07-10T19:50:02 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5555",
"html_url": "https://github.com/ollama/ollama/pull/5555",
"diff_url": "https://github.com/ollama/ollama/pull/5555.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5555.patch",
"merged_at": "2024-07-10T19:50:02"
} | Some users are experienging runner startup errors due to not having these msvc redist libraries on their host
Fixes #4657
```
> dumpbin /dependents .\llm\build\windows\amd64\cpu\bin\ollama_llama_server.exe
Microsoft (R) COFF/PE Dumper Version 14.29.30154.0
Copyright (C) Microsoft Corporation. All rights rese... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5555/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5555/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7428 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7428/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7428/comments | https://api.github.com/repos/ollama/ollama/issues/7428/events | https://github.com/ollama/ollama/pull/7428 | 2,625,008,534 | PR_kwDOJ0Z1Ps6Abo89 | 7,428 | Soften windows clang requirement | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-10-30T18:42:27 | 2024-10-30T19:28:39 | 2024-10-30T19:28:36 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7428",
"html_url": "https://github.com/ollama/ollama/pull/7428",
"diff_url": "https://github.com/ollama/ollama/pull/7428.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7428.patch",
"merged_at": "2024-10-30T19:28:36"
} | This will no longer error if built with regular gcc on windows. To help triage issues that may come in related to different compilers, the runner now reports the compiler used by cgo.
Example runner output when compiled with GCC
```
time=2024-10-30T11:29:49.863-07:00 level=INFO source=runner.go:877 msg=system inf... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7428/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7428/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8031 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8031/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8031/comments | https://api.github.com/repos/ollama/ollama/issues/8031/events | https://github.com/ollama/ollama/pull/8031 | 2,730,911,786 | PR_kwDOJ0Z1Ps6Ev7B9 | 8,031 | build: fix typo in override variable | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-12-10T18:42:22 | 2024-12-10T18:51:20 | 2024-12-10T18:51:16 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8031",
"html_url": "https://github.com/ollama/ollama/pull/8031",
"diff_url": "https://github.com/ollama/ollama/pull/8031.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8031.patch",
"merged_at": "2024-12-10T18:51:16"
} | The "F" was missing. | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8031/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8031/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7465 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7465/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7465/comments | https://api.github.com/repos/ollama/ollama/issues/7465/events | https://github.com/ollama/ollama/issues/7465 | 2,629,738,547 | I_kwDOJ0Z1Ps6cvqAz | 7,465 | Error raised when using llama3.2 vision | {
"login": "uzhao",
"id": 5157125,
"node_id": "MDQ6VXNlcjUxNTcxMjU=",
"avatar_url": "https://avatars.githubusercontent.com/u/5157125?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/uzhao",
"html_url": "https://github.com/uzhao",
"followers_url": "https://api.github.com/users/uzhao/follower... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 11 | 2024-11-01T19:33:30 | 2024-11-07T21:03:58 | 2024-11-05T01:37:26 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
ollama run x/llama3.2-vision:latest "describe this image: /home/papillon/Downloads/objectdetection.jpg"
Added image '/home/papillon/Downloads/objectdetection.jpg'
Error: POST predict: Post "http://127.0.0.1:32927/completion": EOF
ollama run x/llama3.2-vision:latest "hello"
Error: POST pred... | {
"login": "uzhao",
"id": 5157125,
"node_id": "MDQ6VXNlcjUxNTcxMjU=",
"avatar_url": "https://avatars.githubusercontent.com/u/5157125?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/uzhao",
"html_url": "https://github.com/uzhao",
"followers_url": "https://api.github.com/users/uzhao/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7465/reactions",
"total_count": 2,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/ollama/ollama/issues/7465/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8651 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8651/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8651/comments | https://api.github.com/repos/ollama/ollama/issues/8651/events | https://github.com/ollama/ollama/issues/8651 | 2,817,464,249 | I_kwDOJ0Z1Ps6n7xe5 | 8,651 | Intel ARC 770 memory does not support | {
"login": "yiteei",
"id": 77902908,
"node_id": "MDQ6VXNlcjc3OTAyOTA4",
"avatar_url": "https://avatars.githubusercontent.com/u/77902908?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yiteei",
"html_url": "https://github.com/yiteei",
"followers_url": "https://api.github.com/users/yiteei/fo... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2025-01-29T07:42:05 | 2025-01-29T23:28:53 | 2025-01-29T23:28:52 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?

Windows 11 24H2
Intel ARC 770
Intel I5-12600K
### OS
Windows
### GPU
Intel
### CPU
Intel
### Ollama version
0.5.7 | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8651/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8651/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8673 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8673/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8673/comments | https://api.github.com/repos/ollama/ollama/issues/8673/events | https://github.com/ollama/ollama/pull/8673 | 2,819,364,464 | PR_kwDOJ0Z1Ps6JbG7c | 8,673 | test: byte pair encoding | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | open | false | null | [] | null | 0 | 2025-01-29T21:27:48 | 2025-01-29T23:17:55 | null | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8673",
"html_url": "https://github.com/ollama/ollama/pull/8673",
"diff_url": "https://github.com/ollama/ollama/pull/8673.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8673.patch",
"merged_at": null
} | Adding a basic unit test for the bpe tokenizer.
Tests on the `next` branch are failing, so this test pipeline will also, but the tokenizer tests specifically pass. | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8673/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8673/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5365 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5365/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5365/comments | https://api.github.com/repos/ollama/ollama/issues/5365/events | https://github.com/ollama/ollama/pull/5365 | 2,381,182,795 | PR_kwDOJ0Z1Ps5z7q5T | 5,365 | convert gemma2 | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 1 | 2024-06-28T20:54:43 | 2024-08-21T18:48:46 | 2024-08-21T18:48:43 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5365",
"html_url": "https://github.com/ollama/ollama/pull/5365",
"diff_url": "https://github.com/ollama/ollama/pull/5365.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5365.patch",
"merged_at": "2024-08-21T18:48:43"
} | resolves #6426 | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5365/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5365/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7911 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7911/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7911/comments | https://api.github.com/repos/ollama/ollama/issues/7911/events | https://github.com/ollama/ollama/issues/7911 | 2,712,175,967 | I_kwDOJ0Z1Ps6hqIVf | 7,911 | Inference on an AMD mid-range GPU achieves 1.5 tokens/s, while LM Studio can reach 65 tokens/s | {
"login": "blizzardwj",
"id": 19178835,
"node_id": "MDQ6VXNlcjE5MTc4ODM1",
"avatar_url": "https://avatars.githubusercontent.com/u/19178835?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/blizzardwj",
"html_url": "https://github.com/blizzardwj",
"followers_url": "https://api.github.com/use... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q... | closed | false | null | [] | null | 1 | 2024-12-02T14:42:26 | 2024-12-23T08:02:59 | 2024-12-23T08:02:59 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I'm just testing out Llama 3.1 8B q4_k_m with “tell me a joke". The performance gap between them is pretty huge. I'm using Ollama + Nvidia for work, but for home fun, I went with a 7800XT. I'm wondering if the performance issue is due to my setup or a limitation in Ollama itself.
Both are run... | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7911/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7911/timeline | null | not_planned | false |
https://api.github.com/repos/ollama/ollama/issues/8622 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8622/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8622/comments | https://api.github.com/repos/ollama/ollama/issues/8622/events | https://github.com/ollama/ollama/issues/8622 | 2,814,520,394 | I_kwDOJ0Z1Ps6nwixK | 8,622 | Support for Zero-shot Text Classification Models | {
"login": "BrainSlugs83",
"id": 5217366,
"node_id": "MDQ6VXNlcjUyMTczNjY=",
"avatar_url": "https://avatars.githubusercontent.com/u/5217366?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BrainSlugs83",
"html_url": "https://github.com/BrainSlugs83",
"followers_url": "https://api.github.com... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2025-01-28T03:09:05 | 2025-01-28T03:09:05 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | It would be helpful to developers if ollama supported zero-shot text classification models, such as [`deberta-v3-large-tasksource-nli`](https://huggingface.co/sileod/deberta-v3-large-tasksource-nli) or other offshoots of BERT, which are fairly small models, that allow you do things like pass in a list of categories and... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8622/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8622/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/5999 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5999/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5999/comments | https://api.github.com/repos/ollama/ollama/issues/5999/events | https://github.com/ollama/ollama/pull/5999 | 2,432,982,820 | PR_kwDOJ0Z1Ps52ns7l | 5,999 | fix nil deref in auth.go | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2024-07-26T21:10:57 | 2024-07-26T21:28:36 | 2024-07-26T21:28:34 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5999",
"html_url": "https://github.com/ollama/ollama/pull/5999",
"diff_url": "https://github.com/ollama/ollama/pull/5999.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5999.patch",
"merged_at": "2024-07-26T21:28:34"
} | `server/auth.go` uses `makeRequest` with `nil` `registryOptions` so `registryOptions.CheckRedirect` will dereference `nil` and panic | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5999/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5999/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/623 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/623/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/623/comments | https://api.github.com/repos/ollama/ollama/issues/623/events | https://github.com/ollama/ollama/issues/623 | 1,915,956,307 | I_kwDOJ0Z1Ps5yMyxT | 623 | Mistral model can't be pulled | {
"login": "instanceofme",
"id": 402483,
"node_id": "MDQ6VXNlcjQwMjQ4Mw==",
"avatar_url": "https://avatars.githubusercontent.com/u/402483?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/instanceofme",
"html_url": "https://github.com/instanceofme",
"followers_url": "https://api.github.com/u... | [] | closed | false | null | [] | null | 4 | 2023-09-27T16:39:04 | 2023-09-27T20:21:44 | 2023-09-27T20:21:44 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi,
I saw the [page about mistral](https://ollama.ai/library/mistral) but trying to pull it does not work, is it too soon / some files need to propagate through a CDN?
```
ollama pull mistral
pulling manifest
Error: pull model manifest: model not found
``` | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/623/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/623/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1022 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1022/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1022/comments | https://api.github.com/repos/ollama/ollama/issues/1022/events | https://github.com/ollama/ollama/pull/1022 | 1,980,150,711 | PR_kwDOJ0Z1Ps5evheQ | 1,022 | instead of static number of parameters for each model family, get the real number from the tensors | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-11-06T22:05:55 | 2023-11-09T01:55:47 | 2023-11-09T01:55:47 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1022",
"html_url": "https://github.com/ollama/ollama/pull/1022",
"diff_url": "https://github.com/ollama/ollama/pull/1022.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1022.patch",
"merged_at": "2023-11-09T01:55:46"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1022/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1022/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4903 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4903/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4903/comments | https://api.github.com/repos/ollama/ollama/issues/4903/events | https://github.com/ollama/ollama/issues/4903 | 2,340,234,272 | I_kwDOJ0Z1Ps6LfSQg | 4,903 | Intel/neural-chat-7b-v3 prompts itself | {
"login": "0x2E16CF0F",
"id": 151734472,
"node_id": "U_kgDOCQtIyA",
"avatar_url": "https://avatars.githubusercontent.com/u/151734472?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/0x2E16CF0F",
"html_url": "https://github.com/0x2E16CF0F",
"followers_url": "https://api.github.com/users/0x2... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 0 | 2024-06-07T11:26:53 | 2024-06-07T11:26:53 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
After conversing Intel/neural-chat-7b-v3 for a long time, asking for more details on certain topics, the system starts prompting itself. It both happens in CLI as well as using a WebUI.
I am not sure whether this is caused by ollama or by an issue with Intel's model.
Here is an example. For ... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4903/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4903/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/5061 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5061/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5061/comments | https://api.github.com/repos/ollama/ollama/issues/5061/events | https://github.com/ollama/ollama/issues/5061 | 2,354,789,925 | I_kwDOJ0Z1Ps6MWz4l | 5,061 | Ollama doesn't start in Docker! | {
"login": "samanthacarapathy",
"id": 172900308,
"node_id": "U_kgDOCk4_1A",
"avatar_url": "https://avatars.githubusercontent.com/u/172900308?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/samanthacarapathy",
"html_url": "https://github.com/samanthacarapathy",
"followers_url": "https://api... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q... | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 2 | 2024-06-15T11:57:58 | 2024-07-03T23:26:38 | 2024-07-03T23:26:38 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Hi guys, I'm trying to deploy Ollama to Hugging Face space using the Docker SDK. I've added this Git repo to Hugging Face since spaces are hosted on the Hugging Face Hub as Git repo and using this Dockerfile without any changes but, it keeps building. I could never get ollama to successfully run... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5061/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5061/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6837 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6837/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6837/comments | https://api.github.com/repos/ollama/ollama/issues/6837/events | https://github.com/ollama/ollama/pull/6837 | 2,530,549,422 | PR_kwDOJ0Z1Ps57uzG- | 6,837 | Add support for the ppc64le architecture | {
"login": "mkumatag",
"id": 12646029,
"node_id": "MDQ6VXNlcjEyNjQ2MDI5",
"avatar_url": "https://avatars.githubusercontent.com/u/12646029?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mkumatag",
"html_url": "https://github.com/mkumatag",
"followers_url": "https://api.github.com/users/mku... | [] | closed | false | null | [] | null | 5 | 2024-09-17T09:21:22 | 2024-11-21T10:18:59 | 2024-11-21T10:18:59 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/6837",
"html_url": "https://github.com/ollama/ollama/pull/6837",
"diff_url": "https://github.com/ollama/ollama/pull/6837.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6837.patch",
"merged_at": null
} | This PR will add the support for [IBM POWER Architecture](https://en.wikipedia.org/wiki/IBM_POWER_architecture). | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6837/reactions",
"total_count": 7,
"+1": 7,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6837/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5607 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5607/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5607/comments | https://api.github.com/repos/ollama/ollama/issues/5607/events | https://github.com/ollama/ollama/pull/5607 | 2,401,363,173 | PR_kwDOJ0Z1Ps51ADBS | 5,607 | Bump ROCm on windows to 6.1.2 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-07-10T18:02:49 | 2024-07-10T19:47:13 | 2024-07-10T19:47:10 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5607",
"html_url": "https://github.com/ollama/ollama/pull/5607",
"diff_url": "https://github.com/ollama/ollama/pull/5607.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5607.patch",
"merged_at": "2024-07-10T19:47:10"
} | This also adjusts our algorithm to favor our bundled ROCm. I've confirmed VRAM reporting still doesn't work properly so we can't yet enable concurrency by default.
Fixes #5599 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5607/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5607/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1733 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1733/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1733/comments | https://api.github.com/repos/ollama/ollama/issues/1733/events | https://github.com/ollama/ollama/issues/1733 | 2,058,025,449 | I_kwDOJ0Z1Ps56qvnp | 1,733 | Feature request: improve install.sh and release binaries for CPU instructions | {
"login": "oafish",
"id": 1235498,
"node_id": "MDQ6VXNlcjEyMzU0OTg=",
"avatar_url": "https://avatars.githubusercontent.com/u/1235498?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/oafish",
"html_url": "https://github.com/oafish",
"followers_url": "https://api.github.com/users/oafish/foll... | [] | closed | false | null | [] | null | 1 | 2023-12-28T05:58:27 | 2023-12-28T20:42:46 | 2023-12-28T20:42:45 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Is it feasible for precompile multiple binaries for AVX1, AVX2, AVX512 and Openblas just like https://github.com/ggerganov/llama.cpp/releases
The install.sh can detect the platform not only CPU architecture but also the grep cpuinfo to download most suitable binaries.
I hope it is an elegant solution for https://git... | {
"login": "technovangelist",
"id": 633681,
"node_id": "MDQ6VXNlcjYzMzY4MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/technovangelist",
"html_url": "https://github.com/technovangelist",
"followers_url": "https://api.git... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1733/reactions",
"total_count": 2,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 2,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1733/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5529 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5529/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5529/comments | https://api.github.com/repos/ollama/ollama/issues/5529/events | https://github.com/ollama/ollama/issues/5529 | 2,394,049,342 | I_kwDOJ0Z1Ps6Osks- | 5,529 | Support glm3 and glm4 | {
"login": "Forevery1",
"id": 19872771,
"node_id": "MDQ6VXNlcjE5ODcyNzcx",
"avatar_url": "https://avatars.githubusercontent.com/u/19872771?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Forevery1",
"html_url": "https://github.com/Forevery1",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 8 | 2024-07-07T13:08:29 | 2024-10-17T10:22:20 | 2024-07-09T06:15:09 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Currently, llama.cpp has merged a PR for this model, which means that llama.cpp already supports gml3 and glm4. We hope to add support for this model @jmorganca
https://github.com/ggerganov/llama.cpp/pull/8031 | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5529/reactions",
"total_count": 10,
"+1": 10,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5529/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7603 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7603/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7603/comments | https://api.github.com/repos/ollama/ollama/issues/7603/events | https://github.com/ollama/ollama/issues/7603 | 2,647,578,686 | I_kwDOJ0Z1Ps6dztg- | 7,603 | API: JSON response is empty | {
"login": "boessu",
"id": 2807976,
"node_id": "MDQ6VXNlcjI4MDc5NzY=",
"avatar_url": "https://avatars.githubusercontent.com/u/2807976?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/boessu",
"html_url": "https://github.com/boessu",
"followers_url": "https://api.github.com/users/boessu/foll... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2024-11-10T19:47:12 | 2024-11-11T18:23:04 | 2024-11-11T18:23:04 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Hello
I'm not sure if that is a model specific problem, but sometimes I get a (somewhat) empty JSON response. It doesn't happen if the format is not set to json.
Example:
```
C:\Users\Boessu>curl http://localhost:11434/api/generate -d "{\"model\":\"llama3.2\",\"prompt\":\"Why is the sky ... | {
"login": "boessu",
"id": 2807976,
"node_id": "MDQ6VXNlcjI4MDc5NzY=",
"avatar_url": "https://avatars.githubusercontent.com/u/2807976?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/boessu",
"html_url": "https://github.com/boessu",
"followers_url": "https://api.github.com/users/boessu/foll... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7603/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7603/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3893 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3893/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3893/comments | https://api.github.com/repos/ollama/ollama/issues/3893/events | https://github.com/ollama/ollama/pull/3893 | 2,262,282,515 | PR_kwDOJ0Z1Ps5tp-Ae | 3,893 | add param count tag to latest push by default | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 0 | 2024-04-24T22:22:34 | 2024-04-24T22:57:35 | 2024-04-24T22:57:35 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3893",
"html_url": "https://github.com/ollama/ollama/pull/3893",
"diff_url": "https://github.com/ollama/ollama/pull/3893.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3893.patch",
"merged_at": null
} | - add a tag for the model parameter count by default when a new latest model is pushed
- allow skipping this default with an environment flag
This change automatically adds an additional parameter count flag when a new `latest` tag of a model is pushed. This should help users to conform to our expected tagging form... | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3893/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3893/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2603 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2603/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2603/comments | https://api.github.com/repos/ollama/ollama/issues/2603/events | https://github.com/ollama/ollama/issues/2603 | 2,143,428,045 | I_kwDOJ0Z1Ps5_wh3N | 2,603 | Issue with anything-llm in connection with the port binding to an IP in a virtual docker network | {
"login": "fukuro-kun",
"id": 29835293,
"node_id": "MDQ6VXNlcjI5ODM1Mjkz",
"avatar_url": "https://avatars.githubusercontent.com/u/29835293?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/fukuro-kun",
"html_url": "https://github.com/fukuro-kun",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 1 | 2024-02-20T02:49:31 | 2024-02-20T18:57:46 | 2024-02-20T18:57:46 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### Backgound
When I set up the Docker container of https://github.com/Mintplex-Labs/anything-llm, and started a conversation, I received the following error:
> llama:streaming - could not stream chat. Error: connect ECONNREFUSED 172.17.0.1:11434
although `OLLAMA_BASE_PATH='http://host.docker.internal:11434` ... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2603/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2603/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/188 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/188/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/188/comments | https://api.github.com/repos/ollama/ollama/issues/188/events | https://github.com/ollama/ollama/issues/188 | 1,817,723,365 | I_kwDOJ0Z1Ps5sWEHl | 188 | Unable to compile on windows using standard go installation | {
"login": "FairyTail2000",
"id": 22645621,
"node_id": "MDQ6VXNlcjIyNjQ1NjIx",
"avatar_url": "https://avatars.githubusercontent.com/u/22645621?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/FairyTail2000",
"html_url": "https://github.com/FairyTail2000",
"followers_url": "https://api.githu... | [] | closed | false | null | [] | null | 21 | 2023-07-24T06:22:48 | 2023-09-15T16:12:51 | 2023-09-07T13:27:00 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Steps I followed:
- I installed the newest GoLang using winget
- I cloned the repro
- I executed `go build .`
- After initial library download build fails with the error message:
> \# github.com/jmorganca/ollama/server
> server\routes.go:54:20: undefined: llama.New
- I then checked out tags/v0.0.11
- Same er... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/188/reactions",
"total_count": 2,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/ollama/ollama/issues/188/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1161 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1161/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1161/comments | https://api.github.com/repos/ollama/ollama/issues/1161/events | https://github.com/ollama/ollama/pull/1161 | 1,998,052,342 | PR_kwDOJ0Z1Ps5fsbzA | 1,161 | placeholder environment variables | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 1 | 2023-11-17T00:55:55 | 2023-11-17T22:50:15 | 2023-11-17T22:45:39 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1161",
"html_url": "https://github.com/ollama/ollama/pull/1161",
"diff_url": "https://github.com/ollama/ollama/pull/1161.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1161.patch",
"merged_at": "2023-11-17T22:45:39"
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1161/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1161/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6111 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6111/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6111/comments | https://api.github.com/repos/ollama/ollama/issues/6111/events | https://github.com/ollama/ollama/issues/6111 | 2,441,176,030 | I_kwDOJ0Z1Ps6RgWPe | 6,111 | Request: add octopus-v4 | {
"login": "mak448a",
"id": 94062293,
"node_id": "U_kgDOBZtG1Q",
"avatar_url": "https://avatars.githubusercontent.com/u/94062293?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mak448a",
"html_url": "https://github.com/mak448a",
"followers_url": "https://api.github.com/users/mak448a/follow... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | open | false | null | [] | null | 3 | 2024-08-01T00:25:58 | 2024-09-12T21:10:17 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Could you add octopus v4? (I haven't tested for malware, but it looks somewhat legit)
https://huggingface.co/NexaAIDev/Octopus-v4
https://huggingface.co/NexaAIDev/octopus-v4-gguf | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6111/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6111/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/6465 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6465/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6465/comments | https://api.github.com/repos/ollama/ollama/issues/6465/events | https://github.com/ollama/ollama/pull/6465 | 2,481,323,837 | PR_kwDOJ0Z1Ps55JzVR | 6,465 | Adding 'Ollama App' as community integrations | {
"login": "JHubi1",
"id": 61345690,
"node_id": "MDQ6VXNlcjYxMzQ1Njkw",
"avatar_url": "https://avatars.githubusercontent.com/u/61345690?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/JHubi1",
"html_url": "https://github.com/JHubi1",
"followers_url": "https://api.github.com/users/JHubi1/fo... | [] | closed | false | null | [] | null | 2 | 2024-08-22T17:08:06 | 2024-10-15T16:57:33 | 2024-10-15T16:57:33 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/6465",
"html_url": "https://github.com/ollama/ollama/pull/6465",
"diff_url": "https://github.com/ollama/ollama/pull/6465.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6465.patch",
"merged_at": "2024-10-15T16:57:33"
} | A little while back, I created an app to interact with Ollama from your phone called 'Ollama App'. I know, very creative, but the name stuck. I'd be very happy if it could be added to the integrations list.
At the moment, I only recommend using the mobile version, it's the most stable one, and the only one with a st... | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6465/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6465/timeline | null | null | true |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.