url stringlengths 51 54 | repository_url stringclasses 1
value | labels_url stringlengths 65 68 | comments_url stringlengths 60 63 | events_url stringlengths 58 61 | html_url stringlengths 39 44 | id int64 1.78B 2.82B | node_id stringlengths 18 19 | number int64 1 8.69k | title stringlengths 1 382 | user dict | labels listlengths 0 5 | state stringclasses 2
values | locked bool 1
class | assignee dict | assignees listlengths 0 2 | milestone null | comments int64 0 323 | created_at timestamp[s] | updated_at timestamp[s] | closed_at timestamp[s] | author_association stringclasses 4
values | sub_issues_summary dict | active_lock_reason null | draft bool 2
classes | pull_request dict | body stringlengths 2 118k ⌀ | closed_by dict | reactions dict | timeline_url stringlengths 60 63 | performed_via_github_app null | state_reason stringclasses 4
values | is_pull_request bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/ollama/ollama/issues/3963 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3963/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3963/comments | https://api.github.com/repos/ollama/ollama/issues/3963/events | https://github.com/ollama/ollama/pull/3963 | 2,266,554,114 | PR_kwDOJ0Z1Ps5t4mC3 | 3,963 | windows: init `cmakeTargets` in `init_vars` | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 1 | 2024-04-26T21:57:05 | 2024-04-26T22:40:31 | 2024-04-26T22:40:31 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3963",
"html_url": "https://github.com/ollama/ollama/pull/3963",
"diff_url": "https://github.com/ollama/ollama/pull/3963.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3963.patch",
"merged_at": null
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3963/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3963/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/952 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/952/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/952/comments | https://api.github.com/repos/ollama/ollama/issues/952/events | https://github.com/ollama/ollama/pull/952 | 1,969,294,048 | PR_kwDOJ0Z1Ps5eK5xF | 952 | support raw generation requests | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 0 | 2023-10-30T22:01:26 | 2023-11-08T22:05:03 | 2023-11-08T22:05:02 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/952",
"html_url": "https://github.com/ollama/ollama/pull/952",
"diff_url": "https://github.com/ollama/ollama/pull/952.diff",
"patch_url": "https://github.com/ollama/ollama/pull/952.patch",
"merged_at": "2023-11-08T22:05:02"
} | - add the optional `raw` generate request parameter to bypass prompt formatting and response context
Add a `raw` parameter to `/generate` requests that allow directly specifying the prompt without the Ollama server applying additional formatting.
```bash
curl -X "POST" -d '{"model":"mistral", "prompt": "[INST] hi ... | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/952/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/952/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1082 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1082/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1082/comments | https://api.github.com/repos/ollama/ollama/issues/1082/events | https://github.com/ollama/ollama/pull/1082 | 1,988,676,727 | PR_kwDOJ0Z1Ps5fMqVc | 1,082 | Add support for Multimodel models | {
"login": "mattapperson",
"id": 192433,
"node_id": "MDQ6VXNlcjE5MjQzMw==",
"avatar_url": "https://avatars.githubusercontent.com/u/192433?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mattapperson",
"html_url": "https://github.com/mattapperson",
"followers_url": "https://api.github.com/u... | [] | closed | false | null | [] | null | 12 | 2023-11-11T02:57:28 | 2023-11-22T00:22:46 | 2023-11-22T00:22:46 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1082",
"html_url": "https://github.com/ollama/ollama/pull/1082",
"diff_url": "https://github.com/ollama/ollama/pull/1082.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1082.patch",
"merged_at": null
} | Interactive cli usage:
```bash
/set image add <image id int> <path to image file>
Please tell me what text is in this photo [img-<image id int>]
```
For the API I added support for the `image_data` prop with the type of `[]{id: int, data: string(base64)}`for the generate endpoint.
To support this, modelfile... | {
"login": "mattapperson",
"id": 192433,
"node_id": "MDQ6VXNlcjE5MjQzMw==",
"avatar_url": "https://avatars.githubusercontent.com/u/192433?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mattapperson",
"html_url": "https://github.com/mattapperson",
"followers_url": "https://api.github.com/u... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1082/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 1,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1082/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8141 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8141/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8141/comments | https://api.github.com/repos/ollama/ollama/issues/8141/events | https://github.com/ollama/ollama/issues/8141 | 2,745,757,781 | I_kwDOJ0Z1Ps6jqPBV | 8,141 | command r plus model update | {
"login": "olumolu",
"id": 162728301,
"node_id": "U_kgDOCbMJbQ",
"avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/olumolu",
"html_url": "https://github.com/olumolu",
"followers_url": "https://api.github.com/users/olumolu/foll... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 1 | 2024-12-17T18:48:33 | 2024-12-23T08:13:29 | 2024-12-23T08:13:29 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | > https://docs.cohere.com/docs/command-r-plus
They already have updated there model can we have the updated new model in olllama thanks | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8141/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8141/timeline | null | not_planned | false |
https://api.github.com/repos/ollama/ollama/issues/5864 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5864/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5864/comments | https://api.github.com/repos/ollama/ollama/issues/5864/events | https://github.com/ollama/ollama/pull/5864 | 2,423,919,879 | PR_kwDOJ0Z1Ps52Je6A | 5,864 | Bump Go patch version | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-07-22T23:17:38 | 2024-07-22T23:34:22 | 2024-07-22T23:34:19 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5864",
"html_url": "https://github.com/ollama/ollama/pull/5864",
"diff_url": "https://github.com/ollama/ollama/pull/5864.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5864.patch",
"merged_at": "2024-07-22T23:34:18"
} | Fixes #5774 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5864/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5864/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/321 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/321/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/321/comments | https://api.github.com/repos/ollama/ollama/issues/321/events | https://github.com/ollama/ollama/pull/321 | 1,846,003,602 | PR_kwDOJ0Z1Ps5XruCs | 321 | length check for parameters | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-08-10T23:09:25 | 2023-08-10T23:23:11 | 2023-08-10T23:23:10 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/321",
"html_url": "https://github.com/ollama/ollama/pull/321",
"diff_url": "https://github.com/ollama/ollama/pull/321.diff",
"patch_url": "https://github.com/ollama/ollama/pull/321.patch",
"merged_at": "2023-08-10T23:23:10"
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/321/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/321/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1392 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1392/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1392/comments | https://api.github.com/repos/ollama/ollama/issues/1392/events | https://github.com/ollama/ollama/pull/1392 | 2,026,923,082 | PR_kwDOJ0Z1Ps5hN7DP | 1,392 | chat api endpoint | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [] | closed | false | null | [] | null | 3 | 2023-12-05T18:57:45 | 2023-12-11T07:33:21 | 2023-12-05T19:57:33 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1392",
"html_url": "https://github.com/ollama/ollama/pull/1392",
"diff_url": "https://github.com/ollama/ollama/pull/1392.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1392.patch",
"merged_at": "2023-12-05T19:57:33"
} | - add a new `/api/chat` API endpoint that takes an array of `message` objects. This endpoint is an alternative to `/api/generate`.
- deprecate generation context and template, but continue to support them
- rebuild chat content from messages
This changes adds a `/api/chat` endpoint to the API which takes an array ... | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1392/reactions",
"total_count": 2,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 2,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1392/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1940 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1940/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1940/comments | https://api.github.com/repos/ollama/ollama/issues/1940/events | https://github.com/ollama/ollama/issues/1940 | 2,077,872,259 | I_kwDOJ0Z1Ps572dCD | 1,940 | `CUDA error 100` after detecting GPU libraries on system | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg... | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 10 | 2024-01-12T01:13:13 | 2024-02-19T19:50:24 | 2024-02-19T19:50:24 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | It seems that upon detecting an Nvidia card, `ollama` may error with `CUDA error 100`:
```
Jan 11 15:37:50 LR9135SQP ollama[5616]: 2024/01/11 15:37:50 gpu.go:88: Detecting GPU type
Jan 11 15:37:50 LR9135SQP ollama[5616]: 2024/01/11 15:37:50 gpu.go:203: Searching for GPU management library libnvidia-ml.so
Jan 11 1... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1940/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1940/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6839 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6839/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6839/comments | https://api.github.com/repos/ollama/ollama/issues/6839/events | https://github.com/ollama/ollama/issues/6839 | 2,531,251,465 | I_kwDOJ0Z1Ps6W39UJ | 6,839 | ollama request llama3.1 fail. | {
"login": "microbitcswcss",
"id": 129246500,
"node_id": "U_kgDOB7QlJA",
"avatar_url": "https://avatars.githubusercontent.com/u/129246500?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/microbitcswcss",
"html_url": "https://github.com/microbitcswcss",
"followers_url": "https://api.github.c... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 5 | 2024-09-17T14:00:01 | 2024-10-28T07:39:53 | 2024-10-28T07:39:52 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
like title i cant request llama3.1
on windows 10
it can be used one month ago,
cmd.exe:
C:\Users\123>ollama run llama3.1
2024/09/17 21:53:18 config.go:45: WARN invalid port, using default port="\\Users\\123\\AppData\\Local\\Programs\\Ollama\\ollama app.exe" default=11434
2024/09/17 21:53:1... | {
"login": "microbitcswcss",
"id": 129246500,
"node_id": "U_kgDOB7QlJA",
"avatar_url": "https://avatars.githubusercontent.com/u/129246500?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/microbitcswcss",
"html_url": "https://github.com/microbitcswcss",
"followers_url": "https://api.github.c... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6839/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6839/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/220 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/220/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/220/comments | https://api.github.com/repos/ollama/ollama/issues/220/events | https://github.com/ollama/ollama/issues/220 | 1,822,783,975 | I_kwDOJ0Z1Ps5spXnn | 220 | Sending input with a + crashes | {
"login": "benr75",
"id": 661,
"node_id": "MDQ6VXNlcjY2MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/661?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/benr75",
"html_url": "https://github.com/benr75",
"followers_url": "https://api.github.com/users/benr75/followers",
"f... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 4 | 2023-07-26T16:37:12 | 2023-08-04T14:28:22 | 2023-08-04T14:28:22 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Maybe an url escaping issue? I didn't dig much deeper but wanted to report.
```
$ ollama run llama2:13b
>>> I cannot send input with a + in it
Error: Post "http://127.0.0.1:11434/api/generate": EOF
```
| {
"login": "benr75",
"id": 661,
"node_id": "MDQ6VXNlcjY2MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/661?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/benr75",
"html_url": "https://github.com/benr75",
"followers_url": "https://api.github.com/users/benr75/followers",
"f... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/220/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/220/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8248 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8248/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8248/comments | https://api.github.com/repos/ollama/ollama/issues/8248/events | https://github.com/ollama/ollama/issues/8248 | 2,759,846,975 | I_kwDOJ0Z1Ps6kf-w_ | 8,248 | Unexpected Connection Closure and GPU Memory Not Releasing | {
"login": "BronyaKaslana06",
"id": 97924591,
"node_id": "U_kgDOBdY17w",
"avatar_url": "https://avatars.githubusercontent.com/u/97924591?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BronyaKaslana06",
"html_url": "https://github.com/BronyaKaslana06",
"followers_url": "https://api.github.... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 7 | 2024-12-26T14:40:51 | 2025-01-01T02:27:55 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
### Problem Description
I am using Ubuntu 22.04 and making network requests to a local Ollama service with Python to run a series of models sequentially. After each model runs, it is unloaded using the following Python code. However, the task encounters an issue after reaching a certain poi... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8248/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8248/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/7455 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7455/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7455/comments | https://api.github.com/repos/ollama/ollama/issues/7455/events | https://github.com/ollama/ollama/pull/7455 | 2,627,779,798 | PR_kwDOJ0Z1Ps6AkeyU | 7,455 | Add basic mllama integration tests | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-10-31T20:37:06 | 2024-11-01T00:25:51 | 2024-11-01T00:25:48 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7455",
"html_url": "https://github.com/ollama/ollama/pull/7455",
"diff_url": "https://github.com/ollama/ollama/pull/7455.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7455.patch",
"merged_at": "2024-11-01T00:25:48"
} | Draft until I run it against a few more systems and make sure the timeout values are reasonable | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7455/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7455/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8379 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8379/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8379/comments | https://api.github.com/repos/ollama/ollama/issues/8379/events | https://github.com/ollama/ollama/issues/8379 | 2,781,273,717 | I_kwDOJ0Z1Ps6lxt51 | 8,379 | Inquiry About Adding vikhyatk/moondream2 to Ollama | {
"login": "Saber120",
"id": 108297159,
"node_id": "U_kgDOBnR7xw",
"avatar_url": "https://avatars.githubusercontent.com/u/108297159?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Saber120",
"html_url": "https://github.com/Saber120",
"followers_url": "https://api.github.com/users/Saber120/... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 2 | 2025-01-10T22:03:19 | 2025-01-13T19:29:39 | 2025-01-13T19:29:38 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi Ollama team, I was wondering if you plan to add the model vikhyatk/moondream2, which was just released yesterday, to your platform | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8379/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8379/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1753 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1753/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1753/comments | https://api.github.com/repos/ollama/ollama/issues/1753/events | https://github.com/ollama/ollama/issues/1753 | 2,061,198,535 | I_kwDOJ0Z1Ps5622TH | 1,753 | [enhancement] Support FROM gpt-4:0314 | {
"login": "nathanleclaire",
"id": 1476820,
"node_id": "MDQ6VXNlcjE0NzY4MjA=",
"avatar_url": "https://avatars.githubusercontent.com/u/1476820?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nathanleclaire",
"html_url": "https://github.com/nathanleclaire",
"followers_url": "https://api.gith... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2024-01-01T00:11:55 | 2024-01-02T11:26:07 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | It would be nice if you could have a `Modelfile` that had an OpenAI model as a base. Then Ollama inference would call the OpenAI APIs instead of local inference with the built parameters. I know it seems like maybe it sorta defeats the point but I really like having the "Ollama facade" available. I use the March versio... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1753/reactions",
"total_count": 5,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 5,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1753/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/226 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/226/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/226/comments | https://api.github.com/repos/ollama/ollama/issues/226/events | https://github.com/ollama/ollama/pull/226 | 1,824,955,065 | PR_kwDOJ0Z1Ps5Wk6Px | 226 | refactor scan multiline for reuse | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-07-27T18:32:03 | 2023-07-27T18:45:45 | 2023-07-27T18:45:41 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/226",
"html_url": "https://github.com/ollama/ollama/pull/226",
"diff_url": "https://github.com/ollama/ollama/pull/226.diff",
"patch_url": "https://github.com/ollama/ollama/pull/226.patch",
"merged_at": "2023-07-27T18:45:41"
} | It's not obvious values are ingested verbatim when not using multiline so `"` are included in the template. Instead, ingest the value inside the quotes | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/226/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/226/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1866 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1866/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1866/comments | https://api.github.com/repos/ollama/ollama/issues/1866/events | https://github.com/ollama/ollama/issues/1866 | 2,072,359,110 | I_kwDOJ0Z1Ps57hbDG | 1,866 | Where is the model file stored? | {
"login": "ParisNeo",
"id": 827993,
"node_id": "MDQ6VXNlcjgyNzk5Mw==",
"avatar_url": "https://avatars.githubusercontent.com/u/827993?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParisNeo",
"html_url": "https://github.com/ParisNeo",
"followers_url": "https://api.github.com/users/ParisNe... | [] | closed | false | null | [] | null | 4 | 2024-01-09T13:19:37 | 2024-01-11T16:49:16 | 2024-01-09T17:30:51 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi there, I need to make a modification to the model file. Can you please tell me where do you store the model file?
Best regards | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1866/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1866/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5531 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5531/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5531/comments | https://api.github.com/repos/ollama/ollama/issues/5531/events | https://github.com/ollama/ollama/issues/5531 | 2,394,121,820 | I_kwDOJ0Z1Ps6Os2Zc | 5,531 | `ollama create --quantize` does not show proper error if quantizing an unsupported model architecture | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "joshyan1",
"id": 76125168,
"node_id": "MDQ6VXNlcjc2MTI1MTY4",
"avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joshyan1",
"html_url": "https://github.com/joshyan1",
"followers_url": "https://api.github.com/users/jos... | [
{
"login": "joshyan1",
"id": 76125168,
"node_id": "MDQ6VXNlcjc2MTI1MTY4",
"avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joshyan1",
"html_url": "https://github.com/joshyan1",
"followers_url": "https://api.... | null | 0 | 2024-07-07T16:27:14 | 2024-07-12T00:24:30 | 2024-07-12T00:24:30 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
When quantizing a model with `ollama create -q`, unknown model architecture errors are not properly bubbled up. E.g. from the logs:
```
llama_model_loader: - type f32: 121 tensors
llama_model_loader: - type f16: 162 tensors
llama_model_quantize: failed to quantize: unknown model archit... | {
"login": "joshyan1",
"id": 76125168,
"node_id": "MDQ6VXNlcjc2MTI1MTY4",
"avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joshyan1",
"html_url": "https://github.com/joshyan1",
"followers_url": "https://api.github.com/users/jos... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5531/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5531/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6806 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6806/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6806/comments | https://api.github.com/repos/ollama/ollama/issues/6806/events | https://github.com/ollama/ollama/issues/6806 | 2,526,589,141 | I_kwDOJ0Z1Ps6WmLDV | 6,806 | slow | {
"login": "ayttop",
"id": 178673810,
"node_id": "U_kgDOCqZYkg",
"avatar_url": "https://avatars.githubusercontent.com/u/178673810?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ayttop",
"html_url": "https://github.com/ayttop",
"followers_url": "https://api.github.com/users/ayttop/follower... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q... | closed | false | null | [] | null | 9 | 2024-09-14T20:03:00 | 2024-09-25T21:08:43 | 2024-09-25T21:08:43 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
The program is very slow
3.10 ollama
### OS
Windows
### GPU
Intel
### CPU
Intel
### Ollama version
3.10 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6806/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6806/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1471 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1471/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1471/comments | https://api.github.com/repos/ollama/ollama/issues/1471/events | https://github.com/ollama/ollama/issues/1471 | 2,036,377,833 | I_kwDOJ0Z1Ps55YKjp | 1,471 | Publishing model fails with "Error: unable to push ..." | {
"login": "stephenwithav",
"id": 54563,
"node_id": "MDQ6VXNlcjU0NTYz",
"avatar_url": "https://avatars.githubusercontent.com/u/54563?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/stephenwithav",
"html_url": "https://github.com/stephenwithav",
"followers_url": "https://api.github.com/user... | [] | closed | false | null | [] | null | 5 | 2023-12-11T19:25:53 | 2024-11-03T01:37:33 | 2024-01-20T00:25:23 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I followed the [import doc](https://github.com/jmorganca/ollama/blob/7a1b37ac64f0fb0585e279a0a840707843511ed3/docs/import.md?plain=1#L108) steps, but I'm getting the error in the title.
I've:
1. Created an account.
2. Copied my `id_ed25519.pub` file to `~/.ollama/` and `/usr/share/ollama/.ollama`.
3. Added the ... | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1471/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1471/timeline | null | not_planned | false |
https://api.github.com/repos/ollama/ollama/issues/3923 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3923/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3923/comments | https://api.github.com/repos/ollama/ollama/issues/3923/events | https://github.com/ollama/ollama/pull/3923 | 2,264,567,354 | PR_kwDOJ0Z1Ps5tx0Ne | 3,923 | precalculate output tensor memory for metal and mmap | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2024-04-25T22:00:38 | 2024-04-25T23:34:18 | 2024-04-25T23:34:17 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3923",
"html_url": "https://github.com/ollama/ollama/pull/3923",
"diff_url": "https://github.com/ollama/ollama/pull/3923.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3923.patch",
"merged_at": "2024-04-25T23:34:17"
} | on metal with mmap, the output tensors are always allocated even if the offloaded layers < total layers + 1. other backends are unaffected | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3923/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3923/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4428 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4428/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4428/comments | https://api.github.com/repos/ollama/ollama/issues/4428/events | https://github.com/ollama/ollama/issues/4428 | 2,295,167,222 | I_kwDOJ0Z1Ps6IzXj2 | 4,428 | Ollama Api not working after adding format:"json" | {
"login": "usmandilmeer",
"id": 51738693,
"node_id": "MDQ6VXNlcjUxNzM4Njkz",
"avatar_url": "https://avatars.githubusercontent.com/u/51738693?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/usmandilmeer",
"html_url": "https://github.com/usmandilmeer",
"followers_url": "https://api.github.c... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg... | closed | false | null | [] | null | 2 | 2024-05-14T11:23:41 | 2024-05-14T21:46:32 | 2024-05-14T21:46:32 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Ollama api is giving empty response if i add add format as json

### OS
Windows
### GPU
AMD
### CPU
Intel
### Ollama version
0.1.37 | {
"login": "usmandilmeer",
"id": 51738693,
"node_id": "MDQ6VXNlcjUxNzM4Njkz",
"avatar_url": "https://avatars.githubusercontent.com/u/51738693?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/usmandilmeer",
"html_url": "https://github.com/usmandilmeer",
"followers_url": "https://api.github.c... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4428/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4428/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5588 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5588/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5588/comments | https://api.github.com/repos/ollama/ollama/issues/5588/events | https://github.com/ollama/ollama/pull/5588 | 2,399,649,431 | PR_kwDOJ0Z1Ps506G4S | 5,588 | Remove `GGML_CUDA_FORCE_MMQ=on` from build | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 1 | 2024-07-10T03:08:09 | 2024-07-10T20:17:14 | 2024-07-10T20:17:13 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5588",
"html_url": "https://github.com/ollama/ollama/pull/5588",
"diff_url": "https://github.com/ollama/ollama/pull/5588.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5588.patch",
"merged_at": "2024-07-10T20:17:13"
} | Removing this build flag seems to cause mmq to be decided at runtime vs always forcing it on. This stops some cuda calls from failing on V100 and other CC 7.0 cards. This may also help with [RDNA3 cards](https://github.com/ggerganov/llama.cpp/blob/a59f8fdc85e1119d470d8766e29617962549d993/docs/build.md?plain=1#L175)
... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5588/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5588/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1614 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1614/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1614/comments | https://api.github.com/repos/ollama/ollama/issues/1614/events | https://github.com/ollama/ollama/pull/1614 | 2,049,321,465 | PR_kwDOJ0Z1Ps5iZ9OH | 1,614 | fix: set template without triple quotes | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 1 | 2023-12-19T19:37:53 | 2024-01-09T17:36:26 | 2024-01-09T17:36:25 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1614",
"html_url": "https://github.com/ollama/ollama/pull/1614",
"diff_url": "https://github.com/ollama/ollama/pull/1614.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1614.patch",
"merged_at": "2024-01-09T17:36:25"
} | this changes updates `/set` to better handle multiline strings. `/set` now correctly sets template or system without using triple quotes
```
>>> /set template {{ .Prompt }}
Set prompt template.
```
additionally, use a strings.Builder instead of concatenating string values for prompt building
```
>>> """hel... | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1614/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1614/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/473 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/473/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/473/comments | https://api.github.com/repos/ollama/ollama/issues/473/events | https://github.com/ollama/ollama/pull/473 | 1,882,917,488 | PR_kwDOJ0Z1Ps5Zn-ay | 473 | create manifests directory | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-09-06T00:12:07 | 2023-09-06T00:37:42 | 2023-09-06T00:37:41 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/473",
"html_url": "https://github.com/ollama/ollama/pull/473",
"diff_url": "https://github.com/ollama/ollama/pull/473.diff",
"patch_url": "https://github.com/ollama/ollama/pull/473.patch",
"merged_at": "2023-09-06T00:37:41"
} | `ollama list` on a brand new install will panic because the manifests directory doesn't exist | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/473/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/473/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1070 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1070/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1070/comments | https://api.github.com/repos/ollama/ollama/issues/1070/events | https://github.com/ollama/ollama/issues/1070 | 1,986,849,875 | I_kwDOJ0Z1Ps52bOxT | 1,070 | API for models on `ollama.com` | {
"login": "danemadsen",
"id": 11537699,
"node_id": "MDQ6VXNlcjExNTM3Njk5",
"avatar_url": "https://avatars.githubusercontent.com/u/11537699?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/danemadsen",
"html_url": "https://github.com/danemadsen",
"followers_url": "https://api.github.com/use... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 6573197867,
"node_id": ... | closed | false | null | [] | null | 12 | 2023-11-10T04:23:05 | 2024-11-21T17:25:46 | 2024-11-21T17:25:46 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi I was wondering if you could add a way to either search for, or get a list of models available to pull off ollama.ai.
Currently the https://ollama.ai/library endpoint serves model information as HTML, it would be better if it was served as JSON so it could be better integrated into software | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1070/reactions",
"total_count": 18,
"+1": 18,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1070/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6120 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6120/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6120/comments | https://api.github.com/repos/ollama/ollama/issues/6120/events | https://github.com/ollama/ollama/issues/6120 | 2,442,618,478 | I_kwDOJ0Z1Ps6Rl2Zu | 6,120 | Outputting the response leaves a bunch of control characters. | {
"login": "jjasghar",
"id": 810824,
"node_id": "MDQ6VXNlcjgxMDgyNA==",
"avatar_url": "https://avatars.githubusercontent.com/u/810824?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jjasghar",
"html_url": "https://github.com/jjasghar",
"followers_url": "https://api.github.com/users/jjasgha... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 9 | 2024-08-01T14:25:11 | 2024-11-22T16:04:55 | 2024-11-22T16:04:55 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I get a bunch of control characters. How can I get a clean output?
_Originally posted by @pongnguy in https://github.com/ollama/ollama/issues/1710#issuecomment-2142294830_
It seems when you run something like: `ollama run granite-code:20b show me a python fuction that does fizzbuzz > ... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6120/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6120/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8127 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8127/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8127/comments | https://api.github.com/repos/ollama/ollama/issues/8127/events | https://github.com/ollama/ollama/pull/8127 | 2,743,777,670 | PR_kwDOJ0Z1Ps6FboqQ | 8,127 | llm: loosen format check to default to no format | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 0 | 2024-12-17T02:21:37 | 2024-12-17T04:25:55 | 2024-12-17T02:45:47 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8127",
"html_url": "https://github.com/ollama/ollama/pull/8127",
"diff_url": "https://github.com/ollama/ollama/pull/8127.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8127.patch",
"merged_at": "2024-12-17T02:45:47"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8127/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8127/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1925 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1925/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1925/comments | https://api.github.com/repos/ollama/ollama/issues/1925/events | https://github.com/ollama/ollama/issues/1925 | 2,077,046,512 | I_kwDOJ0Z1Ps57zTbw | 1,925 | Ollama is running in background in MacOS | {
"login": "MagzhanUnited",
"id": 123943870,
"node_id": "U_kgDOB2M7vg",
"avatar_url": "https://avatars.githubusercontent.com/u/123943870?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/MagzhanUnited",
"html_url": "https://github.com/MagzhanUnited",
"followers_url": "https://api.github.com/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 11 | 2024-01-11T16:04:05 | 2025-01-29T08:51:22 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Even if I exit the ollama app I can see the ollama among run processes | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1925/reactions",
"total_count": 3,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1925/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/1806 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1806/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1806/comments | https://api.github.com/repos/ollama/ollama/issues/1806/events | https://github.com/ollama/ollama/issues/1806 | 2,067,309,672 | I_kwDOJ0Z1Ps57OKRo | 1,806 | Add cli switch to show generation time and tokens/sec output time | {
"login": "thawkins",
"id": 14537,
"node_id": "MDQ6VXNlcjE0NTM3",
"avatar_url": "https://avatars.githubusercontent.com/u/14537?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/thawkins",
"html_url": "https://github.com/thawkins",
"followers_url": "https://api.github.com/users/thawkins/foll... | [] | closed | false | null | [] | null | 3 | 2024-01-05T12:56:05 | 2024-01-08T18:56:01 | 2024-01-08T18:56:01 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Would it be possible to add a metrics switch to show net generation time and output time with tokens/seconds. This would make comparing the performance of LLMs easier. | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1806/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1806/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4451 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4451/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4451/comments | https://api.github.com/repos/ollama/ollama/issues/4451/events | https://github.com/ollama/ollama/pull/4451 | 2,297,709,787 | PR_kwDOJ0Z1Ps5vhZ4t | 4,451 | add ability to create a client without env file | {
"login": "Tyrell04",
"id": 43107913,
"node_id": "MDQ6VXNlcjQzMTA3OTEz",
"avatar_url": "https://avatars.githubusercontent.com/u/43107913?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Tyrell04",
"html_url": "https://github.com/Tyrell04",
"followers_url": "https://api.github.com/users/Tyr... | [] | closed | false | null | [] | null | 3 | 2024-05-15T11:57:58 | 2024-05-19T04:40:18 | 2024-05-18T03:08:30 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4451",
"html_url": "https://github.com/ollama/ollama/pull/4451",
"diff_url": "https://github.com/ollama/ollama/pull/4451.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4451.patch",
"merged_at": null
} | If you use Ollama as a Go package, you are forced to use an environment variable as a configuration to change the host. With the CreateClient function, you are free to use your own configuration passed into the Ollama client. | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4451/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4451/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7324 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7324/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7324/comments | https://api.github.com/repos/ollama/ollama/issues/7324/events | https://github.com/ollama/ollama/pull/7324 | 2,606,172,201 | PR_kwDOJ0Z1Ps5_f5eG | 7,324 | docs: use mingw32-make rather than MSYS Makefiles | {
"login": "shou692199",
"id": 45505768,
"node_id": "MDQ6VXNlcjQ1NTA1NzY4",
"avatar_url": "https://avatars.githubusercontent.com/u/45505768?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/shou692199",
"html_url": "https://github.com/shou692199",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 0 | 2024-10-22T18:26:14 | 2024-10-31T13:39:53 | 2024-10-31T13:39:53 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7324",
"html_url": "https://github.com/ollama/ollama/pull/7324",
"diff_url": "https://github.com/ollama/ollama/pull/7324.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7324.patch",
"merged_at": null
} | Since I'm trying to port ggml-sycl to llama package, MSYS Makefiles doesn't support icx compiler but MinGW Makefiles does, and it also works well with cpu, rocm and cuda. I think it's an ideal choice. | {
"login": "shou692199",
"id": 45505768,
"node_id": "MDQ6VXNlcjQ1NTA1NzY4",
"avatar_url": "https://avatars.githubusercontent.com/u/45505768?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/shou692199",
"html_url": "https://github.com/shou692199",
"followers_url": "https://api.github.com/use... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7324/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7324/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2292 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2292/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2292/comments | https://api.github.com/repos/ollama/ollama/issues/2292/events | https://github.com/ollama/ollama/issues/2292 | 2,110,861,549 | I_kwDOJ0Z1Ps590TDt | 2,292 | Manifest file? | {
"login": "EquaTechnologies",
"id": 107852330,
"node_id": "U_kgDOBm2yKg",
"avatar_url": "https://avatars.githubusercontent.com/u/107852330?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/EquaTechnologies",
"html_url": "https://github.com/EquaTechnologies",
"followers_url": "https://api.gi... | [] | closed | false | null | [] | null | 1 | 2024-01-31T19:57:51 | 2024-02-20T04:05:47 | 2024-02-20T04:05:47 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | qua@equa-Swift-SF314-54:~$ ollama run orca
pulling manifest
Error: pull model manifest: file does not exist
equa@equa-Swift-SF314-54:~$ | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2292/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2292/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6240 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6240/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6240/comments | https://api.github.com/repos/ollama/ollama/issues/6240/events | https://github.com/ollama/ollama/issues/6240 | 2,454,100,567 | I_kwDOJ0Z1Ps6SRppX | 6,240 | Not executed in gpu amd rx 6750 GRE | {
"login": "21307369",
"id": 47931342,
"node_id": "MDQ6VXNlcjQ3OTMxMzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47931342?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/21307369",
"html_url": "https://github.com/21307369",
"followers_url": "https://api.github.com/users/213... | [
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] | closed | false | null | [] | null | 3 | 2024-08-07T18:39:39 | 2024-09-03T19:10:36 | 2024-08-09T19:25:52 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | <img width="898" alt="image" src="https://github.com/user-attachments/assets/87a706b4-7749-443f-86ea-86a3c7de1cc1">
2024/08/08 14:17:22 routes.go:1108: INFO server config env="map[OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_KEEP_ALIVE:5m0s OLLAMA_LLM_LIBRARY: OLLAMA_MAX_... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6240/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6240/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5511 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5511/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5511/comments | https://api.github.com/repos/ollama/ollama/issues/5511/events | https://github.com/ollama/ollama/pull/5511 | 2,393,285,539 | PR_kwDOJ0Z1Ps50kwRU | 5,511 | llm: fix missing dylibs by restoring old build behavior on Linux and macOS | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 0 | 2024-07-06T01:00:34 | 2024-07-08T16:55:51 | 2024-07-06T01:48:32 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5511",
"html_url": "https://github.com/ollama/ollama/pull/5511",
"diff_url": "https://github.com/ollama/ollama/pull/5511.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5511.patch",
"merged_at": "2024-07-06T01:48:32"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5511/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5511/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7660 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7660/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7660/comments | https://api.github.com/repos/ollama/ollama/issues/7660/events | https://github.com/ollama/ollama/issues/7660 | 2,657,322,190 | I_kwDOJ0Z1Ps6eY4TO | 7,660 | on ollama model specifics web sites , add comaptible copypaste <name>:<tag> text | {
"login": "fxmbsw7",
"id": 39368685,
"node_id": "MDQ6VXNlcjM5MzY4Njg1",
"avatar_url": "https://avatars.githubusercontent.com/u/39368685?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/fxmbsw7",
"html_url": "https://github.com/fxmbsw7",
"followers_url": "https://api.github.com/users/fxmbsw... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 6573197867,
"node_id": ... | closed | false | {
"login": "hoyyeva",
"id": 63033505,
"node_id": "MDQ6VXNlcjYzMDMzNTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoyyeva",
"html_url": "https://github.com/hoyyeva",
"followers_url": "https://api.github.com/users/hoyyev... | [
{
"login": "hoyyeva",
"id": 63033505,
"node_id": "MDQ6VXNlcjYzMDMzNTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoyyeva",
"html_url": "https://github.com/hoyyeva",
"followers_url": "https://api.git... | null | 9 | 2024-11-14T02:16:13 | 2024-12-01T07:31:54 | 2024-11-15T02:24:09 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | eg , user visits page of a model
chooses select list , and selects some model version
there on that version of model page i want
say , for qwen2.5-coder.on its 14b-instruct-q4_K_M page
simply qwen2.5-coder:14b-instruct-q4_K_M
name:tag
like hf s use that button with ollama copypaste run cmd in
oh and yea , pe... | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7660/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7660/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/630 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/630/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/630/comments | https://api.github.com/repos/ollama/ollama/issues/630/events | https://github.com/ollama/ollama/issues/630 | 1,916,976,222 | I_kwDOJ0Z1Ps5yQrxe | 630 | Error: failed to start a llama runner | {
"login": "azhang",
"id": 940275,
"node_id": "MDQ6VXNlcjk0MDI3NQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/940275?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/azhang",
"html_url": "https://github.com/azhang",
"followers_url": "https://api.github.com/users/azhang/follow... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api... | null | 28 | 2023-09-28T08:25:32 | 2024-07-12T19:18:02 | 2023-10-12T15:16:38 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | When I run
ollama run mistral
it downloads properly but then fails to run it, with the following error:
Error: failed to start a llama runner
I'm running this on my intel mbp with 64g ram | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/630/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/630/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3661 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3661/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3661/comments | https://api.github.com/repos/ollama/ollama/issues/3661/events | https://github.com/ollama/ollama/pull/3661 | 2,244,821,343 | PR_kwDOJ0Z1Ps5svQOm | 3,661 | better checking for OLLAMA_HOST variable | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | [] | closed | false | null | [] | null | 0 | 2024-04-15T23:39:21 | 2024-04-29T23:14:08 | 2024-04-29T23:14:07 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3661",
"html_url": "https://github.com/ollama/ollama/pull/3661",
"diff_url": "https://github.com/ollama/ollama/pull/3661.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3661.patch",
"merged_at": "2024-04-29T23:14:07"
} | This change adds better validation to the `OLLAMA_HOST` variable when used with `ollama serve`. It should work with both IPv4 and IPv6, and includes unit tests. | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3661/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3661/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1109 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1109/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1109/comments | https://api.github.com/repos/ollama/ollama/issues/1109/events | https://github.com/ollama/ollama/issues/1109 | 1,990,754,546 | I_kwDOJ0Z1Ps52qIDy | 1,109 | Exploring Multimodal LLMs: Incorporating Image as Input in ModelFile. | {
"login": "saifullah-N",
"id": 79261602,
"node_id": "MDQ6VXNlcjc5MjYxNjAy",
"avatar_url": "https://avatars.githubusercontent.com/u/79261602?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/saifullah-N",
"html_url": "https://github.com/saifullah-N",
"followers_url": "https://api.github.com/... | [] | closed | false | null | [] | null | 1 | 2023-11-13T14:18:07 | 2023-11-13T17:42:57 | 2023-11-13T17:42:57 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi, I'd like to experiment with multimodal language models that can handle both images and text as input. Is there a way to input an image.
My ModelFile looks like this
```
FROM ./my_model.gguf
PARAMERTER..
SYSTEM You are An Artist describe the image
```
I'm trying to run this on macos.
M1 chip | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1109/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1109/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3728 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3728/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3728/comments | https://api.github.com/repos/ollama/ollama/issues/3728/events | https://github.com/ollama/ollama/issues/3728 | 2,249,868,609 | I_kwDOJ0Z1Ps6GGkVB | 3,728 | ollama prune | {
"login": "Tobiadefami",
"id": 48369656,
"node_id": "MDQ6VXNlcjQ4MzY5NjU2",
"avatar_url": "https://avatars.githubusercontent.com/u/48369656?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Tobiadefami",
"html_url": "https://github.com/Tobiadefami",
"followers_url": "https://api.github.com/... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 2 | 2024-04-18T06:35:53 | 2024-04-18T08:19:21 | 2024-04-18T08:19:21 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I think ollama needs a prune command, something similar to `docker system prune`. Just to keep the environment tidy | {
"login": "Tobiadefami",
"id": 48369656,
"node_id": "MDQ6VXNlcjQ4MzY5NjU2",
"avatar_url": "https://avatars.githubusercontent.com/u/48369656?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Tobiadefami",
"html_url": "https://github.com/Tobiadefami",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3728/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3728/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3660 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3660/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3660/comments | https://api.github.com/repos/ollama/ollama/issues/3660/events | https://github.com/ollama/ollama/issues/3660 | 2,244,762,677 | I_kwDOJ0Z1Ps6FzFw1 | 3,660 | Exception using command-r-plus:104b-q4_0 | {
"login": "mehdiataei",
"id": 29686241,
"node_id": "MDQ6VXNlcjI5Njg2MjQx",
"avatar_url": "https://avatars.githubusercontent.com/u/29686241?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mehdiataei",
"html_url": "https://github.com/mehdiataei",
"followers_url": "https://api.github.com/use... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 7 | 2024-04-15T22:49:24 | 2024-04-17T00:41:10 | 2024-04-17T00:41:10 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Ollama: exception done_getting_tensors: wrong number of tensors; expected 642, got 514
### What did you expect to see?
_No response_
### Steps to reproduce
_No response_
### Are there any recent changes that introduced the issue?
_No response_
### OS
_No response_
### Architecture
_No ... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3660/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3660/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2792 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2792/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2792/comments | https://api.github.com/repos/ollama/ollama/issues/2792/events | https://github.com/ollama/ollama/issues/2792 | 2,157,751,613 | I_kwDOJ0Z1Ps6AnK09 | 2,792 | Subsequent generation requests hang after successful generation request with `num_predict: 0` | {
"login": "stanier",
"id": 8770519,
"node_id": "MDQ6VXNlcjg3NzA1MTk=",
"avatar_url": "https://avatars.githubusercontent.com/u/8770519?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/stanier",
"html_url": "https://github.com/stanier",
"followers_url": "https://api.github.com/users/stanier/... | [] | closed | false | null | [] | null | 1 | 2024-02-27T22:54:24 | 2024-05-10T01:13:35 | 2024-05-10T01:13:35 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | If you submit a generation request with `num_predict: 0`, the request will be handled successfully but all subsequent generation requests will hang indefinitely regardless of their `num_predict` values.
Below is an example of how to reproduce this behavior:
```
curl http://localhost:11434/api/generate -d '{
"... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2792/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2792/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1849 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1849/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1849/comments | https://api.github.com/repos/ollama/ollama/issues/1849/events | https://github.com/ollama/ollama/pull/1849 | 2,069,374,142 | PR_kwDOJ0Z1Ps5jatea | 1,849 | Accomodate split cuda lib dir | {
"login": "mraiser",
"id": 1268027,
"node_id": "MDQ6VXNlcjEyNjgwMjc=",
"avatar_url": "https://avatars.githubusercontent.com/u/1268027?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mraiser",
"html_url": "https://github.com/mraiser",
"followers_url": "https://api.github.com/users/mraiser/... | [] | closed | false | null | [] | null | 9 | 2024-01-08T00:24:45 | 2024-02-06T00:01:17 | 2024-02-06T00:01:17 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1849",
"html_url": "https://github.com/ollama/ollama/pull/1849",
"diff_url": "https://github.com/ollama/ollama/pull/1849.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1849.patch",
"merged_at": "2024-02-06T00:01:17"
} | Makes it a little easier to compile when cuda lib dir is split up as in nixos. | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1849/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1849/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6786 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6786/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6786/comments | https://api.github.com/repos/ollama/ollama/issues/6786/events | https://github.com/ollama/ollama/issues/6786 | 2,523,912,567 | I_kwDOJ0Z1Ps6Wb9l3 | 6,786 | Isn't it time to move onto Omni models? | {
"login": "Meshwa428",
"id": 135232056,
"node_id": "U_kgDOCA96OA",
"avatar_url": "https://avatars.githubusercontent.com/u/135232056?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Meshwa428",
"html_url": "https://github.com/Meshwa428",
"followers_url": "https://api.github.com/users/Meshwa... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2024-09-13T05:34:46 | 2024-10-23T17:17:40 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | There is a model that I found today called
*LLaMa 3.1 8b Omni* and it is a speech to speech model with very low latency ensuring the best experience for local models.
But as for Ollama it doesn't support such models, although VLMs are there but having these Omni models on your local device is just 🤌🏻 too Good to be ... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6786/reactions",
"total_count": 9,
"+1": 9,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6786/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/5244 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5244/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5244/comments | https://api.github.com/repos/ollama/ollama/issues/5244/events | https://github.com/ollama/ollama/pull/5244 | 2,368,887,556 | PR_kwDOJ0Z1Ps5zS5Hb | 5,244 | llm: suppress large allocations for GGUF arrays | {
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers"... | [] | closed | false | null | [] | null | 1 | 2024-06-23T21:01:06 | 2024-06-23T22:32:04 | 2024-06-23T22:32:04 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | true | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5244",
"html_url": "https://github.com/ollama/ollama/pull/5244",
"diff_url": "https://github.com/ollama/ollama/pull/5244.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5244.patch",
"merged_at": null
} | This introduces a little array type for holding GGUF arrays that prevents the array from growing too large. It preserves the total size of the array, but limits the number of elements that are actually allocated.
GGUF arrays that are extremely large, such as tokens, etc, are generally uninteresting to users, and are... | {
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers"... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5244/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5244/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/836 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/836/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/836/comments | https://api.github.com/repos/ollama/ollama/issues/836/events | https://github.com/ollama/ollama/issues/836 | 1,949,521,611 | I_kwDOJ0Z1Ps50M1bL | 836 | Support setting parameters without a modelfile | {
"login": "chigkim",
"id": 22120994,
"node_id": "MDQ6VXNlcjIyMTIwOTk0",
"avatar_url": "https://avatars.githubusercontent.com/u/22120994?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/chigkim",
"html_url": "https://github.com/chigkim",
"followers_url": "https://api.github.com/users/chigki... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | [
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/us... | null | 3 | 2023-10-18T11:30:51 | 2024-05-05T23:14:26 | 2024-01-03T03:38:43 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I know you can set temperature using modelfile or api, but it would be much more convenient to be able to do it in cli.
For example:
`ollama run llama-2 -temperture 0.1` when loading
and/or
`/set temperature 0.1` in runtime
Thanks for your consideration! | {
"login": "chigkim",
"id": 22120994,
"node_id": "MDQ6VXNlcjIyMTIwOTk0",
"avatar_url": "https://avatars.githubusercontent.com/u/22120994?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/chigkim",
"html_url": "https://github.com/chigkim",
"followers_url": "https://api.github.com/users/chigki... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/836/reactions",
"total_count": 5,
"+1": 5,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/836/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4747 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4747/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4747/comments | https://api.github.com/repos/ollama/ollama/issues/4747/events | https://github.com/ollama/ollama/issues/4747 | 2,327,194,470 | I_kwDOJ0Z1Ps6Ktitm | 4,747 | Running multiple models simultaneously, always using one card | {
"login": "leoHostProject",
"id": 87935281,
"node_id": "MDQ6VXNlcjg3OTM1Mjgx",
"avatar_url": "https://avatars.githubusercontent.com/u/87935281?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/leoHostProject",
"html_url": "https://github.com/leoHostProject",
"followers_url": "https://api.gi... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 4 | 2024-05-31T07:58:11 | 2024-05-31T18:58:02 | 2024-05-31T18:57:10 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Running multiple models simultaneously, always using one card,but i have 4 cards and download 4model
When multiple users are using it at the same time, always clear the first card and then load other models instead of using my other idle cards
### OS
Linux
### GPU
Nvidia
### CPU
Intel
... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4747/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4747/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4583 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4583/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4583/comments | https://api.github.com/repos/ollama/ollama/issues/4583/events | https://github.com/ollama/ollama/pull/4583 | 2,311,914,296 | PR_kwDOJ0Z1Ps5wRyOi | 4,583 | Add new community integration (TypingMind) | {
"login": "trungdq88",
"id": 4214509,
"node_id": "MDQ6VXNlcjQyMTQ1MDk=",
"avatar_url": "https://avatars.githubusercontent.com/u/4214509?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/trungdq88",
"html_url": "https://github.com/trungdq88",
"followers_url": "https://api.github.com/users/tr... | [] | closed | false | null | [] | null | 2 | 2024-05-23T04:31:39 | 2024-06-10T09:27:58 | 2024-06-10T09:27:57 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4583",
"html_url": "https://github.com/ollama/ollama/pull/4583",
"diff_url": "https://github.com/ollama/ollama/pull/4583.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4583.patch",
"merged_at": null
} | Cheers! | {
"login": "trungdq88",
"id": 4214509,
"node_id": "MDQ6VXNlcjQyMTQ1MDk=",
"avatar_url": "https://avatars.githubusercontent.com/u/4214509?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/trungdq88",
"html_url": "https://github.com/trungdq88",
"followers_url": "https://api.github.com/users/tr... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4583/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4583/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5688 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5688/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5688/comments | https://api.github.com/repos/ollama/ollama/issues/5688/events | https://github.com/ollama/ollama/pull/5688 | 2,407,318,220 | PR_kwDOJ0Z1Ps51T_VY | 5,688 | Serve static files | {
"login": "1feralcat",
"id": 51179976,
"node_id": "MDQ6VXNlcjUxMTc5OTc2",
"avatar_url": "https://avatars.githubusercontent.com/u/51179976?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/1feralcat",
"html_url": "https://github.com/1feralcat",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 1 | 2024-07-14T08:06:08 | 2024-08-14T16:46:19 | 2024-08-14T16:46:19 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5688",
"html_url": "https://github.com/ollama/ollama/pull/5688",
"diff_url": "https://github.com/ollama/ollama/pull/5688.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5688.patch",
"merged_at": null
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5688/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5688/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5707 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5707/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5707/comments | https://api.github.com/repos/ollama/ollama/issues/5707/events | https://github.com/ollama/ollama/pull/5707 | 2,409,438,713 | PR_kwDOJ0Z1Ps51bLDN | 5,707 | Image Embeddings | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjha... | [] | closed | false | null | [] | null | 1 | 2024-07-15T19:13:38 | 2024-11-21T10:11:28 | 2024-11-21T10:11:28 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | true | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5707",
"html_url": "https://github.com/ollama/ollama/pull/5707",
"diff_url": "https://github.com/ollama/ollama/pull/5707.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5707.patch",
"merged_at": null
} | rough work | {
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/us... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5707/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5707/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6117 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6117/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6117/comments | https://api.github.com/repos/ollama/ollama/issues/6117/events | https://github.com/ollama/ollama/issues/6117 | 2,442,322,095 | I_kwDOJ0Z1Ps6RkuCv | 6,117 | Add Gemma 2 2b base/ text/ pre-trained model to registry | {
"login": "nviraj",
"id": 8409854,
"node_id": "MDQ6VXNlcjg0MDk4NTQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/8409854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nviraj",
"html_url": "https://github.com/nviraj",
"followers_url": "https://api.github.com/users/nviraj/foll... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 1 | 2024-08-01T12:25:31 | 2024-08-02T12:19:55 | 2024-08-02T12:19:55 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Apologies, if I am missing something but can you upload quants for [base 2b](https://huggingface.co/google/gemma-2-2b) as well?
All I see are instruct ones [here](https://ollama.com/library/gemma2/tags)
Thanks! | {
"login": "nviraj",
"id": 8409854,
"node_id": "MDQ6VXNlcjg0MDk4NTQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/8409854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nviraj",
"html_url": "https://github.com/nviraj",
"followers_url": "https://api.github.com/users/nviraj/foll... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6117/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6117/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4652 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4652/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4652/comments | https://api.github.com/repos/ollama/ollama/issues/4652/events | https://github.com/ollama/ollama/pull/4652 | 2,317,888,554 | PR_kwDOJ0Z1Ps5wmNS4 | 4,652 | Ensure `nvidia` and `nvidia_uvm` kernel modules are loaded in `install.sh` script and at startup | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 0 | 2024-05-26T18:53:36 | 2024-05-26T21:57:18 | 2024-05-26T21:57:17 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4652",
"html_url": "https://github.com/ollama/ollama/pull/4652",
"diff_url": "https://github.com/ollama/ollama/pull/4652.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4652.patch",
"merged_at": "2024-05-26T21:57:17"
} | Make sure the `nvidia` and `nvidia_uvm` kernel modules are loaded on install.
Nvidia has a daemon that takes care of that `nvidia-persistenced` for restarts, so add both to its config file as well
Lastly, make sure the kernel modules are loaded when re-running the install script, even if drivers are already insta... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4652/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4652/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1988 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1988/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1988/comments | https://api.github.com/repos/ollama/ollama/issues/1988/events | https://github.com/ollama/ollama/pull/1988 | 2,080,801,827 | PR_kwDOJ0Z1Ps5kBthD | 1,988 | Fix typo in arm mac arch script | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 1 | 2024-01-14T16:34:14 | 2024-01-14T16:45:25 | 2024-01-14T16:45:18 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1988",
"html_url": "https://github.com/ollama/ollama/pull/1988",
"diff_url": "https://github.com/ollama/ollama/pull/1988.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1988.patch",
"merged_at": "2024-01-14T16:45:18"
} | null | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1988/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1988/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3401 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3401/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3401/comments | https://api.github.com/repos/ollama/ollama/issues/3401/events | https://github.com/ollama/ollama/issues/3401 | 2,214,382,654 | I_kwDOJ0Z1Ps6D_Mw- | 3,401 | Document `OLLAMA_DEBUG` in `ollama serve` `-h` docs | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 5667396210,
"node_id": ... | closed | false | null | [] | null | 2 | 2024-03-29T00:27:47 | 2024-05-13T20:20:58 | 2024-05-13T20:20:58 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What are you trying to do?
_No response_
### How should we solve this?
_No response_
### What is the impact of not solving this?
_No response_
### Anything else?
_No response_ | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3401/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3401/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1648 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1648/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1648/comments | https://api.github.com/repos/ollama/ollama/issues/1648/events | https://github.com/ollama/ollama/issues/1648 | 2,051,491,566 | I_kwDOJ0Z1Ps56R0bu | 1,648 | gpu issues | {
"login": "RootnuII",
"id": 66104474,
"node_id": "MDQ6VXNlcjY2MTA0NDc0",
"avatar_url": "https://avatars.githubusercontent.com/u/66104474?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/RootnuII",
"html_url": "https://github.com/RootnuII",
"followers_url": "https://api.github.com/users/Roo... | [] | closed | false | null | [] | null | 5 | 2023-12-21T00:38:19 | 2023-12-23T15:16:26 | 2023-12-23T09:59:42 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | warning: gpu support may not be enabled, check that you have installed GPU drivers: nvidia-smi command failed
i have installed all drivers for my gpu:1070ti
still nothing maybe i installed the wrong drivers can someone point me in the right direction
| {
"login": "RootnuII",
"id": 66104474,
"node_id": "MDQ6VXNlcjY2MTA0NDc0",
"avatar_url": "https://avatars.githubusercontent.com/u/66104474?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/RootnuII",
"html_url": "https://github.com/RootnuII",
"followers_url": "https://api.github.com/users/Roo... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1648/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1648/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2828 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2828/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2828/comments | https://api.github.com/repos/ollama/ollama/issues/2828/events | https://github.com/ollama/ollama/pull/2828 | 2,160,642,796 | PR_kwDOJ0Z1Ps5oQuDE | 2,828 | update llama.cpp submodule commit to `87c91c07` | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | [] | closed | false | null | [] | null | 0 | 2024-02-29T07:59:29 | 2024-02-29T17:42:09 | 2024-02-29T17:42:08 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2828",
"html_url": "https://github.com/ollama/ollama/pull/2828",
"diff_url": "https://github.com/ollama/ollama/pull/2828.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2828.patch",
"merged_at": "2024-02-29T17:42:08"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2828/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2828/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1501 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1501/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1501/comments | https://api.github.com/repos/ollama/ollama/issues/1501/events | https://github.com/ollama/ollama/issues/1501 | 2,039,679,106 | I_kwDOJ0Z1Ps55kwiC | 1,501 | macOS environment variable not working | {
"login": "brandoncarl",
"id": 5427971,
"node_id": "MDQ6VXNlcjU0Mjc5NzE=",
"avatar_url": "https://avatars.githubusercontent.com/u/5427971?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/brandoncarl",
"html_url": "https://github.com/brandoncarl",
"followers_url": "https://api.github.com/us... | [] | closed | false | null | [] | null | 9 | 2023-12-13T13:08:29 | 2025-01-11T11:41:21 | 2024-01-22T23:41:41 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hello - thanks for the great repository. I wanted to alert you to the fact that the OLLAMA_MODELS path appears to be having no impact. This is true for `pull`, `rull` and serving.
```
$ echo $OLLAMA_MODELS
(prints appropriate directory)
$ ollama run <model>
(downloads to ~/.ollama/..)
$ OLLAMA_MODELS=<direc... | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1501/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1501/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4947 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4947/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4947/comments | https://api.github.com/repos/ollama/ollama/issues/4947/events | https://github.com/ollama/ollama/issues/4947 | 2,342,185,507 | I_kwDOJ0Z1Ps6Lmuoj | 4,947 | Cannot update ollama on windows through "restart to update" | {
"login": "TonyBlur",
"id": 101348912,
"node_id": "U_kgDOBgp2MA",
"avatar_url": "https://avatars.githubusercontent.com/u/101348912?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/TonyBlur",
"html_url": "https://github.com/TonyBlur",
"followers_url": "https://api.github.com/users/TonyBlur/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 11 | 2024-06-09T11:52:04 | 2024-06-10T11:42:01 | 2024-06-10T11:42:01 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
cannot update ollama on windows through clicking "restart to update" at taskbar, it has no response
### OS
Windows
### GPU
AMD
### CPU
AMD
### Ollama version
0.1.41 | {
"login": "TonyBlur",
"id": 101348912,
"node_id": "U_kgDOBgp2MA",
"avatar_url": "https://avatars.githubusercontent.com/u/101348912?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/TonyBlur",
"html_url": "https://github.com/TonyBlur",
"followers_url": "https://api.github.com/users/TonyBlur/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4947/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4947/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5918 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5918/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5918/comments | https://api.github.com/repos/ollama/ollama/issues/5918/events | https://github.com/ollama/ollama/issues/5918 | 2,427,947,507 | I_kwDOJ0Z1Ps6Qt4nz | 5,918 | Llama3.1 70b-instruct-q4_1 buggy | {
"login": "velaia",
"id": 1515904,
"node_id": "MDQ6VXNlcjE1MTU5MDQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/1515904?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/velaia",
"html_url": "https://github.com/velaia",
"followers_url": "https://api.github.com/users/velaia/foll... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 14 | 2024-07-24T15:54:38 | 2024-09-15T17:35:22 | 2024-09-15T17:18:31 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
When I run the **70b-instruct-q4_1** version of Llama3.1 ollama gives a buggy reply:
My sample request:
> ➜ ollama-tests curl http://localhost:11434/api/chat -d '{
"model": "llama3.1:70b-instruct-q4_1",
"messages": [
{
"role": "system",
"content": "You a... | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5918/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5918/timeline | null | not_planned | false |
https://api.github.com/repos/ollama/ollama/issues/2496 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2496/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2496/comments | https://api.github.com/repos/ollama/ollama/issues/2496/events | https://github.com/ollama/ollama/issues/2496 | 2,134,803,651 | I_kwDOJ0Z1Ps5_PoTD | 2,496 | default num_thread incorrect on some large core count system (non-hyperthreading) | {
"login": "mokkin",
"id": 2938748,
"node_id": "MDQ6VXNlcjI5Mzg3NDg=",
"avatar_url": "https://avatars.githubusercontent.com/u/2938748?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mokkin",
"html_url": "https://github.com/mokkin",
"followers_url": "https://api.github.com/users/mokkin/foll... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 35 | 2024-02-14T17:10:39 | 2024-08-05T22:20:08 | 2024-08-05T22:20:08 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I have tested Ollama on different machines yet, but no matter how many cores or RAM I have, it's only using 50% of the cores and just a very few GB of RAM.
For example now I'm running `ollama rum llama2:70b` on 16 core server with 32 GB of RAM, but while prompting only eight cores are used and just around 1 GB of RAM.... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2496/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2496/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3754 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3754/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3754/comments | https://api.github.com/repos/ollama/ollama/issues/3754/events | https://github.com/ollama/ollama/issues/3754 | 2,253,202,923 | I_kwDOJ0Z1Ps6GTSXr | 3,754 | ollama rm Support deletion through ID | {
"login": "eryajf",
"id": 33259379,
"node_id": "MDQ6VXNlcjMzMjU5Mzc5",
"avatar_url": "https://avatars.githubusercontent.com/u/33259379?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/eryajf",
"html_url": "https://github.com/eryajf",
"followers_url": "https://api.github.com/users/eryajf/fo... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2024-04-19T14:48:40 | 2024-04-19T14:48:40 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Just like Docker RM supports deletion through ID, have you considered supporting model deletion through ID? If so, I am willing to research this area and submit a PR | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3754/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3754/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/225 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/225/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/225/comments | https://api.github.com/repos/ollama/ollama/issues/225/events | https://github.com/ollama/ollama/pull/225 | 1,824,949,943 | PR_kwDOJ0Z1Ps5Wk5I4 | 225 | add stop conditions | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 0 | 2023-07-27T18:28:50 | 2023-07-28T00:20:57 | 2023-07-28T00:20:56 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/225",
"html_url": "https://github.com/ollama/ollama/pull/225",
"diff_url": "https://github.com/ollama/ollama/pull/225.diff",
"patch_url": "https://github.com/ollama/ollama/pull/225.patch",
"merged_at": "2023-07-28T00:20:56"
} | resolves #140
resolves #217 | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/225/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/225/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1653 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1653/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1653/comments | https://api.github.com/repos/ollama/ollama/issues/1653/events | https://github.com/ollama/ollama/issues/1653 | 2,052,245,322 | I_kwDOJ0Z1Ps56UsdK | 1,653 | shell autocompletion | {
"login": "teto",
"id": 886074,
"node_id": "MDQ6VXNlcjg4NjA3NA==",
"avatar_url": "https://avatars.githubusercontent.com/u/886074?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/teto",
"html_url": "https://github.com/teto",
"followers_url": "https://api.github.com/users/teto/followers",
... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 6 | 2023-12-21T11:28:48 | 2024-12-04T05:19:01 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | hi, I would like to add shell autocompletion for ollama to my linux distribution.
I dont know go but the cli parser seems to be using https://github.com/spf13/cobra, the same as github CLI `gh`.
The nix package generates the autocompletion like this:
https://github.com/teto/nixpkgs/blob/dbaa028d61848fda534a04ce2145... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1653/reactions",
"total_count": 23,
"+1": 23,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1653/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/3365 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3365/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3365/comments | https://api.github.com/repos/ollama/ollama/issues/3365/events | https://github.com/ollama/ollama/issues/3365 | 2,209,739,153 | I_kwDOJ0Z1Ps6DtfGR | 3,365 | Unable to make cors work in docker container | {
"login": "gbarton",
"id": 1878953,
"node_id": "MDQ6VXNlcjE4Nzg5NTM=",
"avatar_url": "https://avatars.githubusercontent.com/u/1878953?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gbarton",
"html_url": "https://github.com/gbarton",
"followers_url": "https://api.github.com/users/gbarton/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 3 | 2024-03-27T03:24:08 | 2024-05-02T02:15:59 | 2024-03-27T22:44:09 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I have been trying to get cors to work inside a docker container with no luck so far. Attached is my docker-compose.yaml for starting ollama. I have tried many variations on the OLLAMA_HOST and OLLAMA_ORIGINS env vars to no luck. Does anyone know the magic combo to make it work?
### What did ... | {
"login": "gbarton",
"id": 1878953,
"node_id": "MDQ6VXNlcjE4Nzg5NTM=",
"avatar_url": "https://avatars.githubusercontent.com/u/1878953?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gbarton",
"html_url": "https://github.com/gbarton",
"followers_url": "https://api.github.com/users/gbarton/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3365/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3365/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2686 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2686/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2686/comments | https://api.github.com/repos/ollama/ollama/issues/2686/events | https://github.com/ollama/ollama/pull/2686 | 2,149,472,407 | PR_kwDOJ0Z1Ps5nqz9L | 2,686 | update README to add Gemma 2B, 7B model in Model Library Table | {
"login": "adminazhar",
"id": 20738252,
"node_id": "MDQ6VXNlcjIwNzM4MjUy",
"avatar_url": "https://avatars.githubusercontent.com/u/20738252?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/adminazhar",
"html_url": "https://github.com/adminazhar",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 0 | 2024-02-22T16:36:57 | 2024-02-22T18:07:47 | 2024-02-22T18:07:47 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2686",
"html_url": "https://github.com/ollama/ollama/pull/2686",
"diff_url": "https://github.com/ollama/ollama/pull/2686.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2686.patch",
"merged_at": "2024-02-22T18:07:47"
} | @pfrankov @jmorganca Kindly Review | {
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/us... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2686/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2686/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2911 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2911/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2911/comments | https://api.github.com/repos/ollama/ollama/issues/2911/events | https://github.com/ollama/ollama/issues/2911 | 2,166,737,391 | I_kwDOJ0Z1Ps6BJcnv | 2,911 | System Monitor detected that internet is used by Ollama. | {
"login": "kukidevs",
"id": 113847173,
"node_id": "U_kgDOBskrhQ",
"avatar_url": "https://avatars.githubusercontent.com/u/113847173?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kukidevs",
"html_url": "https://github.com/kukidevs",
"followers_url": "https://api.github.com/users/kukidevs/... | [] | closed | false | null | [] | null | 2 | 2024-03-04T12:29:53 | 2025-01-30T01:41:17 | 2024-03-04T17:50:06 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | 
What does it send/receive? I was using it through MacOS terminal, installed the app from the website. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2911/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2911/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2200 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2200/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2200/comments | https://api.github.com/repos/ollama/ollama/issues/2200/events | https://github.com/ollama/ollama/issues/2200 | 2,101,705,603 | I_kwDOJ0Z1Ps59RXuD | 2,200 | ollama is stuck when translate language | {
"login": "MunhakKim",
"id": 13396311,
"node_id": "MDQ6VXNlcjEzMzk2MzEx",
"avatar_url": "https://avatars.githubusercontent.com/u/13396311?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/MunhakKim",
"html_url": "https://github.com/MunhakKim",
"followers_url": "https://api.github.com/users/... | [
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] | closed | false | null | [] | null | 2 | 2024-01-26T07:10:48 | 2024-03-11T23:14:33 | 2024-03-11T23:14:33 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ollama is stuck when i ask to translate language.
In my case, i put it in like '한글로' for prompt which means 'in Korean' after English output.
Sometimes ollama could translate perfectly and stable, but mostly ollama is stuck.
I think if there is a problem in text generation, It should be pause not is broken.
In th... | {
"login": "MunhakKim",
"id": 13396311,
"node_id": "MDQ6VXNlcjEzMzk2MzEx",
"avatar_url": "https://avatars.githubusercontent.com/u/13396311?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/MunhakKim",
"html_url": "https://github.com/MunhakKim",
"followers_url": "https://api.github.com/users/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2200/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2200/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/870 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/870/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/870/comments | https://api.github.com/repos/ollama/ollama/issues/870/events | https://github.com/ollama/ollama/pull/870 | 1,955,433,858 | PR_kwDOJ0Z1Ps5dcU35 | 870 | Added a minimalist React UI for Ollama models to ReadME.md | {
"login": "richawo",
"id": 35015261,
"node_id": "MDQ6VXNlcjM1MDE1MjYx",
"avatar_url": "https://avatars.githubusercontent.com/u/35015261?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/richawo",
"html_url": "https://github.com/richawo",
"followers_url": "https://api.github.com/users/richaw... | [] | closed | false | null | [] | null | 4 | 2023-10-21T10:50:25 | 2023-10-23T14:44:39 | 2023-10-23T14:44:39 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/870",
"html_url": "https://github.com/ollama/ollama/pull/870",
"diff_url": "https://github.com/ollama/ollama/pull/870.diff",
"patch_url": "https://github.com/ollama/ollama/pull/870.patch",
"merged_at": "2023-10-23T14:44:39"
} | I discussed it a few times in the discord, and a few people seem to be using it, so it would be good to add.
Demo video:
https://github.com/jmorganca/ollama/assets/35015261/d50f7036-cdf2-44ed-9bb0-fdbed6a4ec66
I'll be maintaining/improving it a lot over the coming weeks, and some contributors reached out to get ... | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/870/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 1,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/870/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2587 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2587/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2587/comments | https://api.github.com/repos/ollama/ollama/issues/2587/events | https://github.com/ollama/ollama/issues/2587 | 2,141,539,255 | I_kwDOJ0Z1Ps5_pUu3 | 2,587 | Running on GPU | {
"login": "shersoni610",
"id": 57876250,
"node_id": "MDQ6VXNlcjU3ODc2MjUw",
"avatar_url": "https://avatars.githubusercontent.com/u/57876250?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/shersoni610",
"html_url": "https://github.com/shersoni610",
"followers_url": "https://api.github.com/... | [] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 29 | 2024-02-19T05:27:39 | 2024-02-29T23:47:57 | 2024-02-29T23:47:57 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hello,
It seems, the response time of llama2:7b is slow on my linux machine. I am not sure if the code
is running on Nvidia card.
In a python code, how to ensure that Ollama models run on GPU?
| {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2587/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2587/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5742 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5742/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5742/comments | https://api.github.com/repos/ollama/ollama/issues/5742/events | https://github.com/ollama/ollama/pull/5742 | 2,413,170,315 | PR_kwDOJ0Z1Ps51nl1k | 5,742 | Added code to use swap memory in linux | {
"login": "avinash-ghadshi",
"id": 65444549,
"node_id": "MDQ6VXNlcjY1NDQ0NTQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/65444549?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/avinash-ghadshi",
"html_url": "https://github.com/avinash-ghadshi",
"followers_url": "https://api... | [] | closed | false | null | [] | null | 5 | 2024-07-17T09:54:10 | 2024-09-03T16:54:39 | 2024-09-03T16:54:39 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5742",
"html_url": "https://github.com/ollama/ollama/pull/5742",
"diff_url": "https://github.com/ollama/ollama/pull/5742.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5742.patch",
"merged_at": null
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5742/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5742/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6420 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6420/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6420/comments | https://api.github.com/repos/ollama/ollama/issues/6420/events | https://github.com/ollama/ollama/issues/6420 | 2,473,039,397 | I_kwDOJ0Z1Ps6TZ5Yl | 6,420 | Is the speed of the Olama running model related to the CUDA version? | {
"login": "TianWuYuJiangHenShou",
"id": 20592000,
"node_id": "MDQ6VXNlcjIwNTkyMDAw",
"avatar_url": "https://avatars.githubusercontent.com/u/20592000?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/TianWuYuJiangHenShou",
"html_url": "https://github.com/TianWuYuJiangHenShou",
"followers_url... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg... | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 3 | 2024-08-19T10:40:53 | 2024-09-26T00:13:40 | 2024-09-26T00:13:40 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I deployed qwen2:72B with the latest version of Ollama, but I found that the loading speed of Ollama models varies greatly under different nvidia driver version.
# driver:535.183.06 | cuda version:12.2
ollama version:0.3.4
Time of Loading Model:29s
# driver:515.105.01 | cuda version:11.7
... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6420/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6420/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1418 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1418/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1418/comments | https://api.github.com/repos/ollama/ollama/issues/1418/events | https://github.com/ollama/ollama/issues/1418 | 2,031,151,545 | I_kwDOJ0Z1Ps55EOm5 | 1,418 | Make `notus` model available on `ollama` | {
"login": "davidberenstein1957",
"id": 25269220,
"node_id": "MDQ6VXNlcjI1MjY5MjIw",
"avatar_url": "https://avatars.githubusercontent.com/u/25269220?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/davidberenstein1957",
"html_url": "https://github.com/davidberenstein1957",
"followers_url": ... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 13 | 2023-12-07T16:53:09 | 2024-01-11T07:03:57 | 2024-01-11T07:03:57 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | We just hosted our [Argilla community meetup](https://www.youtube.com/playlist?list=PLBmuFBJ5cjcbsr49KFoC4DQoo3ZWT7q_d) and got the question if we would be able to make [Notus](https://argilla.io/blog/notus7b/) available on ollama. What would it take, and how can we help with this? | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1418/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1418/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1693 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1693/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1693/comments | https://api.github.com/repos/ollama/ollama/issues/1693/events | https://github.com/ollama/ollama/issues/1693 | 2,054,975,190 | I_kwDOJ0Z1Ps56fG7W | 1,693 | Possible to increase speed / efficiency of model? | {
"login": "theyluvEnething",
"id": 115583828,
"node_id": "U_kgDOBuOrVA",
"avatar_url": "https://avatars.githubusercontent.com/u/115583828?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/theyluvEnething",
"html_url": "https://github.com/theyluvEnething",
"followers_url": "https://api.githu... | [] | closed | false | null | [] | null | 5 | 2023-12-24T03:49:05 | 2024-06-22T13:47:35 | 2024-03-12T00:16:40 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I'm trying out the Dolphin-Mixture model and it's quite fun, but really slow. (my specs are 64gb 3200mj ram, i7 4.5gh cpu and a 1080 ti)
but still it takes some time to start answering and when it starts it writes at maybe 2-3 words a second. Is it possible to make improvements to this? | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1693/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1693/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4172 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4172/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4172/comments | https://api.github.com/repos/ollama/ollama/issues/4172/events | https://github.com/ollama/ollama/pull/4172 | 2,279,565,700 | PR_kwDOJ0Z1Ps5ukaTU | 4,172 | Add support for the BSD platforms | {
"login": "prep",
"id": 2017915,
"node_id": "MDQ6VXNlcjIwMTc5MTU=",
"avatar_url": "https://avatars.githubusercontent.com/u/2017915?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/prep",
"html_url": "https://github.com/prep",
"followers_url": "https://api.github.com/users/prep/followers",
... | [] | open | false | null | [] | null | 6 | 2024-05-05T14:42:06 | 2024-09-08T12:35:59 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4172",
"html_url": "https://github.com/ollama/ollama/pull/4172",
"diff_url": "https://github.com/ollama/ollama/pull/4172.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4172.patch",
"merged_at": null
} | This PR adds the ability to compile and run Ollama on various BSD platforms, specifically DragonFly, FreeBSD, NetBSD and OpenBSD. A couple notes:
1. It includes support for Vulkan to offload work to the GPU.
2. It's only been tested on OpenBSD, but hopefully interested parties in this community can test the other p... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4172/reactions",
"total_count": 4,
"+1": 4,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4172/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8169 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8169/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8169/comments | https://api.github.com/repos/ollama/ollama/issues/8169/events | https://github.com/ollama/ollama/pull/8169 | 2,749,784,967 | PR_kwDOJ0Z1Ps6FwPej | 8,169 | Update go.mod | {
"login": "Squishedmac",
"id": 88924339,
"node_id": "MDQ6VXNlcjg4OTI0MzM5",
"avatar_url": "https://avatars.githubusercontent.com/u/88924339?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Squishedmac",
"html_url": "https://github.com/Squishedmac",
"followers_url": "https://api.github.com/... | [] | closed | false | null | [] | null | 0 | 2024-12-19T09:50:26 | 2024-12-19T10:51:33 | 2024-12-19T10:51:33 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8169",
"html_url": "https://github.com/ollama/ollama/pull/8169",
"diff_url": "https://github.com/ollama/ollama/pull/8169.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8169.patch",
"merged_at": null
} | Fixed the CVE-2024-45337 severity on wiz scan for the docker image, changing the version of golang.org/x/crypto to the latest version should fix the issue | {
"login": "Squishedmac",
"id": 88924339,
"node_id": "MDQ6VXNlcjg4OTI0MzM5",
"avatar_url": "https://avatars.githubusercontent.com/u/88924339?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Squishedmac",
"html_url": "https://github.com/Squishedmac",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8169/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8169/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6890 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6890/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6890/comments | https://api.github.com/repos/ollama/ollama/issues/6890/events | https://github.com/ollama/ollama/issues/6890 | 2,538,032,802 | I_kwDOJ0Z1Ps6XR06i | 6,890 | "/show parameters" command causes crashes when running Qwen 2.5 models, on version 0.3.11 | {
"login": "cmdntfnd",
"id": 142593649,
"node_id": "U_kgDOCH_OcQ",
"avatar_url": "https://avatars.githubusercontent.com/u/142593649?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/cmdntfnd",
"html_url": "https://github.com/cmdntfnd",
"followers_url": "https://api.github.com/users/cmdntfnd/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 1 | 2024-09-20T06:54:55 | 2024-12-13T17:07:10 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
This only happens after changing the parameters through /set parameter command.
Here's an example:
PS H:\ztmp> ollama run qwen2.5
>>> /show parameters
No parameters were specified for this model.
>>> /set parameter top_k 1
Set parameter 'top_k' to '1'
>>> /show parameters
error: couldn't... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6890/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6890/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/1020 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1020/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1020/comments | https://api.github.com/repos/ollama/ollama/issues/1020/events | https://github.com/ollama/ollama/pull/1020 | 1,979,730,196 | PR_kwDOJ0Z1Ps5euED0 | 1,020 | Add ModelFusion community integration | {
"login": "lgrammel",
"id": 205036,
"node_id": "MDQ6VXNlcjIwNTAzNg==",
"avatar_url": "https://avatars.githubusercontent.com/u/205036?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lgrammel",
"html_url": "https://github.com/lgrammel",
"followers_url": "https://api.github.com/users/lgramme... | [] | closed | false | null | [] | null | 0 | 2023-11-06T17:54:56 | 2023-11-06T18:46:16 | 2023-11-06T18:46:16 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1020",
"html_url": "https://github.com/ollama/ollama/pull/1020",
"diff_url": "https://github.com/ollama/ollama/pull/1020.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1020.patch",
"merged_at": "2023-11-06T18:46:16"
} | null | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1020/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1020/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5656 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5656/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5656/comments | https://api.github.com/repos/ollama/ollama/issues/5656/events | https://github.com/ollama/ollama/pull/5656 | 2,406,435,963 | PR_kwDOJ0Z1Ps51RF2g | 5,656 | llm: reorder gguf tensors | {
"login": "joshyan1",
"id": 76125168,
"node_id": "MDQ6VXNlcjc2MTI1MTY4",
"avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joshyan1",
"html_url": "https://github.com/joshyan1",
"followers_url": "https://api.github.com/users/jos... | [] | closed | false | null | [] | null | 1 | 2024-07-12T22:38:11 | 2024-08-14T16:49:46 | 2024-08-14T16:49:45 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5656",
"html_url": "https://github.com/ollama/ollama/pull/5656",
"diff_url": "https://github.com/ollama/ollama/pull/5656.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5656.patch",
"merged_at": null
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5656/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5656/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5629 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5629/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5629/comments | https://api.github.com/repos/ollama/ollama/issues/5629/events | https://github.com/ollama/ollama/issues/5629 | 2,403,190,147 | I_kwDOJ0Z1Ps6PPcWD | 5,629 | Crashing or gibberish output on 3x Radeon GPUs | {
"login": "darwinvelez58",
"id": 118543481,
"node_id": "U_kgDOBxDUeQ",
"avatar_url": "https://avatars.githubusercontent.com/u/118543481?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/darwinvelez58",
"html_url": "https://github.com/darwinvelez58",
"followers_url": "https://api.github.com/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6433346500,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf3UTxA... | open | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 22 | 2024-07-11T13:26:43 | 2024-11-27T22:33:13 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
While running mixtral:8x7b-instruct-v0.1-q4_K_M on my physical machine with x3 7900 XTX I got this error:
```
[root@5dc6ecf27031 /]# ollama run mixtral:8x7b-instruct-v0.1-q4_K_M
Error: llama runner process has terminated: signal: segmentation fault (core dumped)
[root@5dc6ecf27031 /]#
... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5629/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5629/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/1707 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1707/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1707/comments | https://api.github.com/repos/ollama/ollama/issues/1707/events | https://github.com/ollama/ollama/issues/1707 | 2,055,634,785 | I_kwDOJ0Z1Ps56hn9h | 1,707 | [Bug] Calling embedding endpoint within short time will receive `None` | {
"login": "samx81",
"id": 20315145,
"node_id": "MDQ6VXNlcjIwMzE1MTQ1",
"avatar_url": "https://avatars.githubusercontent.com/u/20315145?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/samx81",
"html_url": "https://github.com/samx81",
"followers_url": "https://api.github.com/users/samx81/fo... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2023-12-25T11:44:12 | 2024-05-06T23:44:40 | 2024-05-06T23:44:40 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I'm currently using `llama_index` to make LLM do a document QA,
but I notice the embedding endpoint often output:
```json
{"embedding":null}
```
when I do `condense_question` chat mode which calls LLM to rephrase question (and match context) before actually answer.
Like:
```python
llm = Ollama(model="dolphin... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1707/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1707/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7605 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7605/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7605/comments | https://api.github.com/repos/ollama/ollama/issues/7605/events | https://github.com/ollama/ollama/issues/7605 | 2,647,941,751 | I_kwDOJ0Z1Ps6d1GJ3 | 7,605 | [Model request]GOT-OCR2.0 | {
"login": "monkeycc",
"id": 6490927,
"node_id": "MDQ6VXNlcjY0OTA5Mjc=",
"avatar_url": "https://avatars.githubusercontent.com/u/6490927?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/monkeycc",
"html_url": "https://github.com/monkeycc",
"followers_url": "https://api.github.com/users/monke... | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | open | false | null | [] | null | 0 | 2024-11-11T02:30:06 | 2024-12-18T01:46:27 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | https://github.com/Ucas-HaoranWei/GOT-OCR2.0 | {
"login": "monkeycc",
"id": 6490927,
"node_id": "MDQ6VXNlcjY0OTA5Mjc=",
"avatar_url": "https://avatars.githubusercontent.com/u/6490927?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/monkeycc",
"html_url": "https://github.com/monkeycc",
"followers_url": "https://api.github.com/users/monke... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7605/reactions",
"total_count": 6,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 3,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7605/timeline | null | reopened | false |
https://api.github.com/repos/ollama/ollama/issues/7774 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7774/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7774/comments | https://api.github.com/repos/ollama/ollama/issues/7774/events | https://github.com/ollama/ollama/issues/7774 | 2,677,843,192 | I_kwDOJ0Z1Ps6fnKT4 | 7,774 | Toolcall stream | {
"login": "ZHOUxiaohe1987",
"id": 59469405,
"node_id": "MDQ6VXNlcjU5NDY5NDA1",
"avatar_url": "https://avatars.githubusercontent.com/u/59469405?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ZHOUxiaohe1987",
"html_url": "https://github.com/ZHOUxiaohe1987",
"followers_url": "https://api.gi... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 0 | 2024-11-21T03:21:31 | 2024-11-21T03:22:11 | 2024-11-21T03:21:38 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | null | {
"login": "ZHOUxiaohe1987",
"id": 59469405,
"node_id": "MDQ6VXNlcjU5NDY5NDA1",
"avatar_url": "https://avatars.githubusercontent.com/u/59469405?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ZHOUxiaohe1987",
"html_url": "https://github.com/ZHOUxiaohe1987",
"followers_url": "https://api.gi... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7774/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7774/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8597 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8597/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8597/comments | https://api.github.com/repos/ollama/ollama/issues/8597/events | https://github.com/ollama/ollama/issues/8597 | 2,811,701,408 | I_kwDOJ0Z1Ps6nlyig | 8,597 | Error: llama runner process has terminated: error loading model: unable to allocate CUDA0 buffer (4x L40S, 384GB system RAM, Deepseek-R1) | {
"login": "orlyandico",
"id": 1325420,
"node_id": "MDQ6VXNlcjEzMjU0MjA=",
"avatar_url": "https://avatars.githubusercontent.com/u/1325420?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/orlyandico",
"html_url": "https://github.com/orlyandico",
"followers_url": "https://api.github.com/users... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 17 | 2025-01-26T17:56:11 | 2025-01-30T07:47:39 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I am aware that 4 x L40S only has 192GB of VRAM, but the machine also has 384GB of system RAM. The error seems to indicate that 45108 MiB of RAM is being allocated with cudaMalloc and this is failing. This is very close to the GPU limit (46068 MiB). On my home setup (2x P40, admittedly not tryi... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8597/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8597/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/7270 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7270/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7270/comments | https://api.github.com/repos/ollama/ollama/issues/7270/events | https://github.com/ollama/ollama/pull/7270 | 2,599,159,642 | PR_kwDOJ0Z1Ps5_LPMN | 7,270 | Show supported types list in help message for `ollama create` | {
"login": "shirayu",
"id": 963961,
"node_id": "MDQ6VXNlcjk2Mzk2MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/963961?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/shirayu",
"html_url": "https://github.com/shirayu",
"followers_url": "https://api.github.com/users/shirayu/fo... | [] | closed | false | null | [] | null | 2 | 2024-10-19T13:15:48 | 2024-11-27T06:43:00 | 2024-11-27T06:43:00 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7270",
"html_url": "https://github.com/ollama/ollama/pull/7270",
"diff_url": "https://github.com/ollama/ollama/pull/7270.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7270.patch",
"merged_at": null
} | This PR adds supported types list to the help message for `ollama create`
```console
$ go run main.go create -h
Create a model from a Modelfile
Usage:
ollama create MODEL [flags]
Flags:
-f, --file string Name of the Modelfile (default "Modelfile")
-h, --help help for create
... | {
"login": "shirayu",
"id": 963961,
"node_id": "MDQ6VXNlcjk2Mzk2MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/963961?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/shirayu",
"html_url": "https://github.com/shirayu",
"followers_url": "https://api.github.com/users/shirayu/fo... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7270/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7270/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/478 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/478/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/478/comments | https://api.github.com/repos/ollama/ollama/issues/478/events | https://github.com/ollama/ollama/pull/478 | 1,884,825,856 | PR_kwDOJ0Z1Ps5ZufRs | 478 | remove unused openssh key types | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | [] | closed | false | null | [] | null | 1 | 2023-09-06T22:13:05 | 2023-09-06T22:18:55 | 2023-09-06T22:18:54 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/478",
"html_url": "https://github.com/ollama/ollama/pull/478",
"diff_url": "https://github.com/ollama/ollama/pull/478.diff",
"patch_url": "https://github.com/ollama/ollama/pull/478.patch",
"merged_at": "2023-09-06T22:18:54"
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/478/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/478/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7369 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7369/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7369/comments | https://api.github.com/repos/ollama/ollama/issues/7369/events | https://github.com/ollama/ollama/pull/7369 | 2,615,356,549 | PR_kwDOJ0Z1Ps5_94Uy | 7,369 | Fix deepseek deseret regex | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [] | closed | false | null | [] | null | 0 | 2024-10-25T23:37:50 | 2024-10-26T21:58:57 | 2024-10-26T21:58:54 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7369",
"html_url": "https://github.com/ollama/ollama/pull/7369",
"diff_url": "https://github.com/ollama/ollama/pull/7369.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7369.patch",
"merged_at": "2024-10-26T21:58:54"
} | On windows compiled with gcc the c++ regex library failed to handle the characters
Without any changes, loading the model in the Go server crashes with
```
llama_model_load: error loading model: error loading model vocabulary: wstring_convert::from_bytes
```
The patch for unicode.cpp gets past the wide charact... | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7369/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7369/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4416 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4416/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4416/comments | https://api.github.com/repos/ollama/ollama/issues/4416/events | https://github.com/ollama/ollama/pull/4416 | 2,294,175,355 | PR_kwDOJ0Z1Ps5vVSj2 | 4,416 | don't abort when an invalid model name is used in /save | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | [] | closed | false | null | [] | null | 0 | 2024-05-14T01:42:17 | 2024-05-14T01:48:29 | 2024-05-14T01:48:28 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4416",
"html_url": "https://github.com/ollama/ollama/pull/4416",
"diff_url": "https://github.com/ollama/ollama/pull/4416.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4416.patch",
"merged_at": "2024-05-14T01:48:28"
} | Fixes #3852 | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4416/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4416/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2774 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2774/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2774/comments | https://api.github.com/repos/ollama/ollama/issues/2774/events | https://github.com/ollama/ollama/issues/2774 | 2,155,807,968 | I_kwDOJ0Z1Ps6AfwTg | 2,774 | What is the different between /api/generate and /api/chat? | {
"login": "owenzhao",
"id": 2182896,
"node_id": "MDQ6VXNlcjIxODI4OTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/2182896?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/owenzhao",
"html_url": "https://github.com/owenzhao",
"followers_url": "https://api.github.com/users/owenz... | [] | closed | false | null | [] | null | 16 | 2024-02-27T06:53:37 | 2024-09-09T18:39:13 | 2024-03-12T06:33:16 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I mean if I give them the same prompt and input, the answers will be the same. Right? Then why they are two different API?
Or is chat auto context? I mean when using /api/chat, the answer will automatically include the previous conversation? And the /api/generate only answer for the present? | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2774/reactions",
"total_count": 14,
"+1": 14,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2774/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8232 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8232/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8232/comments | https://api.github.com/repos/ollama/ollama/issues/8232/events | https://github.com/ollama/ollama/pull/8232 | 2,757,927,665 | PR_kwDOJ0Z1Ps6GLO5h | 8,232 | Remove redundant image_url format check in fromChatRequest | {
"login": "muzzlol",
"id": 116030715,
"node_id": "U_kgDOBup8-w",
"avatar_url": "https://avatars.githubusercontent.com/u/116030715?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/muzzlol",
"html_url": "https://github.com/muzzlol",
"followers_url": "https://api.github.com/users/muzzlol/foll... | [] | closed | false | {
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/... | [
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "htt... | null | 1 | 2024-12-24T15:22:37 | 2025-01-13T19:32:37 | 2025-01-13T19:32:37 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8232",
"html_url": "https://github.com/ollama/ollama/pull/8232",
"diff_url": "https://github.com/ollama/ollama/pull/8232.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8232.patch",
"merged_at": null
} | Improves the `fromChatRequest` function by removing redundant checks for Ollama's direct string format and improving error messages for better debugging.
Summary of changes:
- Removed the redundant check for `image_url` in Ollama format.
- Clarified and enhanced error messages in the conversion logic. | {
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8232/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8232/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8604 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8604/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8604/comments | https://api.github.com/repos/ollama/ollama/issues/8604/events | https://github.com/ollama/ollama/pull/8604 | 2,812,382,560 | PR_kwDOJ0Z1Ps6JDQsP | 8,604 | readme: add AI Toolkit for VSCode to community integrations | {
"login": "XiaofuHuang",
"id": 49138419,
"node_id": "MDQ6VXNlcjQ5MTM4NDE5",
"avatar_url": "https://avatars.githubusercontent.com/u/49138419?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/XiaofuHuang",
"html_url": "https://github.com/XiaofuHuang",
"followers_url": "https://api.github.com/... | [] | closed | false | null | [] | null | 0 | 2025-01-27T08:34:26 | 2025-01-30T13:12:11 | 2025-01-27T08:36:23 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8604",
"html_url": "https://github.com/ollama/ollama/pull/8604",
"diff_url": "https://github.com/ollama/ollama/pull/8604.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8604.patch",
"merged_at": "2025-01-27T08:36:23"
} | Add [AI Toolkit for VSCode](https://marketplace.visualstudio.com/items?itemName=ms-windows-ai-studio.windows-ai-studio) to community integrations.
AI Toolkit for VSCode ollama support: https://aka.ms/ai-tooklit/ollama-docs | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8604/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8604/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2949 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2949/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2949/comments | https://api.github.com/repos/ollama/ollama/issues/2949/events | https://github.com/ollama/ollama/pull/2949 | 2,171,044,978 | PR_kwDOJ0Z1Ps5o0Jnk | 2,949 | add api.NewClient | {
"login": "jackielii",
"id": 360983,
"node_id": "MDQ6VXNlcjM2MDk4Mw==",
"avatar_url": "https://avatars.githubusercontent.com/u/360983?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jackielii",
"html_url": "https://github.com/jackielii",
"followers_url": "https://api.github.com/users/jack... | [] | closed | false | null | [] | null | 1 | 2024-03-06T09:31:58 | 2024-05-04T08:35:08 | 2024-05-04T08:35:07 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2949",
"html_url": "https://github.com/ollama/ollama/pull/2949",
"diff_url": "https://github.com/ollama/ollama/pull/2949.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2949.patch",
"merged_at": null
} | fix #2948
| {
"login": "jackielii",
"id": 360983,
"node_id": "MDQ6VXNlcjM2MDk4Mw==",
"avatar_url": "https://avatars.githubusercontent.com/u/360983?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jackielii",
"html_url": "https://github.com/jackielii",
"followers_url": "https://api.github.com/users/jack... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2949/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2949/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1374 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1374/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1374/comments | https://api.github.com/repos/ollama/ollama/issues/1374/events | https://github.com/ollama/ollama/issues/1374 | 2,024,279,782 | I_kwDOJ0Z1Ps54qA7m | 1,374 | Out of memory error on model that previously worked fine after update to version 0.1.13 | {
"login": "madsamjp",
"id": 49611363,
"node_id": "MDQ6VXNlcjQ5NjExMzYz",
"avatar_url": "https://avatars.githubusercontent.com/u/49611363?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/madsamjp",
"html_url": "https://github.com/madsamjp",
"followers_url": "https://api.github.com/users/mad... | [] | closed | false | null | [] | null | 27 | 2023-12-04T16:40:23 | 2024-01-08T21:42:02 | 2024-01-08T21:42:02 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I configured a model to run entirely in VRAM using the following Modelfile:
```
FROM deepseek-coder:33b-instruct-q5_K_S
PARAMETER num_gpu 65
PARAMETER num_ctx 2048
```
I had no issues with running this, it would use about 22GB of my 4090's 24GB VRAM without issue. It would generate responses very quickly whic... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1374/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1374/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/938 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/938/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/938/comments | https://api.github.com/repos/ollama/ollama/issues/938/events | https://github.com/ollama/ollama/issues/938 | 1,966,217,268 | I_kwDOJ0Z1Ps51Mhg0 | 938 | CLI windows size detection not accurate on WSL | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/Br... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 0 | 2023-10-27T21:33:22 | 2024-02-20T00:56:13 | 2024-02-20T00:56:13 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | When backspacing characters from a long input on wsl a "float extra character" may not be deleted.
Reproduce by going to a newline and backspacing from the last character.
<img width="717" alt="Screenshot 2023-10-27 at 17 30 48" src="https://github.com/jmorganca/ollama/assets/5853428/9e18a84b-3071-4b67-aaba-43f2... | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmor... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/938/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/938/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3458 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3458/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3458/comments | https://api.github.com/repos/ollama/ollama/issues/3458/events | https://github.com/ollama/ollama/pull/3458 | 2,220,673,215 | PR_kwDOJ0Z1Ps5rcjJg | 3,458 | fix model needLoad always be true | {
"login": "mofanke",
"id": 54242816,
"node_id": "MDQ6VXNlcjU0MjQyODE2",
"avatar_url": "https://avatars.githubusercontent.com/u/54242816?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mofanke",
"html_url": "https://github.com/mofanke",
"followers_url": "https://api.github.com/users/mofank... | [] | closed | false | null | [] | null | 1 | 2024-04-02T14:26:53 | 2024-04-03T06:05:55 | 2024-04-03T06:05:54 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3458",
"html_url": "https://github.com/ollama/ollama/pull/3458",
"diff_url": "https://github.com/ollama/ollama/pull/3458.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3458.patch",
"merged_at": null
} | 
loaded.Options.Runner.NumGPU was change to 33 or other value by
``
func New(model string, adapters, projectors []string, opts *api.Options)
``
so `!reflect.DeepEqual(loaded.Options.Runner, opts.Runner)` will always b... | {
"login": "mofanke",
"id": 54242816,
"node_id": "MDQ6VXNlcjU0MjQyODE2",
"avatar_url": "https://avatars.githubusercontent.com/u/54242816?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mofanke",
"html_url": "https://github.com/mofanke",
"followers_url": "https://api.github.com/users/mofank... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3458/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3458/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/729 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/729/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/729/comments | https://api.github.com/repos/ollama/ollama/issues/729/events | https://github.com/ollama/ollama/issues/729 | 1,931,317,404 | I_kwDOJ0Z1Ps5zHZCc | 729 | Unable to pull models behind the proxy | {
"login": "ilyanoskov",
"id": 16443383,
"node_id": "MDQ6VXNlcjE2NDQzMzgz",
"avatar_url": "https://avatars.githubusercontent.com/u/16443383?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ilyanoskov",
"html_url": "https://github.com/ilyanoskov",
"followers_url": "https://api.github.com/use... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 13 | 2023-10-07T11:38:39 | 2024-10-01T11:45:58 | 2023-10-10T16:59:07 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Dear Maintainers,
Thank you very much for creating this project!
I need to set up ollama on Linux behind a proxy, and when pulling I get an error:
```download.go:166: couldn't download blob: Get "https:///...../ollama/docker/registry/v2/blobs/...": tls: first record does not look like a TLS handshake```
I h... | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/729/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/729/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4922 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4922/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4922/comments | https://api.github.com/repos/ollama/ollama/issues/4922/events | https://github.com/ollama/ollama/pull/4922 | 2,341,320,564 | PR_kwDOJ0Z1Ps5x2EeC | 4,922 | cmd: initial functionality for `ollama show <model>` | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjha... | [] | closed | false | null | [] | null | 1 | 2024-06-08T00:44:51 | 2024-06-15T03:03:48 | 2024-06-14T23:18:33 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | true | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4922",
"html_url": "https://github.com/ollama/ollama/pull/4922",
"diff_url": "https://github.com/ollama/ollama/pull/4922.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4922.patch",
"merged_at": null
} | null | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjha... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4922/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4922/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4631 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4631/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4631/comments | https://api.github.com/repos/ollama/ollama/issues/4631/events | https://github.com/ollama/ollama/issues/4631 | 2,316,815,049 | I_kwDOJ0Z1Ps6KF8rJ | 4,631 | Repeating answers in an instance. | {
"login": "Protostarship",
"id": 125519972,
"node_id": "U_kgDOB3tIZA",
"avatar_url": "https://avatars.githubusercontent.com/u/125519972?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Protostarship",
"html_url": "https://github.com/Protostarship",
"followers_url": "https://api.github.com/... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 1 | 2024-05-25T07:58:33 | 2024-05-29T22:52:59 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Hi, i'm using llama2 models, and when i asked the Ai to explain something, it does responds and explain it, but when it reached the end, it instead print out certain line,
```
[INST: <</INST]
```
then repeats another answer regarding the current questions.
### OS
Windows
### GPU
Nvidi... | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4631/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4631/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/6263 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6263/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6263/comments | https://api.github.com/repos/ollama/ollama/issues/6263/events | https://github.com/ollama/ollama/issues/6263 | 2,456,652,905 | I_kwDOJ0Z1Ps6SbYxp | 6,263 | Pull Command Parsing Not Working | {
"login": "chadwickhar08",
"id": 23239905,
"node_id": "MDQ6VXNlcjIzMjM5OTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/23239905?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/chadwickhar08",
"html_url": "https://github.com/chadwickhar08",
"followers_url": "https://api.githu... | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg... | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.gi... | null | 9 | 2024-08-08T21:27:30 | 2024-08-10T00:18:27 | 2024-08-09T19:30:26 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
When running ollama on Windows, attempt to run 'ollama pull llama3.1' results in 'ollama pull llama3.1
pulling manifest
Error: Incorrect function.'
### OS
Windows
### GPU
Nvidia
### CPU
Intel
### Ollama version
ollama --version ollama version is 0.3.4 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhilt... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6263/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6263/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6412 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6412/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6412/comments | https://api.github.com/repos/ollama/ollama/issues/6412/events | https://github.com/ollama/ollama/issues/6412 | 2,472,517,560 | I_kwDOJ0Z1Ps6TX5-4 | 6,412 | for glm4-9b | {
"login": "lcck200121",
"id": 126126214,
"node_id": "U_kgDOB4SIhg",
"avatar_url": "https://avatars.githubusercontent.com/u/126126214?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lcck200121",
"html_url": "https://github.com/lcck200121",
"followers_url": "https://api.github.com/users/lcc... | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2024-08-19T05:54:49 | 2024-08-19T05:54:49 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Do you have any plans to support the tool-calling function of glm4-9b? | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6412/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6412/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/4194 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4194/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4194/comments | https://api.github.com/repos/ollama/ollama/issues/4194/events | https://github.com/ollama/ollama/pull/4194 | 2,280,062,512 | PR_kwDOJ0Z1Ps5umAi_ | 4,194 | docs: delete log `HEAD` | {
"login": "HydenLiu",
"id": 52030677,
"node_id": "MDQ6VXNlcjUyMDMwNjc3",
"avatar_url": "https://avatars.githubusercontent.com/u/52030677?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/HydenLiu",
"html_url": "https://github.com/HydenLiu",
"followers_url": "https://api.github.com/users/Hyd... | [] | closed | false | null | [] | null | 0 | 2024-05-06T05:28:47 | 2024-05-06T17:32:31 | 2024-05-06T17:32:31 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4194",
"html_url": "https://github.com/ollama/ollama/pull/4194",
"diff_url": "https://github.com/ollama/ollama/pull/4194.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4194.patch",
"merged_at": "2024-05-06T17:32:31"
} | null | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4194/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4194/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1804 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1804/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1804/comments | https://api.github.com/repos/ollama/ollama/issues/1804/events | https://github.com/ollama/ollama/issues/1804 | 2,067,249,992 | I_kwDOJ0Z1Ps57N7tI | 1,804 | Code view on codellama vs phi and dolphin-phi | {
"login": "oliverbob",
"id": 23272429,
"node_id": "MDQ6VXNlcjIzMjcyNDI5",
"avatar_url": "https://avatars.githubusercontent.com/u/23272429?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/oliverbob",
"html_url": "https://github.com/oliverbob",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 2 | 2024-01-05T12:19:46 | 2024-03-13T23:53:54 | 2024-03-13T23:53:54 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Dolphin phi and (probably phi code indent):


Phi, indents but has no code view:
![image](https://github.com/jmo... | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/follower... | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1804/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1804/timeline | null | completed | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.