url
stringlengths 51
54
| repository_url
stringclasses 1
value | labels_url
stringlengths 65
68
| comments_url
stringlengths 60
63
| events_url
stringlengths 58
61
| html_url
stringlengths 39
44
| id
int64 1.78B
2.82B
| node_id
stringlengths 18
19
| number
int64 1
8.69k
| title
stringlengths 1
382
| user
dict | labels
listlengths 0
5
| state
stringclasses 2
values | locked
bool 1
class | assignee
dict | assignees
listlengths 0
2
| milestone
null | comments
int64 0
323
| created_at
timestamp[s] | updated_at
timestamp[s] | closed_at
timestamp[s] | author_association
stringclasses 4
values | sub_issues_summary
dict | active_lock_reason
null | draft
bool 2
classes | pull_request
dict | body
stringlengths 2
118k
⌀ | closed_by
dict | reactions
dict | timeline_url
stringlengths 60
63
| performed_via_github_app
null | state_reason
stringclasses 4
values | is_pull_request
bool 2
classes |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/ollama/ollama/issues/6525 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6525/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6525/comments | https://api.github.com/repos/ollama/ollama/issues/6525/events | https://github.com/ollama/ollama/issues/6525 | 2,488,468,101 | I_kwDOJ0Z1Ps6UUwKF | 6,525 | ollama collapses CPU | {
"login": "Hyphaed",
"id": 19622367,
"node_id": "MDQ6VXNlcjE5NjIyMzY3",
"avatar_url": "https://avatars.githubusercontent.com/u/19622367?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Hyphaed",
"html_url": "https://github.com/Hyphaed",
"followers_url": "https://api.github.com/users/Hyphaed/followers",
"following_url": "https://api.github.com/users/Hyphaed/following{/other_user}",
"gists_url": "https://api.github.com/users/Hyphaed/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Hyphaed/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Hyphaed/subscriptions",
"organizations_url": "https://api.github.com/users/Hyphaed/orgs",
"repos_url": "https://api.github.com/users/Hyphaed/repos",
"events_url": "https://api.github.com/users/Hyphaed/events{/privacy}",
"received_events_url": "https://api.github.com/users/Hyphaed/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 9 | 2024-08-27T07:00:05 | 2024-09-16T20:41:39 | 2024-08-27T15:12:18 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
ollama collapses CPU
even when I stop the server the CPU still stuck from 75% to 90%
even when I do have an RTX 3070 and terminal is showind that is using the GPU
there is no error in terminal
I have no verbose since I forcelly shutedown the workstation
### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.3.6 | {
"login": "Hyphaed",
"id": 19622367,
"node_id": "MDQ6VXNlcjE5NjIyMzY3",
"avatar_url": "https://avatars.githubusercontent.com/u/19622367?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Hyphaed",
"html_url": "https://github.com/Hyphaed",
"followers_url": "https://api.github.com/users/Hyphaed/followers",
"following_url": "https://api.github.com/users/Hyphaed/following{/other_user}",
"gists_url": "https://api.github.com/users/Hyphaed/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Hyphaed/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Hyphaed/subscriptions",
"organizations_url": "https://api.github.com/users/Hyphaed/orgs",
"repos_url": "https://api.github.com/users/Hyphaed/repos",
"events_url": "https://api.github.com/users/Hyphaed/events{/privacy}",
"received_events_url": "https://api.github.com/users/Hyphaed/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6525/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6525/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3972 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3972/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3972/comments | https://api.github.com/repos/ollama/ollama/issues/3972/events | https://github.com/ollama/ollama/pull/3972 | 2,266,851,869 | PR_kwDOJ0Z1Ps5t5oCp | 3,972 | Add support for building on Windows ARM64 | {
"login": "hmartinez82",
"id": 1100440,
"node_id": "MDQ6VXNlcjExMDA0NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/1100440?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hmartinez82",
"html_url": "https://github.com/hmartinez82",
"followers_url": "https://api.github.com/users/hmartinez82/followers",
"following_url": "https://api.github.com/users/hmartinez82/following{/other_user}",
"gists_url": "https://api.github.com/users/hmartinez82/gists{/gist_id}",
"starred_url": "https://api.github.com/users/hmartinez82/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/hmartinez82/subscriptions",
"organizations_url": "https://api.github.com/users/hmartinez82/orgs",
"repos_url": "https://api.github.com/users/hmartinez82/repos",
"events_url": "https://api.github.com/users/hmartinez82/events{/privacy}",
"received_events_url": "https://api.github.com/users/hmartinez82/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 7 | 2024-04-27T05:38:44 | 2024-05-08T04:05:14 | 2024-04-28T21:52:59 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3972",
"html_url": "https://github.com/ollama/ollama/pull/3972",
"diff_url": "https://github.com/ollama/ollama/pull/3972.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3972.patch",
"merged_at": "2024-04-28T21:52:58"
} | Part of #2589
- Builds only the cpu runner for ARM64 Also, the existing CMake recipe already enables NEON and Armv8.2 extensions when ARM64 is detected.
- I'll create another PR with build instructions. The main trick is that MSY2 has the CLANGARM64 environment that provides gcc aliases to Clang.
Maintainer changes are enabled. Feel free to adjust the changes to your liking. | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3972/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3972/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5552 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5552/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5552/comments | https://api.github.com/repos/ollama/ollama/issues/5552/events | https://github.com/ollama/ollama/pull/5552 | 2,396,671,216 | PR_kwDOJ0Z1Ps50wDvt | 5,552 | docs | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2024-07-08T22:17:57 | 2024-07-25T23:26:21 | 2024-07-25T23:26:19 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5552",
"html_url": "https://github.com/ollama/ollama/pull/5552",
"diff_url": "https://github.com/ollama/ollama/pull/5552.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5552.patch",
"merged_at": "2024-07-25T23:26:19"
} | part of #5216
part of #5284
part of #5207 | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5552/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5552/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/422 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/422/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/422/comments | https://api.github.com/repos/ollama/ollama/issues/422/events | https://github.com/ollama/ollama/issues/422 | 1,868,154,838 | I_kwDOJ0Z1Ps5vWcfW | 422 | `Error: Post "http://localhost:11434/api/generate": EOF` with long propmts with phind-codellama | {
"login": "tomduncalf",
"id": 5458070,
"node_id": "MDQ6VXNlcjU0NTgwNzA=",
"avatar_url": "https://avatars.githubusercontent.com/u/5458070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tomduncalf",
"html_url": "https://github.com/tomduncalf",
"followers_url": "https://api.github.com/users/tomduncalf/followers",
"following_url": "https://api.github.com/users/tomduncalf/following{/other_user}",
"gists_url": "https://api.github.com/users/tomduncalf/gists{/gist_id}",
"starred_url": "https://api.github.com/users/tomduncalf/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/tomduncalf/subscriptions",
"organizations_url": "https://api.github.com/users/tomduncalf/orgs",
"repos_url": "https://api.github.com/users/tomduncalf/repos",
"events_url": "https://api.github.com/users/tomduncalf/events{/privacy}",
"received_events_url": "https://api.github.com/users/tomduncalf/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 20 | 2023-08-26T16:00:33 | 2024-04-22T09:12:40 | 2023-09-07T11:08:40 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | It seems like if you provide a long prompt (I was using one of 1,000ish tokens according to OpenAI tokenizer) with this model, you get an error `Error: Post "http://localhost:11434/api/generate": EOF`. It may or may not relate to the contents of the prompt as well as the length | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/422/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/422/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7405 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7405/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7405/comments | https://api.github.com/repos/ollama/ollama/issues/7405/events | https://github.com/ollama/ollama/issues/7405 | 2,619,529,082 | I_kwDOJ0Z1Ps6cItd6 | 7,405 | Feature request: Add CLI argument to specify a system prompt | {
"login": "Kerrick",
"id": 552093,
"node_id": "MDQ6VXNlcjU1MjA5Mw==",
"avatar_url": "https://avatars.githubusercontent.com/u/552093?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Kerrick",
"html_url": "https://github.com/Kerrick",
"followers_url": "https://api.github.com/users/Kerrick/followers",
"following_url": "https://api.github.com/users/Kerrick/following{/other_user}",
"gists_url": "https://api.github.com/users/Kerrick/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Kerrick/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Kerrick/subscriptions",
"organizations_url": "https://api.github.com/users/Kerrick/orgs",
"repos_url": "https://api.github.com/users/Kerrick/repos",
"events_url": "https://api.github.com/users/Kerrick/events{/privacy}",
"received_events_url": "https://api.github.com/users/Kerrick/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 1 | 2024-10-28T20:54:54 | 2024-10-29T21:48:49 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I'd like to be able to set the system prompt from the call to `ollama` in my shell, rather than in the conversation. For example:
```sh
ollama run llama3.1 --system="Your nickname is 'Grass' now"
```
...or...
```sh
ollama run llama3.1 -s "system" "Your nickname is 'Grass' now"
```
With this ability, I could set up aliases in my shell profile so that I can run system-prompt-customized versions of a model with a single command.
```sh
alias grass="ollama run llama3.1 --system=\"Your nickname is 'Grass' now\""
```
It'd be even better if I could also specify a path to a text file with a system prompt (e.g. `ollama run llama3.1 --system-file="~/system_prompts/grass.txt"`, but that wouldn't be necessary.
This is unique from #807, whose fix only works in the conversation. | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7405/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7405/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/7620 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7620/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7620/comments | https://api.github.com/repos/ollama/ollama/issues/7620/events | https://github.com/ollama/ollama/pull/7620 | 2,649,452,239 | PR_kwDOJ0Z1Ps6BhB1y | 7,620 | api: fix typo in Golang API types docs | {
"login": "neomantra",
"id": 26842,
"node_id": "MDQ6VXNlcjI2ODQy",
"avatar_url": "https://avatars.githubusercontent.com/u/26842?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/neomantra",
"html_url": "https://github.com/neomantra",
"followers_url": "https://api.github.com/users/neomantra/followers",
"following_url": "https://api.github.com/users/neomantra/following{/other_user}",
"gists_url": "https://api.github.com/users/neomantra/gists{/gist_id}",
"starred_url": "https://api.github.com/users/neomantra/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/neomantra/subscriptions",
"organizations_url": "https://api.github.com/users/neomantra/orgs",
"repos_url": "https://api.github.com/users/neomantra/repos",
"events_url": "https://api.github.com/users/neomantra/events{/privacy}",
"received_events_url": "https://api.github.com/users/neomantra/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2024-11-11T14:10:40 | 2024-12-08T17:32:55 | 2024-11-12T00:21:58 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7620",
"html_url": "https://github.com/ollama/ollama/pull/7620",
"diff_url": "https://github.com/ollama/ollama/pull/7620.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7620.patch",
"merged_at": "2024-11-12T00:21:58"
} | Fixes minor typos and grammar in `api/types.go`
I had only reviewed `client.go` in my commit yesterday, sorry I didn't check this one too. Somehow the last PR had Python in the title, but it and this affect Golang. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7620/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7620/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8485 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8485/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8485/comments | https://api.github.com/repos/ollama/ollama/issues/8485/events | https://github.com/ollama/ollama/issues/8485 | 2,797,657,711 | I_kwDOJ0Z1Ps6mwN5v | 8,485 | [0.5.7] small models are loaded to GPU, but inference is slow and using a lot of CPU | {
"login": "kha84",
"id": 110789576,
"node_id": "U_kgDOBpqDyA",
"avatar_url": "https://avatars.githubusercontent.com/u/110789576?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kha84",
"html_url": "https://github.com/kha84",
"followers_url": "https://api.github.com/users/kha84/followers",
"following_url": "https://api.github.com/users/kha84/following{/other_user}",
"gists_url": "https://api.github.com/users/kha84/gists{/gist_id}",
"starred_url": "https://api.github.com/users/kha84/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/kha84/subscriptions",
"organizations_url": "https://api.github.com/users/kha84/orgs",
"repos_url": "https://api.github.com/users/kha84/repos",
"events_url": "https://api.github.com/users/kha84/events{/privacy}",
"received_events_url": "https://api.github.com/users/kha84/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 6 | 2025-01-19T14:33:04 | 2025-01-20T09:50:51 | 2025-01-20T09:50:51 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Hello there.
Just upgraded from ollama 0.4.x version to the latest one (0.5.7) and immediately noticed that inference of all models (even small ones, like llama 3.2 3B) become very slow. Like orders of magnitude slow. I can see that during inference CPU is being used intensively, even though the model itself is loaded into VRAM of GPU and there's a lot of VRAM is still free (as per nvtop).
OS Ubuntu 22 LTS
RTX 4090
### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.5.7 | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8485/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8485/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4145 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4145/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4145/comments | https://api.github.com/repos/ollama/ollama/issues/4145/events | https://github.com/ollama/ollama/pull/4145 | 2,278,580,561 | PR_kwDOJ0Z1Ps5uhXWg | 4,145 | Fix lint warnings | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2024-05-03T23:44:35 | 2024-05-03T23:53:20 | 2024-05-03T23:53:17 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4145",
"html_url": "https://github.com/ollama/ollama/pull/4145",
"diff_url": "https://github.com/ollama/ollama/pull/4145.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4145.patch",
"merged_at": "2024-05-03T23:53:17"
} | null | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4145/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4145/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/290 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/290/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/290/comments | https://api.github.com/repos/ollama/ollama/issues/290/events | https://github.com/ollama/ollama/pull/290 | 1,837,466,485 | PR_kwDOJ0Z1Ps5XPFzC | 290 | implement loading ggml lora adapters through the modelfile | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2023-08-05T00:21:45 | 2023-08-11T00:23:03 | 2023-08-11T00:23:01 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/290",
"html_url": "https://github.com/ollama/ollama/pull/290",
"diff_url": "https://github.com/ollama/ollama/pull/290.diff",
"patch_url": "https://github.com/ollama/ollama/pull/290.patch",
"merged_at": "2023-08-11T00:23:01"
} | LoRA adapters can be added to Ollama models through the Modelfile and automatically applied when the model is loaded:
```
FROM llama2:13b
TEMPLATE {{ .Prompt }}
ADAPTER ./llama2-13b-storywriter-lora.ggml.bin
```
A few caveats:
* LoRA adapters must be GGML. If the adapter isn't GGML, it can be converted with the `convert-lora-to-ggml.py` script in https://github.com/ggerganov/llama.cpp
* Using adapters with quantized weights might not produce good results
* Using adapters disables mmap
* It's possible to apply multiple adapters but
1. Ordering is important
2. There may be unintended side effects since most adapters are not intended to be layered on other adapters
3. Performance may degrade with more adapters | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/290/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/290/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6651 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6651/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6651/comments | https://api.github.com/repos/ollama/ollama/issues/6651/events | https://github.com/ollama/ollama/issues/6651 | 2,507,064,853 | I_kwDOJ0Z1Ps6VbsYV | 6,651 | The speed of using embedded models is much slower compared to xinference | {
"login": "yushengliao",
"id": 29765903,
"node_id": "MDQ6VXNlcjI5NzY1OTAz",
"avatar_url": "https://avatars.githubusercontent.com/u/29765903?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yushengliao",
"html_url": "https://github.com/yushengliao",
"followers_url": "https://api.github.com/users/yushengliao/followers",
"following_url": "https://api.github.com/users/yushengliao/following{/other_user}",
"gists_url": "https://api.github.com/users/yushengliao/gists{/gist_id}",
"starred_url": "https://api.github.com/users/yushengliao/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yushengliao/subscriptions",
"organizations_url": "https://api.github.com/users/yushengliao/orgs",
"repos_url": "https://api.github.com/users/yushengliao/repos",
"events_url": "https://api.github.com/users/yushengliao/events{/privacy}",
"received_events_url": "https://api.github.com/users/yushengliao/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 5808482718,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWjZpng",
"url": "https://api.github.com/repos/ollama/ollama/labels/performance",
"name": "performance",
"color": "A5B5C6",
"default": false,
"description": ""
}
] | open | false | null | [] | null | 0 | 2024-09-05T07:58:41 | 2024-09-05T16:17:42 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I use the BGE-M3 model and send the same request, especially with xinference taking about 10 seconds and ollama taking about 200 seconds.
I'm sure they all use GPUs.
I found that xinference allocates more video memory, while ollama's video memory usage remains basically unchanged. Perhaps this is the reason for the speed difference? | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6651/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/ollama/ollama/issues/6651/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/891 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/891/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/891/comments | https://api.github.com/repos/ollama/ollama/issues/891/events | https://github.com/ollama/ollama/issues/891 | 1,959,768,778 | I_kwDOJ0Z1Ps50z7LK | 891 | Support remote `ollama create` | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 0 | 2023-10-24T17:47:30 | 2023-11-16T00:41:14 | 2023-11-16T00:41:14 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | `ollama create` should support remote instances of Ollama with `OLLAMA_HOST`
```
OLLAMA_HOST=my-test-host:11434 ollama create my-model -f ./Modelfile
``` | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/891/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/891/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3323 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3323/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3323/comments | https://api.github.com/repos/ollama/ollama/issues/3323/events | https://github.com/ollama/ollama/issues/3323 | 2,204,255,976 | I_kwDOJ0Z1Ps6DYkbo | 3,323 | Feat req: Add "Last updated" sorting to the hub | {
"login": "knoopx",
"id": 100993,
"node_id": "MDQ6VXNlcjEwMDk5Mw==",
"avatar_url": "https://avatars.githubusercontent.com/u/100993?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/knoopx",
"html_url": "https://github.com/knoopx",
"followers_url": "https://api.github.com/users/knoopx/followers",
"following_url": "https://api.github.com/users/knoopx/following{/other_user}",
"gists_url": "https://api.github.com/users/knoopx/gists{/gist_id}",
"starred_url": "https://api.github.com/users/knoopx/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/knoopx/subscriptions",
"organizations_url": "https://api.github.com/users/knoopx/orgs",
"repos_url": "https://api.github.com/users/knoopx/repos",
"events_url": "https://api.github.com/users/knoopx/events{/privacy}",
"received_events_url": "https://api.github.com/users/knoopx/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 6573197867,
"node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw",
"url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com",
"name": "ollama.com",
"color": "ffffff",
"default": false,
"description": ""
}
] | closed | false | null | [] | null | 1 | 2024-03-24T09:09:12 | 2024-07-18T19:04:28 | 2024-07-18T19:04:28 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What are you trying to do?
There's no way to discover recent updates for existing models.
### How should we solve this?
Add "Last updated" sort choice
### What is the impact of not solving this?
No way to find out recently updated models
### Anything else?
_No response_ | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3323/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3323/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7820 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7820/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7820/comments | https://api.github.com/repos/ollama/ollama/issues/7820/events | https://github.com/ollama/ollama/issues/7820 | 2,687,973,667 | I_kwDOJ0Z1Ps6gNzkj | 7,820 | Instant closure when using shell input with piped output. | {
"login": "WyvernDotRed",
"id": 41121402,
"node_id": "MDQ6VXNlcjQxMTIxNDAy",
"avatar_url": "https://avatars.githubusercontent.com/u/41121402?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/WyvernDotRed",
"html_url": "https://github.com/WyvernDotRed",
"followers_url": "https://api.github.com/users/WyvernDotRed/followers",
"following_url": "https://api.github.com/users/WyvernDotRed/following{/other_user}",
"gists_url": "https://api.github.com/users/WyvernDotRed/gists{/gist_id}",
"starred_url": "https://api.github.com/users/WyvernDotRed/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/WyvernDotRed/subscriptions",
"organizations_url": "https://api.github.com/users/WyvernDotRed/orgs",
"repos_url": "https://api.github.com/users/WyvernDotRed/repos",
"events_url": "https://api.github.com/users/WyvernDotRed/events{/privacy}",
"received_events_url": "https://api.github.com/users/WyvernDotRed/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info",
"name": "needs more info",
"color": "BA8041",
"default": false,
"description": "More information is needed to assist"
}
] | open | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 6 | 2024-11-24T16:36:36 | 2024-12-10T21:07:24 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
When running `ollama run [model] | cat` or `ollama run [model] > [file]`, ollama now closes immediately and does not accept any manual input.
`ollama run [model]` still functions as expected.
While `cat | ollama run [model] ...` seems to be the workaround, this requires entering ^d to have the input processed.
Thus closing the input stream and also ollama as a result, expected in this case since input was explicitly closed.
Additionally, thanks to [pull 416](https://github.com/ollama/ollama/pull/416), this removes the ability to queue or follow up with further prompts.
This regression seems specific to 0.4.4.
After downgrading to 0.4.3 or 0.4.2, ollama functions as I expect.
I have not tested older versions and will stick to version 0.4.3 for the time being.
The purpose of the above syntax being to use the ability of the chat functionality for to enter multiple prompts while processing the output by a further script.
Though this will be covered in a following feature request and is only tangentially related to this regression.
Thank you for your time, have a nice rest of your day!
### OS
Linux
### GPU
AMD
### CPU
AMD
### Ollama version
0.4.4 | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7820/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7820/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/7239 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7239/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7239/comments | https://api.github.com/repos/ollama/ollama/issues/7239/events | https://github.com/ollama/ollama/issues/7239 | 2,594,416,445 | I_kwDOJ0Z1Ps6ao6c9 | 7,239 | Add Tab-Enabled Autocomplete for Local Model Parameters in Ollama CLI | {
"login": "lucianoayres",
"id": 20209393,
"node_id": "MDQ6VXNlcjIwMjA5Mzkz",
"avatar_url": "https://avatars.githubusercontent.com/u/20209393?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lucianoayres",
"html_url": "https://github.com/lucianoayres",
"followers_url": "https://api.github.com/users/lucianoayres/followers",
"following_url": "https://api.github.com/users/lucianoayres/following{/other_user}",
"gists_url": "https://api.github.com/users/lucianoayres/gists{/gist_id}",
"starred_url": "https://api.github.com/users/lucianoayres/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lucianoayres/subscriptions",
"organizations_url": "https://api.github.com/users/lucianoayres/orgs",
"repos_url": "https://api.github.com/users/lucianoayres/repos",
"events_url": "https://api.github.com/users/lucianoayres/events{/privacy}",
"received_events_url": "https://api.github.com/users/lucianoayres/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 3 | 2024-10-17T11:20:14 | 2025-01-13T00:46:55 | 2025-01-13T00:46:55 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | It would greatly enhance usability if the Ollama CLI supported tab-autocomplete for model names when using commands like `run`, `show`, `list`, etc. For example:
```bash
# This would autocomplete to something like `llama3.2`, based on the locally available models.
ollama run lla<TAB>
```
Implementing this across all commands where model names are parameters would improve efficiency and user experience.
Thank you! | {
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7239/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7239/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2138 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2138/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2138/comments | https://api.github.com/repos/ollama/ollama/issues/2138/events | https://github.com/ollama/ollama/pull/2138 | 2,094,374,620 | PR_kwDOJ0Z1Ps5kvv-F | 2,138 | Update README.md - Community Integrations - Obsidian Local GPT plugin | {
"login": "pfrankov",
"id": 584632,
"node_id": "MDQ6VXNlcjU4NDYzMg==",
"avatar_url": "https://avatars.githubusercontent.com/u/584632?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pfrankov",
"html_url": "https://github.com/pfrankov",
"followers_url": "https://api.github.com/users/pfrankov/followers",
"following_url": "https://api.github.com/users/pfrankov/following{/other_user}",
"gists_url": "https://api.github.com/users/pfrankov/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pfrankov/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pfrankov/subscriptions",
"organizations_url": "https://api.github.com/users/pfrankov/orgs",
"repos_url": "https://api.github.com/users/pfrankov/repos",
"events_url": "https://api.github.com/users/pfrankov/events{/privacy}",
"received_events_url": "https://api.github.com/users/pfrankov/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2024-01-22T17:11:44 | 2024-02-22T15:52:36 | 2024-02-22T15:52:36 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2138",
"html_url": "https://github.com/ollama/ollama/pull/2138",
"diff_url": "https://github.com/ollama/ollama/pull/2138.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2138.patch",
"merged_at": "2024-02-22T15:52:36"
} | Local GPT plugin for Obsidian mainly relies on Ollama provider

Also works with images
<img width="400" src="https://github.com/pfrankov/obsidian-local-gpt/assets/584632/a05d68fa-5419-4386-ac43-82b9513999ad">
<img width="598" alt="Settings" src="https://github.com/pfrankov/obsidian-local-gpt/assets/584632/6ab2d802-13ed-42be-aab1-6a3f689b18a0">
I'd say that Local GPT plugin is enhanced version of [Obsidian Ollama plugin](https://github.com/hinterdupfinger/obsidian-ollama) in every way.
Tried to resolve merge conflicts in https://github.com/jmorganca/ollama/pull/1662 but accidentally closed it. | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2138/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2138/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5783 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5783/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5783/comments | https://api.github.com/repos/ollama/ollama/issues/5783/events | https://github.com/ollama/ollama/issues/5783 | 2,417,335,104 | I_kwDOJ0Z1Ps6QFZtA | 5,783 | erorr loading models x3 7900 XTX #5708 | {
"login": "darwinvelez58",
"id": 118543481,
"node_id": "U_kgDOBxDUeQ",
"avatar_url": "https://avatars.githubusercontent.com/u/118543481?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/darwinvelez58",
"html_url": "https://github.com/darwinvelez58",
"followers_url": "https://api.github.com/users/darwinvelez58/followers",
"following_url": "https://api.github.com/users/darwinvelez58/following{/other_user}",
"gists_url": "https://api.github.com/users/darwinvelez58/gists{/gist_id}",
"starred_url": "https://api.github.com/users/darwinvelez58/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/darwinvelez58/subscriptions",
"organizations_url": "https://api.github.com/users/darwinvelez58/orgs",
"repos_url": "https://api.github.com/users/darwinvelez58/repos",
"events_url": "https://api.github.com/users/darwinvelez58/events{/privacy}",
"received_events_url": "https://api.github.com/users/darwinvelez58/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2024-07-18T20:28:02 | 2024-07-22T23:07:52 | 2024-07-22T23:07:52 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Few Days ago I report this error #5708, #5710 suppose to fix the issue but I still have the same error.
### OS
Linux
### GPU
AMD
### CPU
AMD
### Ollama version
0.2.6-rocm | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5783/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5783/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1528 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1528/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1528/comments | https://api.github.com/repos/ollama/ollama/issues/1528/events | https://github.com/ollama/ollama/pull/1528 | 2,042,606,696 | PR_kwDOJ0Z1Ps5iDRhN | 1,528 | Add unit test of API routes | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2023-12-14T22:35:10 | 2023-12-15T00:47:41 | 2023-12-15T00:47:40 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1528",
"html_url": "https://github.com/ollama/ollama/pull/1528",
"diff_url": "https://github.com/ollama/ollama/pull/1528.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1528.patch",
"merged_at": "2023-12-15T00:47:40"
} | This change modifies the base server to allow it to be more easily unit tested. It also adds in a simple unit test to "/api/version" to demonstrate how to add unit tests in the future. | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1528/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1528/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5830 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5830/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5830/comments | https://api.github.com/repos/ollama/ollama/issues/5830/events | https://github.com/ollama/ollama/issues/5830 | 2,421,426,851 | I_kwDOJ0Z1Ps6QVAqj | 5,830 | OpenAI endpoint gives 404 | {
"login": "defaultsecurity",
"id": 34036534,
"node_id": "MDQ6VXNlcjM0MDM2NTM0",
"avatar_url": "https://avatars.githubusercontent.com/u/34036534?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/defaultsecurity",
"html_url": "https://github.com/defaultsecurity",
"followers_url": "https://api.github.com/users/defaultsecurity/followers",
"following_url": "https://api.github.com/users/defaultsecurity/following{/other_user}",
"gists_url": "https://api.github.com/users/defaultsecurity/gists{/gist_id}",
"starred_url": "https://api.github.com/users/defaultsecurity/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/defaultsecurity/subscriptions",
"organizations_url": "https://api.github.com/users/defaultsecurity/orgs",
"repos_url": "https://api.github.com/users/defaultsecurity/repos",
"events_url": "https://api.github.com/users/defaultsecurity/events{/privacy}",
"received_events_url": "https://api.github.com/users/defaultsecurity/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2024-07-21T14:20:07 | 2024-07-22T06:18:03 | 2024-07-22T06:18:03 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
- http://localhost:11434/v1/chat/completions (gives 404)
- http://localhost:11434 (shows ollama is running)
Otherwise Ollama is working. I'm not sure what to do.
### OS
Linux
### GPU
Nvidia
### CPU
AMD
### Ollama version
0.2.7 | {
"login": "defaultsecurity",
"id": 34036534,
"node_id": "MDQ6VXNlcjM0MDM2NTM0",
"avatar_url": "https://avatars.githubusercontent.com/u/34036534?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/defaultsecurity",
"html_url": "https://github.com/defaultsecurity",
"followers_url": "https://api.github.com/users/defaultsecurity/followers",
"following_url": "https://api.github.com/users/defaultsecurity/following{/other_user}",
"gists_url": "https://api.github.com/users/defaultsecurity/gists{/gist_id}",
"starred_url": "https://api.github.com/users/defaultsecurity/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/defaultsecurity/subscriptions",
"organizations_url": "https://api.github.com/users/defaultsecurity/orgs",
"repos_url": "https://api.github.com/users/defaultsecurity/repos",
"events_url": "https://api.github.com/users/defaultsecurity/events{/privacy}",
"received_events_url": "https://api.github.com/users/defaultsecurity/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5830/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5830/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2920 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2920/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2920/comments | https://api.github.com/repos/ollama/ollama/issues/2920/events | https://github.com/ollama/ollama/issues/2920 | 2,167,396,829 | I_kwDOJ0Z1Ps6BL9nd | 2,920 | ollama call failed with status code 500 llama 2 | {
"login": "sabahatza",
"id": 135341585,
"node_id": "U_kgDOCBEmEQ",
"avatar_url": "https://avatars.githubusercontent.com/u/135341585?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sabahatza",
"html_url": "https://github.com/sabahatza",
"followers_url": "https://api.github.com/users/sabahatza/followers",
"following_url": "https://api.github.com/users/sabahatza/following{/other_user}",
"gists_url": "https://api.github.com/users/sabahatza/gists{/gist_id}",
"starred_url": "https://api.github.com/users/sabahatza/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/sabahatza/subscriptions",
"organizations_url": "https://api.github.com/users/sabahatza/orgs",
"repos_url": "https://api.github.com/users/sabahatza/repos",
"events_url": "https://api.github.com/users/sabahatza/events{/privacy}",
"received_events_url": "https://api.github.com/users/sabahatza/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 3 | 2024-03-04T17:44:32 | 2024-03-04T17:57:26 | 2024-03-04T17:48:17 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi team,
I am trying to run the llama2 model locally ( I was doing it previously for the last couple of weeks without any problems), but now I face the following error when I am trying to -> ollama run llama2
`Error: error loading model /Users/S_Z/.ollama/models/blobs/sha256:8934d96d3f08982e95922b2b7a2c626a1fe873d7c3b06e8e56d7bc0a1`
No ollama models are running for that matter. for context, I have an M1 Max Macbook. Is there any fix to this? | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2920/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2920/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/601 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/601/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/601/comments | https://api.github.com/repos/ollama/ollama/issues/601/events | https://github.com/ollama/ollama/pull/601 | 1,912,778,743 | PR_kwDOJ0Z1Ps5bMTBe | 601 | Update README.md for linux + cleanup | {
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/users/mchiang0610/followers",
"following_url": "https://api.github.com/users/mchiang0610/following{/other_user}",
"gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions",
"organizations_url": "https://api.github.com/users/mchiang0610/orgs",
"repos_url": "https://api.github.com/users/mchiang0610/repos",
"events_url": "https://api.github.com/users/mchiang0610/events{/privacy}",
"received_events_url": "https://api.github.com/users/mchiang0610/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2023-09-26T06:30:46 | 2023-09-26T06:44:54 | 2023-09-26T06:44:53 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/601",
"html_url": "https://github.com/ollama/ollama/pull/601",
"diff_url": "https://github.com/ollama/ollama/pull/601.diff",
"patch_url": "https://github.com/ollama/ollama/pull/601.patch",
"merged_at": "2023-09-26T06:44:53"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/601/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/601/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5208 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5208/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5208/comments | https://api.github.com/repos/ollama/ollama/issues/5208/events | https://github.com/ollama/ollama/pull/5208 | 2,367,386,322 | PR_kwDOJ0Z1Ps5zOuoI | 5,208 | Support image input for OpenAI chat compatibility | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 8 | 2024-06-22T00:31:49 | 2024-07-30T20:10:05 | 2024-07-14T05:07:45 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5208",
"html_url": "https://github.com/ollama/ollama/pull/5208",
"diff_url": "https://github.com/ollama/ollama/pull/5208.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5208.patch",
"merged_at": "2024-07-14T05:07:45"
} | Supports passing in base64 encoded image into image_url.
E.g.
```
curl http://localhost:11434/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "llava",
"messages": [
{
"role": "user",
"content": [
{
"type": "text",
"text": "What'\''s in this image?"
},
{
"type": "image_url",
"image_url": {
"url": "'$image'"
}
}
]
}
],
"max_tokens": 300
}' | jq
```
```
{
"id": "chatcmpl-659",
"object": "chat.completion",
"created": 1719016156,
"model": "llava",
"system_fingerprint": "fp_ollama",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": " The image shows a cute cartoon of an animal. It appears to be a dog or similar creature, styled with exaggerated features typical in internet memes. The character has big eyes, a round face, and its arms are raised in the air, as if waving or giving a thumbs-up gesture. There's also some motion blur that gives the impression of movement, suggesting the animal might be jumping or dancing. This kind of image is often used in digital communication to convey emotions or add a playful element to text messages. "
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 1,
"completion_tokens": 112,
"total_tokens": 113
}
}
```
Resolves #3690, #5304 | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5208/reactions",
"total_count": 6,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 6,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5208/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1640 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1640/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1640/comments | https://api.github.com/repos/ollama/ollama/issues/1640/events | https://github.com/ollama/ollama/pull/1640 | 2,051,142,771 | PR_kwDOJ0Z1Ps5igMrN | 1,640 | added logprobs (`n_probs`) | {
"login": "janpf",
"id": 9437600,
"node_id": "MDQ6VXNlcjk0Mzc2MDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/9437600?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/janpf",
"html_url": "https://github.com/janpf",
"followers_url": "https://api.github.com/users/janpf/followers",
"following_url": "https://api.github.com/users/janpf/following{/other_user}",
"gists_url": "https://api.github.com/users/janpf/gists{/gist_id}",
"starred_url": "https://api.github.com/users/janpf/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/janpf/subscriptions",
"organizations_url": "https://api.github.com/users/janpf/orgs",
"repos_url": "https://api.github.com/users/janpf/repos",
"events_url": "https://api.github.com/users/janpf/events{/privacy}",
"received_events_url": "https://api.github.com/users/janpf/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | {
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 42 | 2023-12-20T19:26:27 | 2025-01-07T19:25:56 | 2025-01-07T19:25:56 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1640",
"html_url": "https://github.com/ollama/ollama/pull/1640",
"diff_url": "https://github.com/ollama/ollama/pull/1640.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1640.patch",
"merged_at": null
} | As discussed on discord I implemented the feature. It just passes through the probs from the llamacpp server. Sorry, first time writing Go, might have missed something.
https://discord.com/channels/1128867683291627614/1128867684130508875/1187028494228664340 | {
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1640/reactions",
"total_count": 38,
"+1": 17,
"-1": 0,
"laugh": 0,
"hooray": 11,
"confused": 0,
"heart": 10,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1640/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3065 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3065/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3065/comments | https://api.github.com/repos/ollama/ollama/issues/3065/events | https://github.com/ollama/ollama/pull/3065 | 2,180,154,843 | PR_kwDOJ0Z1Ps5pTJW3 | 3,065 | relay load model errors to the client | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2024-03-11T20:25:07 | 2024-03-11T20:48:28 | 2024-03-11T20:48:27 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3065",
"html_url": "https://github.com/ollama/ollama/pull/3065",
"diff_url": "https://github.com/ollama/ollama/pull/3065.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3065.patch",
"merged_at": "2024-03-11T20:48:27"
} | Relay errors on model load, this is needed to help people troubleshoot the specific problem they are experiencing when running a model. This function is a bottle-neck where many different errors can occur. As seen in #2753, there are many issues when the generic "failed to load model" error being reported.
In order to effectively identify and fix issues we need to relay more information to the client.
This change:
- Re-throws load errors, so that the client can output the problem
This change does not:
- Provide enhanced information to user's about how to resolve their problem. We will have to add these as we see the issues people face, if they can't be fixed on the Ollama side.
Errors will be relayed like this:
```
ollama run bad-model
Error: exception invalid model dimensions
```
| {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3065/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3065/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/661 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/661/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/661/comments | https://api.github.com/repos/ollama/ollama/issues/661/events | https://github.com/ollama/ollama/pull/661 | 1,920,414,250 | PR_kwDOJ0Z1Ps5bmHz9 | 661 | Documenting Docker Hub image and OpenAI compatibility | {
"login": "jamesbraza",
"id": 8990777,
"node_id": "MDQ6VXNlcjg5OTA3Nzc=",
"avatar_url": "https://avatars.githubusercontent.com/u/8990777?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jamesbraza",
"html_url": "https://github.com/jamesbraza",
"followers_url": "https://api.github.com/users/jamesbraza/followers",
"following_url": "https://api.github.com/users/jamesbraza/following{/other_user}",
"gists_url": "https://api.github.com/users/jamesbraza/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jamesbraza/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jamesbraza/subscriptions",
"organizations_url": "https://api.github.com/users/jamesbraza/orgs",
"repos_url": "https://api.github.com/users/jamesbraza/repos",
"events_url": "https://api.github.com/users/jamesbraza/events{/privacy}",
"received_events_url": "https://api.github.com/users/jamesbraza/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 2 | 2023-09-30T22:09:47 | 2023-10-25T20:18:44 | 2023-10-24T23:15:30 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/661",
"html_url": "https://github.com/ollama/ollama/pull/661",
"diff_url": "https://github.com/ollama/ollama/pull/661.diff",
"patch_url": "https://github.com/ollama/ollama/pull/661.patch",
"merged_at": null
} | - Closes https://github.com/jmorganca/ollama/issues/538
- Upstreams some knowledge from https://github.com/jmorganca/ollama/issues/546
- Simplifies `brew install` to one line | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/661/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/661/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5233 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5233/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5233/comments | https://api.github.com/repos/ollama/ollama/issues/5233/events | https://github.com/ollama/ollama/issues/5233 | 2,368,103,862 | I_kwDOJ0Z1Ps6NJmW2 | 5,233 | filtering library models based on tags? | {
"login": "itsPreto",
"id": 45348368,
"node_id": "MDQ6VXNlcjQ1MzQ4MzY4",
"avatar_url": "https://avatars.githubusercontent.com/u/45348368?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/itsPreto",
"html_url": "https://github.com/itsPreto",
"followers_url": "https://api.github.com/users/itsPreto/followers",
"following_url": "https://api.github.com/users/itsPreto/following{/other_user}",
"gists_url": "https://api.github.com/users/itsPreto/gists{/gist_id}",
"starred_url": "https://api.github.com/users/itsPreto/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/itsPreto/subscriptions",
"organizations_url": "https://api.github.com/users/itsPreto/orgs",
"repos_url": "https://api.github.com/users/itsPreto/repos",
"events_url": "https://api.github.com/users/itsPreto/events{/privacy}",
"received_events_url": "https://api.github.com/users/itsPreto/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 6573197867,
"node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw",
"url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com",
"name": "ollama.com",
"color": "ffffff",
"default": false,
"description": ""
}
] | open | false | null | [] | null | 0 | 2024-06-23T01:02:38 | 2024-07-08T17:17:16 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | can we get something like this for the models library? would be reallyyyy nice!
<img width="387" alt="Screenshot 2024-06-22 at 9 01 25 PM" src="https://github.com/ollama/ollama/assets/45348368/bd142627-9426-41af-b451-67dc82c427df">
| null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5233/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5233/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/5806 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5806/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5806/comments | https://api.github.com/repos/ollama/ollama/issues/5806/events | https://github.com/ollama/ollama/issues/5806 | 2,420,556,455 | I_kwDOJ0Z1Ps6QRsKn | 5,806 | allowing ollama 3 to access local txt files for a larger memory? | {
"login": "boba1234567890",
"id": 107341969,
"node_id": "U_kgDOBmXokQ",
"avatar_url": "https://avatars.githubusercontent.com/u/107341969?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/boba1234567890",
"html_url": "https://github.com/boba1234567890",
"followers_url": "https://api.github.com/users/boba1234567890/followers",
"following_url": "https://api.github.com/users/boba1234567890/following{/other_user}",
"gists_url": "https://api.github.com/users/boba1234567890/gists{/gist_id}",
"starred_url": "https://api.github.com/users/boba1234567890/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/boba1234567890/subscriptions",
"organizations_url": "https://api.github.com/users/boba1234567890/orgs",
"repos_url": "https://api.github.com/users/boba1234567890/repos",
"events_url": "https://api.github.com/users/boba1234567890/events{/privacy}",
"received_events_url": "https://api.github.com/users/boba1234567890/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 3 | 2024-07-20T04:34:36 | 2024-09-04T04:29:27 | 2024-09-04T04:29:26 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | is there a way to allow ollama 3 to use local txt files for a larger memory and maybe other stuff? | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5806/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5806/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5878 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5878/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5878/comments | https://api.github.com/repos/ollama/ollama/issues/5878/events | https://github.com/ollama/ollama/issues/5878 | 2,425,437,900 | I_kwDOJ0Z1Ps6QkT7M | 5,878 | Apple LLM -> DCLM-7B | {
"login": "dvelez3815",
"id": 40648189,
"node_id": "MDQ6VXNlcjQwNjQ4MTg5",
"avatar_url": "https://avatars.githubusercontent.com/u/40648189?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dvelez3815",
"html_url": "https://github.com/dvelez3815",
"followers_url": "https://api.github.com/users/dvelez3815/followers",
"following_url": "https://api.github.com/users/dvelez3815/following{/other_user}",
"gists_url": "https://api.github.com/users/dvelez3815/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dvelez3815/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dvelez3815/subscriptions",
"organizations_url": "https://api.github.com/users/dvelez3815/orgs",
"repos_url": "https://api.github.com/users/dvelez3815/repos",
"events_url": "https://api.github.com/users/dvelez3815/events{/privacy}",
"received_events_url": "https://api.github.com/users/dvelez3815/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | open | false | null | [] | null | 0 | 2024-07-23T14:59:28 | 2024-07-23T15:00:50 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | apple llm
https://huggingface.co/apple/DCLM-7B

| null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5878/reactions",
"total_count": 5,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 5,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5878/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/7300 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7300/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7300/comments | https://api.github.com/repos/ollama/ollama/issues/7300/events | https://github.com/ollama/ollama/issues/7300 | 2,603,110,739 | I_kwDOJ0Z1Ps6bKFFT | 7,300 | Llama3.2-vision Run Error | {
"login": "mruckman1",
"id": 10116867,
"node_id": "MDQ6VXNlcjEwMTE2ODY3",
"avatar_url": "https://avatars.githubusercontent.com/u/10116867?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mruckman1",
"html_url": "https://github.com/mruckman1",
"followers_url": "https://api.github.com/users/mruckman1/followers",
"following_url": "https://api.github.com/users/mruckman1/following{/other_user}",
"gists_url": "https://api.github.com/users/mruckman1/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mruckman1/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mruckman1/subscriptions",
"organizations_url": "https://api.github.com/users/mruckman1/orgs",
"repos_url": "https://api.github.com/users/mruckman1/repos",
"events_url": "https://api.github.com/users/mruckman1/events{/privacy}",
"received_events_url": "https://api.github.com/users/mruckman1/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 21 | 2024-10-21T16:40:09 | 2024-11-05T16:16:29 | 2024-10-23T01:29:12 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
1. Updated Ollama this morning.
2. Entered `ollama run x/llama3.2-vision` on macbook
3. Got below output:
> pulling manifest
> pulling 652e85aa1e14... 100% ▕████████████████▏ 6.0 GB
> pulling 622429e8d318... 100% ▕████████████████▏ 1.9 GB
> pulling 962e0f69a367... 100% ▕████████████████▏ 163 B
> pulling dc49c86b8ebb... 100% ▕████████████████▏ 30 B
> pulling 6a50468ba2a8... 100% ▕████████████████▏ 498 B
> verifying sha256 digest
> writing manifest
> success
**> Error: llama runner process has terminated: error:Missing required key: clip.has_text_encoder**
Expected: Ollama download without error.
### OS
macOS
### GPU
Apple
### CPU
Apple
### Ollama version
0.3.14 | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7300/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7300/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7437 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7437/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7437/comments | https://api.github.com/repos/ollama/ollama/issues/7437/events | https://github.com/ollama/ollama/pull/7437 | 2,625,863,096 | PR_kwDOJ0Z1Ps6AeYjb | 7,437 | Give unicode test more time to run | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2024-10-31T02:58:26 | 2024-10-31T20:35:33 | 2024-10-31T20:35:31 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7437",
"html_url": "https://github.com/ollama/ollama/pull/7437",
"diff_url": "https://github.com/ollama/ollama/pull/7437.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7437.patch",
"merged_at": "2024-10-31T20:35:31"
} | Some slower GPUs (or partial CPU/GPU loads) can take more than the default 30s to complete this test | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7437/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7437/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/800 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/800/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/800/comments | https://api.github.com/repos/ollama/ollama/issues/800/events | https://github.com/ollama/ollama/pull/800 | 1,944,775,176 | PR_kwDOJ0Z1Ps5c4Ey- | 800 | API docs link fix | {
"login": "richawo",
"id": 35015261,
"node_id": "MDQ6VXNlcjM1MDE1MjYx",
"avatar_url": "https://avatars.githubusercontent.com/u/35015261?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/richawo",
"html_url": "https://github.com/richawo",
"followers_url": "https://api.github.com/users/richawo/followers",
"following_url": "https://api.github.com/users/richawo/following{/other_user}",
"gists_url": "https://api.github.com/users/richawo/gists{/gist_id}",
"starred_url": "https://api.github.com/users/richawo/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/richawo/subscriptions",
"organizations_url": "https://api.github.com/users/richawo/orgs",
"repos_url": "https://api.github.com/users/richawo/repos",
"events_url": "https://api.github.com/users/richawo/events{/privacy}",
"received_events_url": "https://api.github.com/users/richawo/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 3 | 2023-10-16T09:28:16 | 2023-10-21T16:00:22 | 2023-10-21T16:00:21 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/800",
"html_url": "https://github.com/ollama/ollama/pull/800",
"diff_url": "https://github.com/ollama/ollama/pull/800.diff",
"patch_url": "https://github.com/ollama/ollama/pull/800.patch",
"merged_at": null
} | For some reason, the relative API docs link is broken (api is a particular path in Github).
Replaced the API docs link in README.md with the absolute path.
Fixes issue #802. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/800/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/800/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4174 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4174/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4174/comments | https://api.github.com/repos/ollama/ollama/issues/4174/events | https://github.com/ollama/ollama/pull/4174 | 2,279,618,845 | PR_kwDOJ0Z1Ps5ukkak | 4,174 | update libraries for langchain_community + llama3 changed from llama2 | {
"login": "Drlordbasil",
"id": 126736516,
"node_id": "U_kgDOB43YhA",
"avatar_url": "https://avatars.githubusercontent.com/u/126736516?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Drlordbasil",
"html_url": "https://github.com/Drlordbasil",
"followers_url": "https://api.github.com/users/Drlordbasil/followers",
"following_url": "https://api.github.com/users/Drlordbasil/following{/other_user}",
"gists_url": "https://api.github.com/users/Drlordbasil/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Drlordbasil/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Drlordbasil/subscriptions",
"organizations_url": "https://api.github.com/users/Drlordbasil/orgs",
"repos_url": "https://api.github.com/users/Drlordbasil/repos",
"events_url": "https://api.github.com/users/Drlordbasil/events{/privacy}",
"received_events_url": "https://api.github.com/users/Drlordbasil/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 1 | 2024-05-05T16:46:34 | 2024-05-06T02:06:32 | 2024-05-05T23:07:04 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4174",
"html_url": "https://github.com/ollama/ollama/pull/4174",
"diff_url": "https://github.com/ollama/ollama/pull/4174.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4174.patch",
"merged_at": "2024-05-05T23:07:04"
} | Changed:
- run ->invoke for updated lib
- updated langchain libraries for non-depreciated
- updated llama2 to llama3 | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4174/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4174/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3782 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3782/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3782/comments | https://api.github.com/repos/ollama/ollama/issues/3782/events | https://github.com/ollama/ollama/pull/3782 | 2,254,683,690 | PR_kwDOJ0Z1Ps5tQTYN | 3,782 | Critical fix from llama.cpp JSON grammar to forbid un-escaped escape characters in JSON strings | {
"login": "hughescr",
"id": 46348,
"node_id": "MDQ6VXNlcjQ2MzQ4",
"avatar_url": "https://avatars.githubusercontent.com/u/46348?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hughescr",
"html_url": "https://github.com/hughescr",
"followers_url": "https://api.github.com/users/hughescr/followers",
"following_url": "https://api.github.com/users/hughescr/following{/other_user}",
"gists_url": "https://api.github.com/users/hughescr/gists{/gist_id}",
"starred_url": "https://api.github.com/users/hughescr/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/hughescr/subscriptions",
"organizations_url": "https://api.github.com/users/hughescr/orgs",
"repos_url": "https://api.github.com/users/hughescr/repos",
"events_url": "https://api.github.com/users/hughescr/events{/privacy}",
"received_events_url": "https://api.github.com/users/hughescr/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 1 | 2024-04-20T19:18:56 | 2024-06-10T01:53:52 | 2024-06-09T17:57:09 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3782",
"html_url": "https://github.com/ollama/ollama/pull/3782",
"diff_url": "https://github.com/ollama/ollama/pull/3782.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3782.patch",
"merged_at": "2024-06-09T17:57:09"
} | JSON generation is broken, as models can insert control characters inside strings, which violates JSON. For example, with the current JSON grammar, models could generate:
```
{ "key": "value
broken" }
```
This is incorrect, and if a linebreak is wanted in the middle of the string there, it should be:
```
{ "key": "value\nbroken" }
```
The former breaks at least the nodejs JSON parser, and likely many many others, since it's not compliant JSON.
This PR injects the grammar directly from llama.cpp upstream, which prohibits `\x00` through `\x1f` inside JSON strings, fixing the problem. | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3782/reactions",
"total_count": 3,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 2,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3782/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/7608 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7608/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7608/comments | https://api.github.com/repos/ollama/ollama/issues/7608/events | https://github.com/ollama/ollama/issues/7608 | 2,648,022,235 | I_kwDOJ0Z1Ps6d1Zzb | 7,608 | pulling manifest error | {
"login": "the-nine-nation",
"id": 103977945,
"node_id": "U_kgDOBjKT2Q",
"avatar_url": "https://avatars.githubusercontent.com/u/103977945?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/the-nine-nation",
"html_url": "https://github.com/the-nine-nation",
"followers_url": "https://api.github.com/users/the-nine-nation/followers",
"following_url": "https://api.github.com/users/the-nine-nation/following{/other_user}",
"gists_url": "https://api.github.com/users/the-nine-nation/gists{/gist_id}",
"starred_url": "https://api.github.com/users/the-nine-nation/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/the-nine-nation/subscriptions",
"organizations_url": "https://api.github.com/users/the-nine-nation/orgs",
"repos_url": "https://api.github.com/users/the-nine-nation/repos",
"events_url": "https://api.github.com/users/the-nine-nation/events{/privacy}",
"received_events_url": "https://api.github.com/users/the-nine-nation/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2024-11-11T03:32:24 | 2024-11-11T03:34:34 | 2024-11-11T03:34:34 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
How to skip pulling manifest ? because my mechine without internet.
### OS
Linux, Docker
### GPU
_No response_
### CPU
Intel
### Ollama version
newest | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7608/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7608/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3118 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3118/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3118/comments | https://api.github.com/repos/ollama/ollama/issues/3118/events | https://github.com/ollama/ollama/issues/3118 | 2,184,554,272 | I_kwDOJ0Z1Ps6CNacg | 3,118 | ollama RAM use on orangepi 5 | {
"login": "parzzd",
"id": 103915075,
"node_id": "U_kgDOBjGeQw",
"avatar_url": "https://avatars.githubusercontent.com/u/103915075?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/parzzd",
"html_url": "https://github.com/parzzd",
"followers_url": "https://api.github.com/users/parzzd/followers",
"following_url": "https://api.github.com/users/parzzd/following{/other_user}",
"gists_url": "https://api.github.com/users/parzzd/gists{/gist_id}",
"starred_url": "https://api.github.com/users/parzzd/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/parzzd/subscriptions",
"organizations_url": "https://api.github.com/users/parzzd/orgs",
"repos_url": "https://api.github.com/users/parzzd/repos",
"events_url": "https://api.github.com/users/parzzd/events{/privacy}",
"received_events_url": "https://api.github.com/users/parzzd/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 2 | 2024-03-13T17:35:48 | 2024-03-13T18:21:56 | 2024-03-13T18:15:01 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Im trying model on my SBC, it has 16gb RAM, but the execution just uses 1.6gb, making the model take so much time to process.
Is there any parameter or configuration to allow it.
im new on Ollama, so any answer would be appreciated.


| {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3118/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3118/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/393 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/393/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/393/comments | https://api.github.com/repos/ollama/ollama/issues/393/events | https://github.com/ollama/ollama/pull/393 | 1,860,413,586 | PR_kwDOJ0Z1Ps5YcUmI | 393 | use url.URL | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2023-08-22T01:56:35 | 2023-08-22T22:51:34 | 2023-08-22T22:51:33 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/393",
"html_url": "https://github.com/ollama/ollama/pull/393",
"diff_url": "https://github.com/ollama/ollama/pull/393.diff",
"patch_url": "https://github.com/ollama/ollama/pull/393.patch",
"merged_at": "2023-08-22T22:51:33"
} | null | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/393/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/393/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/961 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/961/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/961/comments | https://api.github.com/repos/ollama/ollama/issues/961/events | https://github.com/ollama/ollama/issues/961 | 1,972,308,192 | I_kwDOJ0Z1Ps51jwjg | 961 | garbage output on small models spread to many GPUs | {
"login": "chymian",
"id": 1899961,
"node_id": "MDQ6VXNlcjE4OTk5NjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1899961?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/chymian",
"html_url": "https://github.com/chymian",
"followers_url": "https://api.github.com/users/chymian/followers",
"following_url": "https://api.github.com/users/chymian/following{/other_user}",
"gists_url": "https://api.github.com/users/chymian/gists{/gist_id}",
"starred_url": "https://api.github.com/users/chymian/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/chymian/subscriptions",
"organizations_url": "https://api.github.com/users/chymian/orgs",
"repos_url": "https://api.github.com/users/chymian/repos",
"events_url": "https://api.github.com/users/chymian/events{/privacy}",
"received_events_url": "https://api.github.com/users/chymian/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg",
"url": "https://api.github.com/repos/ollama/ollama/labels/nvidia",
"name": "nvidia",
"color": "8CDB00",
"default": false,
"description": "Issues relating to Nvidia GPUs and CUDA"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 7 | 2023-11-01T12:43:10 | 2024-04-23T15:31:40 | 2024-04-23T15:31:40 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | when loading a small model on multiple GPUs, it produces garbage.
the machine has 4 x 3070 (8GB) and an older i5-7400, UBU 22.04, Cuda 11.8
### How to reproduce
starting the server by hand
```bash
ollama serve
```
```bash
ollama run zephyr
>>> why is the sky blue?
acia##############################################################################^C
```
nvidia-smi shows it loads 4 x 1,2 GB
### Workaround
Till we get some flags to control the loading to a discrete GPU
start the server with `CUDA_VISIBLE_DEVICES`
```bash
CUDA_VISIBLE_DEVICES="3" ollama serve # loads the model on GPU No. 4
```
There seems to be a "memory-low-limit-point", under which this error occures.
| {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/961/reactions",
"total_count": 3,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 2,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/961/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1978 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1978/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1978/comments | https://api.github.com/repos/ollama/ollama/issues/1978/events | https://github.com/ollama/ollama/issues/1978 | 2,080,398,998 | I_kwDOJ0Z1Ps58AF6W | 1,978 | Error "unknown architecture MistralModel" during quantization | {
"login": "philippgille",
"id": 170670,
"node_id": "MDQ6VXNlcjE3MDY3MA==",
"avatar_url": "https://avatars.githubusercontent.com/u/170670?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/philippgille",
"html_url": "https://github.com/philippgille",
"followers_url": "https://api.github.com/users/philippgille/followers",
"following_url": "https://api.github.com/users/philippgille/following{/other_user}",
"gists_url": "https://api.github.com/users/philippgille/gists{/gist_id}",
"starred_url": "https://api.github.com/users/philippgille/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/philippgille/subscriptions",
"organizations_url": "https://api.github.com/users/philippgille/orgs",
"repos_url": "https://api.github.com/users/philippgille/repos",
"events_url": "https://api.github.com/users/philippgille/events{/privacy}",
"received_events_url": "https://api.github.com/users/philippgille/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 2 | 2024-01-13T17:16:28 | 2024-05-07T00:08:53 | 2024-05-06T23:48:21 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hello :wave: , First of all thank you very much for creating and maintaining ollama! It's so simple to use :+1:
Now I wanted to use ollama for creating embeddings, and saw https://huggingface.co/intfloat/e5-mistral-7b-instruct performing very well on the [embeddings benchmark](https://huggingface.co/spaces/mteb/leaderboard). The official ollama model library doesn't contain it yet, so I wanted to create and upload it myself.
But during the quantization step (`docker run --rm -v .:/model:Z ollama/quantize -q q4_0 /model`) I get the error:
> unknown architecture MistralModel
As Mistral is supported by ollama, I'm wondering about this error. The E5 model is based on the Mistral instruct v0.1 one, so I assume it's the same architecture. Right? Is maybe just the `ollama/quantize` image not updated with the support yet? | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1978/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1978/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4861 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4861/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4861/comments | https://api.github.com/repos/ollama/ollama/issues/4861/events | https://github.com/ollama/ollama/issues/4861 | 2,338,549,595 | I_kwDOJ0Z1Ps6LY29b | 4,861 | Jetson - "ollama run" command loads until timeout | {
"login": "Vassar-HARPER-Project",
"id": 171359116,
"node_id": "U_kgDOCja7jA",
"avatar_url": "https://avatars.githubusercontent.com/u/171359116?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Vassar-HARPER-Project",
"html_url": "https://github.com/Vassar-HARPER-Project",
"followers_url": "https://api.github.com/users/Vassar-HARPER-Project/followers",
"following_url": "https://api.github.com/users/Vassar-HARPER-Project/following{/other_user}",
"gists_url": "https://api.github.com/users/Vassar-HARPER-Project/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Vassar-HARPER-Project/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Vassar-HARPER-Project/subscriptions",
"organizations_url": "https://api.github.com/users/Vassar-HARPER-Project/orgs",
"repos_url": "https://api.github.com/users/Vassar-HARPER-Project/repos",
"events_url": "https://api.github.com/users/Vassar-HARPER-Project/events{/privacy}",
"received_events_url": "https://api.github.com/users/Vassar-HARPER-Project/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg",
"url": "https://api.github.com/repos/ollama/ollama/labels/nvidia",
"name": "nvidia",
"color": "8CDB00",
"default": false,
"description": "Issues relating to Nvidia GPUs and CUDA"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 11 | 2024-06-06T15:34:26 | 2024-11-12T18:31:55 | 2024-11-12T18:31:55 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Upon running "ollama run gemma:2b" (though this happens for all tested models: llama3, phi, tinyllama), the loading animation appears and after ~5 minutes (estimate, untimed), the response / result of the command is:
`Error: timed out waiting for llama runner to start - progress 1.00 - `
the server shows this log for this command:
```
2024/06/06 11:21:53 routes.go:1007: INFO server config env="map[OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_HOST: OLLAMA_KEEP_ALIVE: OLLAMA_LLM_LIBRARY: OLLAMA_MAX_LOADED_MODELS:1 OLLAMA_MAX_QUEUE:512 OLLAMA_MAX_VRAM:0 OLLAMA_MODELS: OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:*] OLLAMA_RUNNERS_DIR: OLLAMA_TMPDIR:]"
time=2024-06-06T11:21:53.848-04:00 level=INFO source=images.go:729 msg="total blobs: 11"
time=2024-06-06T11:21:53.849-04:00 level=INFO source=images.go:736 msg="total unused blobs removed: 0"
time=2024-06-06T11:21:53.849-04:00 level=INFO source=routes.go:1053 msg="Listening on 127.0.0.1:11434 (version 0.1.41)"
time=2024-06-06T11:21:53.850-04:00 level=INFO source=payload.go:30 msg="extracting embedded files" dir=/tmp/ollama3794080172/runners
time=2024-06-06T11:21:58.984-04:00 level=INFO source=payload.go:44 msg="Dynamic LLM libraries [cpu cuda_v11]"
time=2024-06-06T11:21:59.082-04:00 level=INFO source=types.go:71 msg="inference compute" id=GPU-42638932-6929-58db-a006-34d50a6799c1 library=cuda compute=8.7 driver=11.4 name=Orin total="29.9 GiB" available="21.7 GiB"
[GIN] 2024/06/06 - 11:22:14 | 200 | 64.512µs | 127.0.0.1 | HEAD "/"
[GIN] 2024/06/06 - 11:22:14 | 200 | 1.232036ms | 127.0.0.1 | POST "/api/show"
[GIN] 2024/06/06 - 11:22:14 | 200 | 717.058µs | 127.0.0.1 | POST "/api/show"
time=2024-06-06T11:22:16.239-04:00 level=INFO source=memory.go:133 msg="offload to gpu" layers.requested=-1 layers.real=19 memory.available="21.7 GiB" memory.required.full="2.6 GiB" memory.required.partial="2.6 GiB" memory.required.kv="36.0 MiB" memory.weights.total="1.6 GiB" memory.weights.repeating="1.0 GiB" memory.weights.nonrepeating="531.5 MiB" memory.graph.full="504.2 MiB" memory.graph.partial="918.6 MiB"
time=2024-06-06T11:22:16.239-04:00 level=INFO source=memory.go:133 msg="offload to gpu" layers.requested=-1 layers.real=19 memory.available="21.7 GiB" memory.required.full="2.6 GiB" memory.required.partial="2.6 GiB" memory.required.kv="36.0 MiB" memory.weights.total="1.6 GiB" memory.weights.repeating="1.0 GiB" memory.weights.nonrepeating="531.5 MiB" memory.graph.full="504.2 MiB" memory.graph.partial="918.6 MiB"
time=2024-06-06T11:22:16.240-04:00 level=INFO source=server.go:341 msg="starting llama server" cmd="/tmp/ollama3794080172/runners/cuda_v11/ollama_llama_server --model /home/harper/.ollama/models/blobs/sha256-c1864a5eb19305c40519da12cc543519e48a0697ecd30e15d5ac228644957d12 --ctx-size 2048 --batch-size 512 --embedding --log-disable --n-gpu-layers 19 --parallel 1 --port 42781"
time=2024-06-06T11:22:16.240-04:00 level=INFO source=sched.go:338 msg="loaded runners" count=1
time=2024-06-06T11:22:16.240-04:00 level=INFO source=server.go:529 msg="waiting for llama runner to start responding"
time=2024-06-06T11:22:16.241-04:00 level=INFO source=server.go:567 msg="waiting for server to become available" status="llm server error"
INFO [main] build info | build=1 commit="5921b8f" tid="281473278327040" timestamp=1717687336
INFO [main] system info | n_threads=8 n_threads_batch=-1 system_info="AVX = 0 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 0 | NEON = 1 | SVE = 0 | ARM_FMA = 1 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 0 | SSSE3 = 0 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="281473278327040" timestamp=1717687336 total_threads=8
INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="7" port="42781" tid="281473278327040" timestamp=1717687336
llama_model_loader: loaded meta data with 21 key-value pairs and 164 tensors from /home/harper/.ollama/models/blobs/sha256-c1864a5eb19305c40519da12cc543519e48a0697ecd30e15d5ac228644957d12 (version GGUF V3 (latest))
llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.
llama_model_loader: - kv 0: general.architecture str = gemma
llama_model_loader: - kv 1: general.name str = gemma-2b-it
llama_model_loader: - kv 2: gemma.context_length u32 = 8192
llama_model_loader: - kv 3: gemma.block_count u32 = 18
llama_model_loader: - kv 4: gemma.embedding_length u32 = 2048
llama_model_loader: - kv 5: gemma.feed_forward_length u32 = 16384
llama_model_loader: - kv 6: gemma.attention.head_count u32 = 8
llama_model_loader: - kv 7: gemma.attention.head_count_kv u32 = 1
llama_model_loader: - kv 8: gemma.attention.key_length u32 = 256
llama_model_loader: - kv 9: gemma.attention.value_length u32 = 256
llama_model_loader: - kv 10: gemma.attention.layer_norm_rms_epsilon f32 = 0.000001
llama_model_loader: - kv 11: tokenizer.ggml.model str = llama
llama_model_loader: - kv 12: tokenizer.ggml.bos_token_id u32 = 2
llama_model_loader: - kv 13: tokenizer.ggml.eos_token_id u32 = 1
llama_model_loader: - kv 14: tokenizer.ggml.padding_token_id u32 = 0
llama_model_loader: - kv 15: tokenizer.ggml.unknown_token_id u32 = 3
llama_model_loader: - kv 16: tokenizer.ggml.tokens arr[str,256128] = ["<pad>", "<eos>", "<bos>", "<unk>", ...
time=2024-06-06T11:22:16.493-04:00 level=INFO source=server.go:567 msg="waiting for server to become available" status="llm server loading model"
llama_model_loader: - kv 17: tokenizer.ggml.scores arr[f32,256128] = [0.000000, 0.000000, 0.000000, 0.0000...
llama_model_loader: - kv 18: tokenizer.ggml.token_type arr[i32,256128] = [3, 3, 3, 2, 1, 1, 1, 1, 1, 1, 1, 1, ...
llama_model_loader: - kv 19: general.quantization_version u32 = 2
llama_model_loader: - kv 20: general.file_type u32 = 2
llama_model_loader: - type f32: 37 tensors
llama_model_loader: - type q4_0: 126 tensors
llama_model_loader: - type q8_0: 1 tensors
llm_load_vocab: special tokens cache size = 388
llm_load_vocab: token to piece cache size = 3.2028 MB
llm_load_print_meta: format = GGUF V3 (latest)
llm_load_print_meta: arch = gemma
llm_load_print_meta: vocab type = SPM
llm_load_print_meta: n_vocab = 256128
llm_load_print_meta: n_merges = 0
llm_load_print_meta: n_ctx_train = 8192
llm_load_print_meta: n_embd = 2048
llm_load_print_meta: n_head = 8
llm_load_print_meta: n_head_kv = 1
llm_load_print_meta: n_layer = 18
llm_load_print_meta: n_rot = 256
llm_load_print_meta: n_embd_head_k = 256
llm_load_print_meta: n_embd_head_v = 256
llm_load_print_meta: n_gqa = 8
llm_load_print_meta: n_embd_k_gqa = 256
llm_load_print_meta: n_embd_v_gqa = 256
llm_load_print_meta: f_norm_eps = 0.0e+00
llm_load_print_meta: f_norm_rms_eps = 1.0e-06
llm_load_print_meta: f_clamp_kqv = 0.0e+00
llm_load_print_meta: f_max_alibi_bias = 0.0e+00
llm_load_print_meta: f_logit_scale = 0.0e+00
llm_load_print_meta: n_ff = 16384
llm_load_print_meta: n_expert = 0
llm_load_print_meta: n_expert_used = 0
llm_load_print_meta: causal attn = 1
llm_load_print_meta: pooling type = 0
llm_load_print_meta: rope type = 2
llm_load_print_meta: rope scaling = linear
llm_load_print_meta: freq_base_train = 10000.0
llm_load_print_meta: freq_scale_train = 1
llm_load_print_meta: n_yarn_orig_ctx = 8192
llm_load_print_meta: rope_finetuned = unknown
llm_load_print_meta: ssm_d_conv = 0
llm_load_print_meta: ssm_d_inner = 0
llm_load_print_meta: ssm_d_state = 0
llm_load_print_meta: ssm_dt_rank = 0
llm_load_print_meta: model type = 2B
llm_load_print_meta: model ftype = Q4_0
llm_load_print_meta: model params = 2.51 B
llm_load_print_meta: model size = 1.56 GiB (5.34 BPW)
llm_load_print_meta: general.name = gemma-2b-it
llm_load_print_meta: BOS token = 2 '<bos>'
llm_load_print_meta: EOS token = 1 '<eos>'
llm_load_print_meta: UNK token = 3 '<unk>'
llm_load_print_meta: PAD token = 0 '<pad>'
llm_load_print_meta: LF token = 227 '<0x0A>'
llm_load_print_meta: EOT token = 107 '<end_of_turn>'
ggml_cuda_init: GGML_CUDA_FORCE_MMQ: yes
ggml_cuda_init: CUDA_USE_TENSOR_CORES: no
ggml_cuda_init: found 1 CUDA devices:
Device 0: Orin, compute capability 8.7, VMM: yes
llm_load_tensors: ggml ctx size = 0.17 MiB
llm_load_tensors: offloading 18 repeating layers to GPU
llm_load_tensors: offloading non-repeating layers to GPU
llm_load_tensors: offloaded 19/19 layers to GPU
llm_load_tensors: CPU buffer size = 531.52 MiB
llm_load_tensors: CUDA0 buffer size = 1594.93 MiB
llama_new_context_with_model: n_ctx = 2048
llama_new_context_with_model: n_batch = 512
llama_new_context_with_model: n_ubatch = 512
llama_new_context_with_model: flash_attn = 0
llama_new_context_with_model: freq_base = 10000.0
llama_new_context_with_model: freq_scale = 1
llama_kv_cache_init: CUDA0 KV buffer size = 36.00 MiB
llama_new_context_with_model: KV self size = 36.00 MiB, K (f16): 18.00 MiB, V (f16): 18.00 MiB
llama_new_context_with_model: CUDA_Host output buffer size = 0.98 MiB
llama_new_context_with_model: CUDA0 compute buffer size = 504.25 MiB
llama_new_context_with_model: CUDA_Host compute buffer size = 8.01 MiB
llama_new_context_with_model: graph nodes = 601
llama_new_context_with_model: graph splits = 2
time=2024-06-06T11:29:16.920-04:00 level=ERROR source=sched.go:344 msg="error loading llama server" error="timed out waiting for llama runner to start - progress 1.00 - "
[GIN] 2024/06/06 - 11:29:16 | 500 | 7m2s | 127.0.0.1 | POST "/api/chat"
time=2024-06-06T11:29:22.037-04:00 level=WARN source=sched.go:512 msg="gpu VRAM usage didn't recover within timeout" seconds=5.117139389
time=2024-06-06T11:29:22.288-04:00 level=WARN source=sched.go:512 msg="gpu VRAM usage didn't recover within timeout" seconds=5.367497274
time=2024-06-06T11:29:22.537-04:00 level=WARN source=sched.go:512 msg="gpu VRAM usage didn't recover within timeout" seconds=5.616874999
```
If it helps, this is running on a [Jetson AGX Orin with 32GB of memory](https://www.seeedstudio.com/AGX-Orin-32GB-H01-Kit-p-5569.html)
### OS
Linux
### GPU
Nvidia
### CPU
Other: 8-core NVIDIA Arm® Cortex A78AE v8.2 64-bit CPU 2MB L2 + 4MB L3
### Ollama version
0.1.41 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4861/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4861/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7141 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7141/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7141/comments | https://api.github.com/repos/ollama/ollama/issues/7141/events | https://github.com/ollama/ollama/pull/7141 | 2,573,860,762 | PR_kwDOJ0Z1Ps59-4qQ | 7,141 | Fix build leakages | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2024-10-08T18:00:51 | 2024-10-08T20:05:03 | 2024-10-08T20:05:00 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7141",
"html_url": "https://github.com/ollama/ollama/pull/7141",
"diff_url": "https://github.com/ollama/ollama/pull/7141.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7141.patch",
"merged_at": "2024-10-08T20:05:00"
} | The recent change to applying patches leaves the submodule dirty based on "new commits" being present. This ensures we clean up so the tree no longer reports dirty after a `go generate ./...` run.
The Makefile was being a bit too aggressive in cleaning things up and would result in deleting the placeholder files which someone might accidentally commit. | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7141/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7141/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6269 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6269/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6269/comments | https://api.github.com/repos/ollama/ollama/issues/6269/events | https://github.com/ollama/ollama/issues/6269 | 2,456,908,603 | I_kwDOJ0Z1Ps6ScXM7 | 6,269 | Please add LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct model | {
"login": "xest",
"id": 4961215,
"node_id": "MDQ6VXNlcjQ5NjEyMTU=",
"avatar_url": "https://avatars.githubusercontent.com/u/4961215?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/xest",
"html_url": "https://github.com/xest",
"followers_url": "https://api.github.com/users/xest/followers",
"following_url": "https://api.github.com/users/xest/following{/other_user}",
"gists_url": "https://api.github.com/users/xest/gists{/gist_id}",
"starred_url": "https://api.github.com/users/xest/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/xest/subscriptions",
"organizations_url": "https://api.github.com/users/xest/orgs",
"repos_url": "https://api.github.com/users/xest/repos",
"events_url": "https://api.github.com/users/xest/events{/privacy}",
"received_events_url": "https://api.github.com/users/xest/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | closed | false | null | [] | null | 9 | 2024-08-09T01:07:48 | 2024-12-10T08:05:04 | 2024-12-10T08:05:04 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | * huggingface: [LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct](https://huggingface.co/LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct) | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6269/reactions",
"total_count": 3,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6269/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3931 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3931/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3931/comments | https://api.github.com/repos/ollama/ollama/issues/3931/events | https://github.com/ollama/ollama/issues/3931 | 2,264,958,659 | I_kwDOJ0Z1Ps6HAIbD | 3,931 | Digest mismatch, file must be downloaded again | {
"login": "tttt-0814",
"id": 39620928,
"node_id": "MDQ6VXNlcjM5NjIwOTI4",
"avatar_url": "https://avatars.githubusercontent.com/u/39620928?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tttt-0814",
"html_url": "https://github.com/tttt-0814",
"followers_url": "https://api.github.com/users/tttt-0814/followers",
"following_url": "https://api.github.com/users/tttt-0814/following{/other_user}",
"gists_url": "https://api.github.com/users/tttt-0814/gists{/gist_id}",
"starred_url": "https://api.github.com/users/tttt-0814/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/tttt-0814/subscriptions",
"organizations_url": "https://api.github.com/users/tttt-0814/orgs",
"repos_url": "https://api.github.com/users/tttt-0814/repos",
"events_url": "https://api.github.com/users/tttt-0814/events{/privacy}",
"received_events_url": "https://api.github.com/users/tttt-0814/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 9 | 2024-04-26T04:52:11 | 2025-01-30T02:39:33 | 2024-05-09T21:08:04 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I tried to pull nomic-embed-text, but got an error below.
I also tried to pull another models, but got the same error.
$ ollama pull nomic-embed-text
pulling manifest
pulling 970aa74c0a90... 100% ▕███████████████████████████████████████████████████████████████████████████████████████████▏ 274 MB
pulling c71d239df917... 100% ▕███████████████████████████████████████████████████████████████████████████████████████████▏ 11 KB
pulling ce4a164fc046... 100% ▕███████████████████████████████████████████████████████████████████████████████████████████▏ 17 B
pulling 31df23ea7daa... 100% ▕███████████████████████████████████████████████████████████████████████████████████████████▏ 420 B
verifying sha256 digest
Error: digest mismatch, file must be downloaded again: want sha256:970aa74c0a90ef7482477cf803618e776e173c007bf957f635f1015bfcfef0e6, got sha256:bea7e365d4085c35b0cfc78f9285682c6c7df7a15ac58f1905542659871024fd
### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.1.32 | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3931/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3931/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4879 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4879/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4879/comments | https://api.github.com/repos/ollama/ollama/issues/4879/events | https://github.com/ollama/ollama/pull/4879 | 2,338,955,742 | PR_kwDOJ0Z1Ps5xuCen | 4,879 | API app/browser access | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2024-06-06T18:56:08 | 2024-06-06T22:19:04 | 2024-06-06T22:19:03 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/4879",
"html_url": "https://github.com/ollama/ollama/pull/4879",
"diff_url": "https://github.com/ollama/ollama/pull/4879.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4879.patch",
"merged_at": "2024-06-06T22:19:03"
} | Fixes #4791
Fixes #3799
Fixes #4388 | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4879/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 1,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4879/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1764 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1764/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1764/comments | https://api.github.com/repos/ollama/ollama/issues/1764/events | https://github.com/ollama/ollama/pull/1764 | 2,063,147,841 | PR_kwDOJ0Z1Ps5jFzXu | 1,764 | keyboard shortcut help | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2024-01-03T01:59:41 | 2024-01-03T02:04:13 | 2024-01-03T02:04:13 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1764",
"html_url": "https://github.com/ollama/ollama/pull/1764",
"diff_url": "https://github.com/ollama/ollama/pull/1764.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1764.patch",
"merged_at": "2024-01-03T02:04:13"
} | This change adds some help in the REPL for using the keyboard shortcut commands. | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1764/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1764/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4200 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4200/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4200/comments | https://api.github.com/repos/ollama/ollama/issues/4200/events | https://github.com/ollama/ollama/issues/4200 | 2,280,915,245 | I_kwDOJ0Z1Ps6H9AEt | 4,200 | http://localhost:11434/api endpoint giving 404 error | {
"login": "ritesh7911",
"id": 64787172,
"node_id": "MDQ6VXNlcjY0Nzg3MTcy",
"avatar_url": "https://avatars.githubusercontent.com/u/64787172?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ritesh7911",
"html_url": "https://github.com/ritesh7911",
"followers_url": "https://api.github.com/users/ritesh7911/followers",
"following_url": "https://api.github.com/users/ritesh7911/following{/other_user}",
"gists_url": "https://api.github.com/users/ritesh7911/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ritesh7911/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ritesh7911/subscriptions",
"organizations_url": "https://api.github.com/users/ritesh7911/orgs",
"repos_url": "https://api.github.com/users/ritesh7911/repos",
"events_url": "https://api.github.com/users/ritesh7911/events{/privacy}",
"received_events_url": "https://api.github.com/users/ritesh7911/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 3 | 2024-05-06T13:42:10 | 2024-05-08T20:27:28 | 2024-05-08T20:27:24 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I am using latest version of windows . As per readme file when I am hitting http://localhost:11434 i am getting "ollama is running" but "http://localhost:11434/api" is giving error 404
### OS
Windows
### GPU
AMD
### CPU
Intel
### Ollama version
0.1.33 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4200/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4200/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3554 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3554/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3554/comments | https://api.github.com/repos/ollama/ollama/issues/3554/events | https://github.com/ollama/ollama/issues/3554 | 2,233,075,542 | I_kwDOJ0Z1Ps6FGgdW | 3,554 | Potential problems with the `llm/ext_server/server.cpp` not accepting `--ubatch-size ` option | {
"login": "jukofyork",
"id": 69222624,
"node_id": "MDQ6VXNlcjY5MjIyNjI0",
"avatar_url": "https://avatars.githubusercontent.com/u/69222624?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jukofyork",
"html_url": "https://github.com/jukofyork",
"followers_url": "https://api.github.com/users/jukofyork/followers",
"following_url": "https://api.github.com/users/jukofyork/following{/other_user}",
"gists_url": "https://api.github.com/users/jukofyork/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jukofyork/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jukofyork/subscriptions",
"organizations_url": "https://api.github.com/users/jukofyork/orgs",
"repos_url": "https://api.github.com/users/jukofyork/repos",
"events_url": "https://api.github.com/users/jukofyork/events{/privacy}",
"received_events_url": "https://api.github.com/users/jukofyork/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | {
"login": "jessegross",
"id": 6468499,
"node_id": "MDQ6VXNlcjY0Njg0OTk=",
"avatar_url": "https://avatars.githubusercontent.com/u/6468499?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jessegross",
"html_url": "https://github.com/jessegross",
"followers_url": "https://api.github.com/users/jessegross/followers",
"following_url": "https://api.github.com/users/jessegross/following{/other_user}",
"gists_url": "https://api.github.com/users/jessegross/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jessegross/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jessegross/subscriptions",
"organizations_url": "https://api.github.com/users/jessegross/orgs",
"repos_url": "https://api.github.com/users/jessegross/repos",
"events_url": "https://api.github.com/users/jessegross/events{/privacy}",
"received_events_url": "https://api.github.com/users/jessegross/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"login": "jessegross",
"id": 6468499,
"node_id": "MDQ6VXNlcjY0Njg0OTk=",
"avatar_url": "https://avatars.githubusercontent.com/u/6468499?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jessegross",
"html_url": "https://github.com/jessegross",
"followers_url": "https://api.github.com/users/jessegross/followers",
"following_url": "https://api.github.com/users/jessegross/following{/other_user}",
"gists_url": "https://api.github.com/users/jessegross/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jessegross/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jessegross/subscriptions",
"organizations_url": "https://api.github.com/users/jessegross/orgs",
"repos_url": "https://api.github.com/users/jessegross/repos",
"events_url": "https://api.github.com/users/jessegross/events{/privacy}",
"received_events_url": "https://api.github.com/users/jessegross/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 1 | 2024-04-09T10:01:50 | 2024-11-23T20:17:29 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Not sure what to list this issue under (it's a potential bug I think).
Recently `llama.cpp` has added an option called `--ubatch-size ` and appears to have changed the default value (and possibly meaning of) the old `--batch-size ` option:
https://github.com/ggerganov/llama.cpp/pull/6017
The PR says:
> By default `n_batch` is 4096, `n_ubatch` is 512
but I think it is actually 2048 and 512 now.
This better explains the difference between the options:
https://github.com/ggerganov/llama.cpp/discussions/6328
There are several potential problems and implications of this change for Ollama:
- It looks like `llm/ext_server/server.cpp` is using some old code that handles the `n_batch` in the old way, but this in turn seems to be importing headers, etc from a newer version of `lamma.cpp` (b2581), who's `examples/server.cpp` code does appear to have the new `n_batch` / `n_ubatch` logic in it.
- It looks like `api/types.go` is setting the `n_batch` value back down to 512, when the new default is supposed to be 2048.
- Since Ollama is using the `--embedding` option there are some potential problems mentioned here: https://github.com/ggerganov/llama.cpp/issues/6263
- Not being able to set the `n_ubatch` value is causing regressions for people using multiple GPUs (https://github.com/ggerganov/llama.cpp/issues/6075 & https://github.com/ggerganov/llama.cpp/issues/6476).
---
https://github.com/ollama/ollama/blob/main/llm/ext_server/server.cpp:
```
printf(" --pooling {none,mean,cls}\n");
printf(" pooling type for embeddings, use model default if unspecified\n");
printf(" -b N, --batch-size N batch size for prompt processing (default: %d)\n", params.n_batch);
printf(" --memory-f32 use f32 instead of f16 for memory key+value (default: disabled)\n");
```
vs:
https://github.com/ggerganov/llama.cpp/blob/37e7854c104301c5b5323ccc40e07699f3a62c3e/examples/server/server.cpp:
```
printf(" --pooling {none,mean,cls} pooling type for embeddings, use model default if unspecified\n");
printf(" -dt N, --defrag-thold N\n");
printf(" KV cache defragmentation threshold (default: %.1f, < 0 - disabled)\n", params.defrag_thold);
printf(" -b N, --batch-size N logical maximum batch size (default: %d)\n", params.n_batch);
printf(" -ub N, --ubatch-size N physical maximum batch size (default: %d)\n", params.n_ubatch);
printf(" --memory-f32 use f32 instead of f16 for memory key+value (default: disabled)\n");
```
---
```
- `-b N`, `--batch-size N`: Set the batch size for prompt processing. Default: `2048`
- `-ub N`, `--ubatch-size N`: Physical maximum batch size. Default: `512`
``` | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3554/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3554/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/4742 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4742/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4742/comments | https://api.github.com/repos/ollama/ollama/issues/4742/events | https://github.com/ollama/ollama/issues/4742 | 2,327,001,333 | I_kwDOJ0Z1Ps6Kszj1 | 4,742 | VRAM allocation error when loading different models with different OLLAMA_VRAM_MAX configurations | {
"login": "hamkido",
"id": 43724352,
"node_id": "MDQ6VXNlcjQzNzI0MzUy",
"avatar_url": "https://avatars.githubusercontent.com/u/43724352?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hamkido",
"html_url": "https://github.com/hamkido",
"followers_url": "https://api.github.com/users/hamkido/followers",
"following_url": "https://api.github.com/users/hamkido/following{/other_user}",
"gists_url": "https://api.github.com/users/hamkido/gists{/gist_id}",
"starred_url": "https://api.github.com/users/hamkido/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/hamkido/subscriptions",
"organizations_url": "https://api.github.com/users/hamkido/orgs",
"repos_url": "https://api.github.com/users/hamkido/repos",
"events_url": "https://api.github.com/users/hamkido/events{/privacy}",
"received_events_url": "https://api.github.com/users/hamkido/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 3 | 2024-05-31T05:38:13 | 2024-06-05T06:34:19 | 2024-05-31T06:31:53 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I have two amd 7900xtx 24g gpu. When using ollama, I encounter different memory allocation errors and exit errors.
1. No OLLAMA_VRAM_MAX configuration
The large model deepseek-llm:67b-chat can be loaded correctly
But if you call something bigger, such as qwen:72b and command-r-plus, the display memory allocation will report an error and exit.
2. Use OLLAMA_VRAM_MAX configuration
Models larger than 67b-q4 load correctly, such as qwen:72b and command-r-plus.
However, smaller models cannot be loaded correctly, and memory allocation errors will be reported and exited, such as deepseek-llm:67b-chat.
There might be some errors in the vram configuration.
### OS
Linux
### GPU
AMD
### CPU
AMD
### Ollama version
0.1.39 | {
"login": "hamkido",
"id": 43724352,
"node_id": "MDQ6VXNlcjQzNzI0MzUy",
"avatar_url": "https://avatars.githubusercontent.com/u/43724352?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hamkido",
"html_url": "https://github.com/hamkido",
"followers_url": "https://api.github.com/users/hamkido/followers",
"following_url": "https://api.github.com/users/hamkido/following{/other_user}",
"gists_url": "https://api.github.com/users/hamkido/gists{/gist_id}",
"starred_url": "https://api.github.com/users/hamkido/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/hamkido/subscriptions",
"organizations_url": "https://api.github.com/users/hamkido/orgs",
"repos_url": "https://api.github.com/users/hamkido/repos",
"events_url": "https://api.github.com/users/hamkido/events{/privacy}",
"received_events_url": "https://api.github.com/users/hamkido/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4742/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4742/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8526 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8526/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8526/comments | https://api.github.com/repos/ollama/ollama/issues/8526/events | https://github.com/ollama/ollama/issues/8526 | 2,803,177,332 | I_kwDOJ0Z1Ps6nFRd0 | 8,526 | how to get English output | {
"login": "jarkkop",
"id": 5814285,
"node_id": "MDQ6VXNlcjU4MTQyODU=",
"avatar_url": "https://avatars.githubusercontent.com/u/5814285?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jarkkop",
"html_url": "https://github.com/jarkkop",
"followers_url": "https://api.github.com/users/jarkkop/followers",
"following_url": "https://api.github.com/users/jarkkop/following{/other_user}",
"gists_url": "https://api.github.com/users/jarkkop/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jarkkop/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jarkkop/subscriptions",
"organizations_url": "https://api.github.com/users/jarkkop/orgs",
"repos_url": "https://api.github.com/users/jarkkop/repos",
"events_url": "https://api.github.com/users/jarkkop/events{/privacy}",
"received_events_url": "https://api.github.com/users/jarkkop/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 1 | 2025-01-22T01:51:56 | 2025-01-22T04:17:26 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
M:\AI\ollama>ollama run deepseek-r1:7b
>>> list philosophers
<think>
</think>
Here is a list of some of the most influential and notable philosophers throughout history, organized by era and
region:
### Ancient Philosophy (c. 600–321 BCE)
- **Thales of Miletus** (c. 624–548 BCE):被认为是第一个哲学家,提出“万物源于水”的学说。
- **Anaximander of Miletus** (c. 570–495 BCE): 提出“无限”概念,并认为万物起源于自然。
- **Anaximenes of Mileti** (c. 510–441 BCE): 认为万物来源于某种原始物质,如“air”(空气)。
### OS
Microsoft Windows [Version 10.0.22635.4800]
### GPU
NVIDIA GeForce GTX 1660 SUPER
### CPU
Processor AMD Ryzen 7 3700X 8-Core Processor, 3600 Mhz, 8 Core(s), 16 Logical Processor(s)
### Ollama version
ollama version is 0.5.7 | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8526/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8526/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/672 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/672/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/672/comments | https://api.github.com/repos/ollama/ollama/issues/672/events | https://github.com/ollama/ollama/pull/672 | 1,922,337,068 | PR_kwDOJ0Z1Ps5bse6w | 672 | Relay default values to llama runner | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 3 | 2023-10-02T17:32:53 | 2023-10-02T18:53:17 | 2023-10-02T18:53:16 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/672",
"html_url": "https://github.com/ollama/ollama/pull/672",
"diff_url": "https://github.com/ollama/ollama/pull/672.diff",
"patch_url": "https://github.com/ollama/ollama/pull/672.patch",
"merged_at": "2023-10-02T18:53:16"
} | Thanks to @hallh for #663. This change cherry-picks that PR, relays all our defaults, and does some re-organizing of the code to make it easier to read. | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/672/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/672/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/8113 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8113/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8113/comments | https://api.github.com/repos/ollama/ollama/issues/8113/events | https://github.com/ollama/ollama/pull/8113 | 2,741,618,542 | PR_kwDOJ0Z1Ps6FULcl | 8,113 | llama: add qwen2vl support | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | open | false | null | [] | null | 0 | 2024-12-16T07:59:10 | 2025-01-15T11:14:59 | null | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | true | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8113",
"html_url": "https://github.com/ollama/ollama/pull/8113",
"diff_url": "https://github.com/ollama/ollama/pull/8113.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8113.patch",
"merged_at": null
} | Still missing: add 4 positions per embedding when creating a batch | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8113/reactions",
"total_count": 5,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 1,
"heart": 0,
"rocket": 0,
"eyes": 4
} | https://api.github.com/repos/ollama/ollama/issues/8113/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/685 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/685/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/685/comments | https://api.github.com/repos/ollama/ollama/issues/685/events | https://github.com/ollama/ollama/issues/685 | 1,923,133,063 | I_kwDOJ0Z1Ps5yoK6H | 685 | Question: where are all the `Modelfile`s? | {
"login": "jamesbraza",
"id": 8990777,
"node_id": "MDQ6VXNlcjg5OTA3Nzc=",
"avatar_url": "https://avatars.githubusercontent.com/u/8990777?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jamesbraza",
"html_url": "https://github.com/jamesbraza",
"followers_url": "https://api.github.com/users/jamesbraza/followers",
"following_url": "https://api.github.com/users/jamesbraza/following{/other_user}",
"gists_url": "https://api.github.com/users/jamesbraza/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jamesbraza/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jamesbraza/subscriptions",
"organizations_url": "https://api.github.com/users/jamesbraza/orgs",
"repos_url": "https://api.github.com/users/jamesbraza/repos",
"events_url": "https://api.github.com/users/jamesbraza/events{/privacy}",
"received_events_url": "https://api.github.com/users/jamesbraza/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 8 | 2023-10-03T01:50:20 | 2023-10-06T15:15:28 | 2023-10-04T02:40:43 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | https://ollama.ai/library has a lot of models. I would like to add a new model, and want to make sure it uses the GPU. So I am looking to refer to `Modelfile`s for models featured on https://ollama.ai/library.
Where is the source `Modelfile`s for the current "built in" models? | {
"login": "technovangelist",
"id": 633681,
"node_id": "MDQ6VXNlcjYzMzY4MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/technovangelist",
"html_url": "https://github.com/technovangelist",
"followers_url": "https://api.github.com/users/technovangelist/followers",
"following_url": "https://api.github.com/users/technovangelist/following{/other_user}",
"gists_url": "https://api.github.com/users/technovangelist/gists{/gist_id}",
"starred_url": "https://api.github.com/users/technovangelist/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/technovangelist/subscriptions",
"organizations_url": "https://api.github.com/users/technovangelist/orgs",
"repos_url": "https://api.github.com/users/technovangelist/repos",
"events_url": "https://api.github.com/users/technovangelist/events{/privacy}",
"received_events_url": "https://api.github.com/users/technovangelist/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/685/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/685/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2246 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2246/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2246/comments | https://api.github.com/repos/ollama/ollama/issues/2246/events | https://github.com/ollama/ollama/pull/2246 | 2,104,401,475 | PR_kwDOJ0Z1Ps5lRIEy | 2,246 | Don't disable GPUs on arm without AVX | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2024-01-28T23:38:00 | 2024-01-29T00:26:58 | 2024-01-29T00:26:55 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/2246",
"html_url": "https://github.com/ollama/ollama/pull/2246",
"diff_url": "https://github.com/ollama/ollama/pull/2246.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2246.patch",
"merged_at": "2024-01-29T00:26:55"
} | AVX is an x86 feature, so ARM should be excluded from the check.
Related to #1979 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2246/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2246/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2489 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2489/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2489/comments | https://api.github.com/repos/ollama/ollama/issues/2489/events | https://github.com/ollama/ollama/issues/2489 | 2,133,991,624 | I_kwDOJ0Z1Ps5_MiDI | 2,489 | what is smallest model that know about comp system administration, network admin, etc? | {
"login": "zinwelzl",
"id": 113045180,
"node_id": "U_kgDOBrzuvA",
"avatar_url": "https://avatars.githubusercontent.com/u/113045180?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/zinwelzl",
"html_url": "https://github.com/zinwelzl",
"followers_url": "https://api.github.com/users/zinwelzl/followers",
"following_url": "https://api.github.com/users/zinwelzl/following{/other_user}",
"gists_url": "https://api.github.com/users/zinwelzl/gists{/gist_id}",
"starred_url": "https://api.github.com/users/zinwelzl/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/zinwelzl/subscriptions",
"organizations_url": "https://api.github.com/users/zinwelzl/orgs",
"repos_url": "https://api.github.com/users/zinwelzl/repos",
"events_url": "https://api.github.com/users/zinwelzl/events{/privacy}",
"received_events_url": "https://api.github.com/users/zinwelzl/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] | closed | false | null | [] | null | 1 | 2024-02-14T10:15:25 | 2024-03-14T00:01:41 | 2024-03-14T00:01:40 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I run ollama locally and need some small model for help with system administration, network administration, etc?
I try few small but are really bad. | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2489/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2489/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/1872 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1872/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1872/comments | https://api.github.com/repos/ollama/ollama/issues/1872/events | https://github.com/ollama/ollama/issues/1872 | 2,072,777,144 | I_kwDOJ0Z1Ps57jBG4 | 1,872 | Error when install on Ubuntu 22.04 | {
"login": "dekogroup",
"id": 126862835,
"node_id": "U_kgDOB4_F8w",
"avatar_url": "https://avatars.githubusercontent.com/u/126862835?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dekogroup",
"html_url": "https://github.com/dekogroup",
"followers_url": "https://api.github.com/users/dekogroup/followers",
"following_url": "https://api.github.com/users/dekogroup/following{/other_user}",
"gists_url": "https://api.github.com/users/dekogroup/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dekogroup/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dekogroup/subscriptions",
"organizations_url": "https://api.github.com/users/dekogroup/orgs",
"repos_url": "https://api.github.com/users/dekogroup/repos",
"events_url": "https://api.github.com/users/dekogroup/events{/privacy}",
"received_events_url": "https://api.github.com/users/dekogroup/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5755339642,
"node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg",
"url": "https://api.github.com/repos/ollama/ollama/labels/linux",
"name": "linux",
"color": "516E70",
"default": false,
"description": ""
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 5 | 2024-01-09T16:55:31 | 2024-03-13T00:13:27 | 2024-03-13T00:13:27 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | curl https://ollama.ai/install.sh | sh
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 8354 0 8354 0 0 16163 0 --:--:-- --:--:-- --:--:-- 16189
>>> Downloading ollama...
######################################################################## 100.0%##O=# # Warning: Failed to open the file /tmp/tmp.AO1TPHxNpB/ollama: No such file or
Warning: directory
0.0%curl: (23) Failure writing output to destination
| {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1872/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1872/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6461 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6461/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6461/comments | https://api.github.com/repos/ollama/ollama/issues/6461/events | https://github.com/ollama/ollama/issues/6461 | 2,480,507,753 | I_kwDOJ0Z1Ps6T2Ytp | 6,461 | "/clear" command is not clearing history | {
"login": "devstefancho",
"id": 61320923,
"node_id": "MDQ6VXNlcjYxMzIwOTIz",
"avatar_url": "https://avatars.githubusercontent.com/u/61320923?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/devstefancho",
"html_url": "https://github.com/devstefancho",
"followers_url": "https://api.github.com/users/devstefancho/followers",
"following_url": "https://api.github.com/users/devstefancho/following{/other_user}",
"gists_url": "https://api.github.com/users/devstefancho/gists{/gist_id}",
"starred_url": "https://api.github.com/users/devstefancho/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/devstefancho/subscriptions",
"organizations_url": "https://api.github.com/users/devstefancho/orgs",
"repos_url": "https://api.github.com/users/devstefancho/repos",
"events_url": "https://api.github.com/users/devstefancho/events{/privacy}",
"received_events_url": "https://api.github.com/users/devstefancho/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2024-08-22T11:07:39 | 2024-08-22T17:00:16 | 2024-08-22T17:00:16 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
`Ctrl` + `l` is clearing history
but `/clear` command is not clearing history
https://github.com/user-attachments/assets/511ee922-9252-41d4-8b5f-ac324a75aaf1
### OS
macOS
### GPU
Apple
### CPU
Apple
### Ollama version
0.3.6 | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6461/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6461/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3473 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3473/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3473/comments | https://api.github.com/repos/ollama/ollama/issues/3473/events | https://github.com/ollama/ollama/pull/3473 | 2,222,435,172 | PR_kwDOJ0Z1Ps5rimQ_ | 3,473 | Add BrainSoup to compatible clients list | {
"login": "Nurgo",
"id": 11637957,
"node_id": "MDQ6VXNlcjExNjM3OTU3",
"avatar_url": "https://avatars.githubusercontent.com/u/11637957?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Nurgo",
"html_url": "https://github.com/Nurgo",
"followers_url": "https://api.github.com/users/Nurgo/followers",
"following_url": "https://api.github.com/users/Nurgo/following{/other_user}",
"gists_url": "https://api.github.com/users/Nurgo/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Nurgo/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Nurgo/subscriptions",
"organizations_url": "https://api.github.com/users/Nurgo/orgs",
"repos_url": "https://api.github.com/users/Nurgo/repos",
"events_url": "https://api.github.com/users/Nurgo/events{/privacy}",
"received_events_url": "https://api.github.com/users/Nurgo/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 1 | 2024-04-03T09:41:58 | 2024-05-06T20:42:16 | 2024-05-06T20:42:16 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3473",
"html_url": "https://github.com/ollama/ollama/pull/3473",
"diff_url": "https://github.com/ollama/ollama/pull/3473.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3473.patch",
"merged_at": "2024-05-06T20:42:16"
} | Hi there,
BrainSoup is a native multi-LLM client for Windows with advanced features such as local document indexing, RAG, multi-modality, multi-agent automation, code interpreter, sandboxed file system and the ability for agents to interact with the local system via customizable events and tools. More information can be found [here](https://www.nurgo-software.com/products/brainsoup).
The latest version is know compatible with Ollama, making BrainSoup an excellent choice for users looking to leverage Ollama's LLMs to their full potential.
I kindly ask for your review of this PR. Thanks for creating such an awesome platform! | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3473/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3473/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6431 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6431/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6431/comments | https://api.github.com/repos/ollama/ollama/issues/6431/events | https://github.com/ollama/ollama/issues/6431 | 2,474,397,214 | I_kwDOJ0Z1Ps6TfE4e | 6,431 | GLM4 tools support | {
"login": "napa3um",
"id": 665538,
"node_id": "MDQ6VXNlcjY2NTUzOA==",
"avatar_url": "https://avatars.githubusercontent.com/u/665538?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/napa3um",
"html_url": "https://github.com/napa3um",
"followers_url": "https://api.github.com/users/napa3um/followers",
"following_url": "https://api.github.com/users/napa3um/following{/other_user}",
"gists_url": "https://api.github.com/users/napa3um/gists{/gist_id}",
"starred_url": "https://api.github.com/users/napa3um/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/napa3um/subscriptions",
"organizations_url": "https://api.github.com/users/napa3um/orgs",
"repos_url": "https://api.github.com/users/napa3um/repos",
"events_url": "https://api.github.com/users/napa3um/events{/privacy}",
"received_events_url": "https://api.github.com/users/napa3um/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2024-08-19T22:57:39 | 2024-08-19T22:57:39 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | GML4 support tools - https://github.com/THUDM/GLM-4/blob/main/finetune_demo/README_en.md
How to fix the template in https://ollama.com/library/glm4 to make the ollama-tools mechanism work? | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6431/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6431/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/4823 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4823/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4823/comments | https://api.github.com/repos/ollama/ollama/issues/4823/events | https://github.com/ollama/ollama/issues/4823 | 2,334,769,831 | I_kwDOJ0Z1Ps6LKcKn | 4,823 | I encountered this error when converting the Tongyi-Finance-14B-Chat-Int4-AWQ model | {
"login": "wangkai111111",
"id": 74865581,
"node_id": "MDQ6VXNlcjc0ODY1NTgx",
"avatar_url": "https://avatars.githubusercontent.com/u/74865581?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/wangkai111111",
"html_url": "https://github.com/wangkai111111",
"followers_url": "https://api.github.com/users/wangkai111111/followers",
"following_url": "https://api.github.com/users/wangkai111111/following{/other_user}",
"gists_url": "https://api.github.com/users/wangkai111111/gists{/gist_id}",
"starred_url": "https://api.github.com/users/wangkai111111/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/wangkai111111/subscriptions",
"organizations_url": "https://api.github.com/users/wangkai111111/orgs",
"repos_url": "https://api.github.com/users/wangkai111111/repos",
"events_url": "https://api.github.com/users/wangkai111111/events{/privacy}",
"received_events_url": "https://api.github.com/users/wangkai111111/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 2 | 2024-06-05T02:27:12 | 2024-06-05T20:39:40 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
`(.venv) [root@bastion ollama]# python llm/llama.cpp/convert-hf-to-gguf.py ./model --outtype f16 --outfile converted.bin
INFO:hf-to-gguf:Loading model: model
INFO:gguf.gguf_writer:gguf: This GGUF file is for Little Endian only
INFO:hf-to-gguf:Set model parameters
INFO:hf-to-gguf:Set model tokenizer
Traceback (most recent call last):
File "/root/model/ollama/llm/llama.cpp/convert-hf-to-gguf.py", line 2865, in <module>
main()
File "/root/model/ollama/llm/llama.cpp/convert-hf-to-gguf.py", line 2850, in main
model_instance.set_vocab()
File "/root/model/ollama/llm/llama.cpp/convert-hf-to-gguf.py", line 1592, in set_vocab
self._set_vocab_qwen()
File "/root/model/ollama/llm/llama.cpp/convert-hf-to-gguf.py", line 530, in _set_vocab_qwen
assert len(merged) == 2
AssertionError
`
### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.1.41 | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4823/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4823/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/7973 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7973/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7973/comments | https://api.github.com/repos/ollama/ollama/issues/7973/events | https://github.com/ollama/ollama/pull/7973 | 2,723,621,546 | PR_kwDOJ0Z1Ps6EW47n | 7,973 | Document that `--format` now supports passing JSON Schemas | {
"login": "joliss",
"id": 524783,
"node_id": "MDQ6VXNlcjUyNDc4Mw==",
"avatar_url": "https://avatars.githubusercontent.com/u/524783?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joliss",
"html_url": "https://github.com/joliss",
"followers_url": "https://api.github.com/users/joliss/followers",
"following_url": "https://api.github.com/users/joliss/following{/other_user}",
"gists_url": "https://api.github.com/users/joliss/gists{/gist_id}",
"starred_url": "https://api.github.com/users/joliss/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/joliss/subscriptions",
"organizations_url": "https://api.github.com/users/joliss/orgs",
"repos_url": "https://api.github.com/users/joliss/repos",
"events_url": "https://api.github.com/users/joliss/events{/privacy}",
"received_events_url": "https://api.github.com/users/joliss/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | open | false | null | [] | null | 0 | 2024-12-06T17:53:16 | 2024-12-12T23:50:09 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7973",
"html_url": "https://github.com/ollama/ollama/pull/7973",
"diff_url": "https://github.com/ollama/ollama/pull/7973.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7973.patch",
"merged_at": null
} | JSON Schema support was added in #7900.
--------
I removed `e.g.` because I don't believe it supports anything else, right? Let me know if that's wrong. | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7973/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7973/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5466 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5466/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5466/comments | https://api.github.com/repos/ollama/ollama/issues/5466/events | https://github.com/ollama/ollama/pull/5466 | 2,389,318,735 | PR_kwDOJ0Z1Ps50XSIm | 5,466 | Fix clip model loading with unicode paths | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 1 | 2024-07-03T19:38:27 | 2024-07-05T15:17:01 | 2024-07-05T15:16:58 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5466",
"html_url": "https://github.com/ollama/ollama/pull/5466",
"diff_url": "https://github.com/ollama/ollama/pull/5466.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5466.patch",
"merged_at": "2024-07-05T15:16:58"
} | On windows, if the model dir contained unicode characters clip models would fail to load. This fixes the file name handling in clip.cpp to support utf16 on windows.
Fixes #5329 #4365 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5466/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5466/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/3581 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3581/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3581/comments | https://api.github.com/repos/ollama/ollama/issues/3581/events | https://github.com/ollama/ollama/issues/3581 | 2,236,296,576 | I_kwDOJ0Z1Ps6FSy2A | 3,581 | MacOS Ollama not binding to 0.0.0.0 | {
"login": "kellerkind84",
"id": 2842721,
"node_id": "MDQ6VXNlcjI4NDI3MjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/2842721?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kellerkind84",
"html_url": "https://github.com/kellerkind84",
"followers_url": "https://api.github.com/users/kellerkind84/followers",
"following_url": "https://api.github.com/users/kellerkind84/following{/other_user}",
"gists_url": "https://api.github.com/users/kellerkind84/gists{/gist_id}",
"starred_url": "https://api.github.com/users/kellerkind84/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/kellerkind84/subscriptions",
"organizations_url": "https://api.github.com/users/kellerkind84/orgs",
"repos_url": "https://api.github.com/users/kellerkind84/repos",
"events_url": "https://api.github.com/users/kellerkind84/events{/privacy}",
"received_events_url": "https://api.github.com/users/kellerkind84/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 32 | 2024-04-10T19:37:49 | 2025-01-22T12:51:14 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
So when set the OLLAMA_HOST to 0.0.0.0, I cannot access Ollama via the IP, but I can still access it via localhost.
### What did you expect to see?
I expect it to be available under <myIP>:11434
### Steps to reproduce
_No response_
### Are there any recent changes that introduced the issue?
_No response_
### OS
macOS
### Architecture
arm64
### Platform
_No response_
### Ollama version
0.1.31
### GPU
Apple
### GPU info
_No response_
### CPU
Apple
### Other software
_No response_ | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3581/reactions",
"total_count": 3,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3581/timeline | null | reopened | false |
https://api.github.com/repos/ollama/ollama/issues/2049 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2049/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2049/comments | https://api.github.com/repos/ollama/ollama/issues/2049/events | https://github.com/ollama/ollama/issues/2049 | 2,088,341,295 | I_kwDOJ0Z1Ps58eY8v | 2,049 | Embedding API could return empty embedding while using completion API from LiteLLM | {
"login": "James4Ever0",
"id": 103997068,
"node_id": "U_kgDOBjLejA",
"avatar_url": "https://avatars.githubusercontent.com/u/103997068?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/James4Ever0",
"html_url": "https://github.com/James4Ever0",
"followers_url": "https://api.github.com/users/James4Ever0/followers",
"following_url": "https://api.github.com/users/James4Ever0/following{/other_user}",
"gists_url": "https://api.github.com/users/James4Ever0/gists{/gist_id}",
"starred_url": "https://api.github.com/users/James4Ever0/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/James4Ever0/subscriptions",
"organizations_url": "https://api.github.com/users/James4Ever0/orgs",
"repos_url": "https://api.github.com/users/James4Ever0/repos",
"events_url": "https://api.github.com/users/James4Ever0/events{/privacy}",
"received_events_url": "https://api.github.com/users/James4Ever0/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677485533,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgJX3Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/embeddings",
"name": "embeddings",
"color": "76BF9F",
"default": false,
"description": "Issues around embeddings"
},
{
"id": 7706482389,
"node_id": "LA_kwDOJ0Z1Ps8AAAABy1eW1Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/api",
"name": "api",
"color": "bfdadc",
"default": false,
"description": ""
}
] | open | false | null | [] | null | 0 | 2024-01-18T13:54:52 | 2024-11-06T19:02:39 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | To reproduce:
Launch a LiteLLM service:
```bash
litellm --model ollama/openhermes2.5-mistral --drop_params
```
Call the service `/completion` API continuously first, meanwhile you call embedding API via Langchain, and hopefully during the very gap (very short) between each `/completion` call you get empty embedding from Langchain.
To call the `/completion` API:
```python
import os
os.environ['OPENAI_API_KEY'] = 'any'
os.environ['OPENAI_API_BASE'] = 'http://0.0.0.0:8000'
from contextlib import contextmanager
from langchain.llms import OpenAI
import tiktoken
def print_center(banner: str):
print(banner.center(50, "="))
class LLM:
"""
A class for running a Language Model Chain.
"""
def __init__(self, prompt: str, temperature=0, gpt_4=False):
"""
Initializes the LLM class.
Args:
prompt (PromptTemplate): The prompt template to use.
temperature (int): The temperature to use for the model.
gpt_4 (bool): Whether to use GPT-4 or Text-Davinci-003.
Side Effects:
Sets the class attributes.
"""
self.prompt = prompt
self.prompt_size = self.number_of_tokens(prompt)
self.temperature = temperature
self.gpt_4 = gpt_4
self.model_name = "gpt-4" if self.gpt_4 else "text-davinci-003"
self.max_tokens = 4097 * 2 if self.gpt_4 else 4097
self.show_init_config()
def show_init_config(self):
print_center("init params")
print(f"Model: {self.model_name}")
print(f"Max Tokens: {self.max_tokens}")
print(f"Prompt Size: {self.prompt_size}")
print(f"Temperature: {self.temperature}")
print_center("init config")
print(self.prompt)
def run(self, query):
"""
Runs the Language Model Chain.
Args:
code (str): The code to use for the chain.
**kwargs (dict): Additional keyword arguments.
Returns:
str: The generated text.
"""
llm = OpenAI(
temperature=self.temperature,
max_tokens=-1,
model_name=self.model_name,
disallowed_special=(), # to suppress error when special tokens within the input text (encode special tokens as normal text)
)
# chain = LLMChain(llm=llm, prompt=self.prompt)
chunk_list = []
print_center("query")
print(query)
print_center("response")
_input = "\n".join([self.prompt, query])
for chunk in llm.stream(input=_input):
print(chunk, end="", flush=True)
chunk_list.append(chunk)
print()
result = "".join(chunk_list)
return result
def number_of_tokens(self, text):
"""
Counts the number of tokens in a given text.
Args:
text (str): The text to count tokens for.
Returns:
int: The number of tokens in the text.
"""
encoding = tiktoken.encoding_for_model("gpt-4")
return len(encoding.encode(text, disallowed_special=()))
@contextmanager
def llm_context(prompt: str, temperature=0, gpt_4=False):
model = LLM(prompt, temperature=temperature, gpt_4=gpt_4)
try:
yield model
finally:
del model
if __name__ == "__main__":
while True:
with llm_context("You are a helpful assistant.") as model:
model.run("Write me a 100 words news.")
```
To call the embedding API:
```python
from langchain.embeddings import OllamaEmbeddings
ollama_emb = OllamaEmbeddings(
model="openhermes2.5-mistral:latest",
)
while True:
embed = ollama_emb.embed_query("Hello world")
print("empty embedding?", embed is None)
``` | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2049/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2049/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/946 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/946/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/946/comments | https://api.github.com/repos/ollama/ollama/issues/946/events | https://github.com/ollama/ollama/issues/946 | 1,967,119,694 | I_kwDOJ0Z1Ps51P91O | 946 | ollama show --modelfile gives incorrect FROM when multiple tags of base model are downloaded. | {
"login": "easp",
"id": 414705,
"node_id": "MDQ6VXNlcjQxNDcwNQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/414705?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/easp",
"html_url": "https://github.com/easp",
"followers_url": "https://api.github.com/users/easp/followers",
"following_url": "https://api.github.com/users/easp/following{/other_user}",
"gists_url": "https://api.github.com/users/easp/gists{/gist_id}",
"starred_url": "https://api.github.com/users/easp/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/easp/subscriptions",
"organizations_url": "https://api.github.com/users/easp/orgs",
"repos_url": "https://api.github.com/users/easp/repos",
"events_url": "https://api.github.com/users/easp/events{/privacy}",
"received_events_url": "https://api.github.com/users/easp/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5667396210,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2acg",
"url": "https://api.github.com/repos/ollama/ollama/labels/good%20first%20issue",
"name": "good first issue",
"color": "7057ff",
"default": true,
"description": "Good for newcomers"
}
] | closed | false | null | [] | null | 6 | 2023-10-29T19:13:09 | 2023-12-04T18:32:40 | 2023-12-04T18:32:40 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I've pulled two tags for codellama and created two new models, one based on each.
```
% ollama list
NAME ID SIZE MODIFIED
[...]
codellama:13b 9f438cb9cd58 7.4 GB 27 hours ago
codellama:13b-16k e86141f13814 7.4 GB 45 hours ago
codellama:34b-16k be95e5b84e7f 16 GB 41 seconds ago
codellama:34b-instruct-q3_K_M f534f618ea64 16 GB 27 hours ago
[...]
```
Note the file sizes; it's clear that ollama isn't confused about which data file to use for which.
However, if I `ollama show --modelfile`, the FROM just lists the base model name, not the tag.
```
% ollama show codellama:13b-16k --modelfile
# Modelfile generated by "ollama show"
# To build a new Modelfile based on this one, replace the FROM line with:
# FROM codellama:13b-16k
FROM library/codellama
TEMPLATE """[INST] <<SYS>>{{ .System }}<</SYS>>
{{ .Prompt }} [/INST]
"""
PARAMETER num_ctx 16384
PARAMETER rope_frequency_base 1e+06
PARAMETER stop "[INST]"
PARAMETER stop "[/INST]"
PARAMETER stop "<<SYS>>"
PARAMETER stop "<</SYS>>"
% ollama show codellama:34b-16k --modelfile
# Modelfile generated by "ollama show"
# To build a new Modelfile based on this one, replace the FROM line with:
# FROM codellama:34b-16k
FROM library/codellama
TEMPLATE """[INST] <<SYS>>{{ .System }}<</SYS>>
{{ .Prompt }} [/INST]
"""
PARAMETER num_ctx 16384
PARAMETER num_gqa 8
PARAMETER rope_frequency_base 1e+06
PARAMETER stop "[INST]"
PARAMETER stop "[/INST]"
PARAMETER stop "<<SYS>>"
PARAMETER stop "<</SYS>>"
```
| {
"login": "technovangelist",
"id": 633681,
"node_id": "MDQ6VXNlcjYzMzY4MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/technovangelist",
"html_url": "https://github.com/technovangelist",
"followers_url": "https://api.github.com/users/technovangelist/followers",
"following_url": "https://api.github.com/users/technovangelist/following{/other_user}",
"gists_url": "https://api.github.com/users/technovangelist/gists{/gist_id}",
"starred_url": "https://api.github.com/users/technovangelist/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/technovangelist/subscriptions",
"organizations_url": "https://api.github.com/users/technovangelist/orgs",
"repos_url": "https://api.github.com/users/technovangelist/repos",
"events_url": "https://api.github.com/users/technovangelist/events{/privacy}",
"received_events_url": "https://api.github.com/users/technovangelist/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/946/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/946/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2743 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2743/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2743/comments | https://api.github.com/repos/ollama/ollama/issues/2743/events | https://github.com/ollama/ollama/issues/2743 | 2,152,745,980 | I_kwDOJ0Z1Ps6AUEv8 | 2,743 | What is the different between "gemma-instruct", "gemma-text" and "gemma". Same to other models. | {
"login": "XinyueZ",
"id": 7869833,
"node_id": "MDQ6VXNlcjc4Njk4MzM=",
"avatar_url": "https://avatars.githubusercontent.com/u/7869833?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/XinyueZ",
"html_url": "https://github.com/XinyueZ",
"followers_url": "https://api.github.com/users/XinyueZ/followers",
"following_url": "https://api.github.com/users/XinyueZ/following{/other_user}",
"gists_url": "https://api.github.com/users/XinyueZ/gists{/gist_id}",
"starred_url": "https://api.github.com/users/XinyueZ/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/XinyueZ/subscriptions",
"organizations_url": "https://api.github.com/users/XinyueZ/orgs",
"repos_url": "https://api.github.com/users/XinyueZ/repos",
"events_url": "https://api.github.com/users/XinyueZ/events{/privacy}",
"received_events_url": "https://api.github.com/users/XinyueZ/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 1 | 2024-02-25T12:39:09 | 2024-02-26T11:01:38 | 2024-02-26T11:01:38 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | null | {
"login": "XinyueZ",
"id": 7869833,
"node_id": "MDQ6VXNlcjc4Njk4MzM=",
"avatar_url": "https://avatars.githubusercontent.com/u/7869833?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/XinyueZ",
"html_url": "https://github.com/XinyueZ",
"followers_url": "https://api.github.com/users/XinyueZ/followers",
"following_url": "https://api.github.com/users/XinyueZ/following{/other_user}",
"gists_url": "https://api.github.com/users/XinyueZ/gists{/gist_id}",
"starred_url": "https://api.github.com/users/XinyueZ/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/XinyueZ/subscriptions",
"organizations_url": "https://api.github.com/users/XinyueZ/orgs",
"repos_url": "https://api.github.com/users/XinyueZ/repos",
"events_url": "https://api.github.com/users/XinyueZ/events{/privacy}",
"received_events_url": "https://api.github.com/users/XinyueZ/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2743/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2743/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5166 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5166/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5166/comments | https://api.github.com/repos/ollama/ollama/issues/5166/events | https://github.com/ollama/ollama/issues/5166 | 2,364,002,946 | I_kwDOJ0Z1Ps6M59KC | 5,166 | In dockerGPU containers ollama still uses the CPU | {
"login": "Zxyy-mo",
"id": 48347974,
"node_id": "MDQ6VXNlcjQ4MzQ3OTc0",
"avatar_url": "https://avatars.githubusercontent.com/u/48347974?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Zxyy-mo",
"html_url": "https://github.com/Zxyy-mo",
"followers_url": "https://api.github.com/users/Zxyy-mo/followers",
"following_url": "https://api.github.com/users/Zxyy-mo/following{/other_user}",
"gists_url": "https://api.github.com/users/Zxyy-mo/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Zxyy-mo/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Zxyy-mo/subscriptions",
"organizations_url": "https://api.github.com/users/Zxyy-mo/orgs",
"repos_url": "https://api.github.com/users/Zxyy-mo/repos",
"events_url": "https://api.github.com/users/Zxyy-mo/events{/privacy}",
"received_events_url": "https://api.github.com/users/Zxyy-mo/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 6 | 2024-06-20T09:43:38 | 2024-06-21T15:38:53 | 2024-06-21T15:38:53 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
## desc
I implemented the deployment following the official Docker GPU container tutorial. And successfully got the graphics card information using nvidia-smi in the Docker container.
I'm using a nvidia discrete graphics card 3090
```info
# ollama ps
NAME ID SIZE PROCESSOR UNTIL
# ollama ps
NAME ID SIZE PROCESSOR UNTIL
qwen2:7b e0d4e1163c58 4.7 GB 100% CPU 4 minutes from now
# nvidia-smi
Thu Jun 20 09:36:11 2024
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 525.147.05 Driver Version: 525.147.05 CUDA Version: 12.0 |
|-------------------------------+----------------------+----------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|===============================+======================+======================|
| 0 NVIDIA GeForce ... Off | 00000000:02:00.0 On | N/A |
| 0% 38C P8 16W / 350W | 114MiB / 24576MiB | 0% Default |
| | | N/A |
+-------------------------------+----------------------+----------------------+
+-----------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=============================================================================|
+-----------------------------------------------------------------------------+
# ollama -v
ollama version is 0.1.44
# ollama -v
ollama version is 0.1.44
# ollama ps
NAME ID SIZE PROCESSOR UNTIL
qwen2:7b e0d4e1163c58 4.7 GB 100% CPU 37 seconds from now
#
```
### OS
Linux, Docker
### GPU
Nvidia
### CPU
Intel
### Ollama version
ollama version is 0.1.44 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5166/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5166/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/8122 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8122/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8122/comments | https://api.github.com/repos/ollama/ollama/issues/8122/events | https://github.com/ollama/ollama/pull/8122 | 2,743,390,539 | PR_kwDOJ0Z1Ps6FaUC2 | 8,122 | build: streamline build | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | open | false | null | [] | null | 0 | 2024-12-16T21:00:53 | 2024-12-16T21:14:49 | null | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8122",
"html_url": "https://github.com/ollama/ollama/pull/8122",
"diff_url": "https://github.com/ollama/ollama/pull/8122.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8122.patch",
"merged_at": null
} | This wiring was intended to make a faster developer build by disabling flash attention but the added complexity and friction on updates makes this less useful | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8122/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8122/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1106 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1106/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1106/comments | https://api.github.com/repos/ollama/ollama/issues/1106/events | https://github.com/ollama/ollama/pull/1106 | 1,989,866,478 | PR_kwDOJ0Z1Ps5fQfz5 | 1,106 | Add Dart library to README.md | {
"login": "breitburg",
"id": 25728414,
"node_id": "MDQ6VXNlcjI1NzI4NDE0",
"avatar_url": "https://avatars.githubusercontent.com/u/25728414?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/breitburg",
"html_url": "https://github.com/breitburg",
"followers_url": "https://api.github.com/users/breitburg/followers",
"following_url": "https://api.github.com/users/breitburg/following{/other_user}",
"gists_url": "https://api.github.com/users/breitburg/gists{/gist_id}",
"starred_url": "https://api.github.com/users/breitburg/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/breitburg/subscriptions",
"organizations_url": "https://api.github.com/users/breitburg/orgs",
"repos_url": "https://api.github.com/users/breitburg/repos",
"events_url": "https://api.github.com/users/breitburg/events{/privacy}",
"received_events_url": "https://api.github.com/users/breitburg/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 1 | 2023-11-13T04:25:49 | 2023-11-14T04:08:36 | 2023-11-13T19:50:42 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/1106",
"html_url": "https://github.com/ollama/ollama/pull/1106",
"diff_url": "https://github.com/ollama/ollama/pull/1106.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1106.patch",
"merged_at": "2023-11-13T19:50:42"
} | Good afternoon!
I have completed the first version of the Ollama library for Dart, making it possible to integrate Ollama into Flutter applications. I thought it would be nice to mention it in the readme file.
 | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1106/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1106/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6317 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6317/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6317/comments | https://api.github.com/repos/ollama/ollama/issues/6317/events | https://github.com/ollama/ollama/issues/6317 | 2,459,986,027 | I_kwDOJ0Z1Ps6SoGhr | 6,317 | Feature request : Tools support of Qwen2 | {
"login": "trinhkiet0105",
"id": 76981747,
"node_id": "MDQ6VXNlcjc2OTgxNzQ3",
"avatar_url": "https://avatars.githubusercontent.com/u/76981747?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/trinhkiet0105",
"html_url": "https://github.com/trinhkiet0105",
"followers_url": "https://api.github.com/users/trinhkiet0105/followers",
"following_url": "https://api.github.com/users/trinhkiet0105/following{/other_user}",
"gists_url": "https://api.github.com/users/trinhkiet0105/gists{/gist_id}",
"starred_url": "https://api.github.com/users/trinhkiet0105/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/trinhkiet0105/subscriptions",
"organizations_url": "https://api.github.com/users/trinhkiet0105/orgs",
"repos_url": "https://api.github.com/users/trinhkiet0105/repos",
"events_url": "https://api.github.com/users/trinhkiet0105/events{/privacy}",
"received_events_url": "https://api.github.com/users/trinhkiet0105/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 5 | 2024-08-12T03:54:30 | 2024-09-02T23:49:18 | 2024-09-02T23:49:18 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | **Why ?**
So i just found out that Qwen2 have tools support. However, ollama current do not have tools support for qwen2 models. And there is a section of [Qwen2 github talking about ollama on tools use](https://github.com/QwenLM/Qwen2?tab=readme-ov-file#-run-locally
). And this seems prombles of ollama in older versions. I think this should be resolvable with version 0.3.x
**Request**
- I would like to request tools support for Qwen2 models
- If possible I would like to know how the process of adding tools support for a model (like ollama team can add tools to Qwen2 directly or participation of Qwen2 team is required)
Thanks for reading | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6317/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6317/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7517 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7517/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7517/comments | https://api.github.com/repos/ollama/ollama/issues/7517/events | https://github.com/ollama/ollama/pull/7517 | 2,636,321,952 | PR_kwDOJ0Z1Ps6A-N6i | 7,517 | Doc updates for supporting Llama3.2 | {
"login": "frances720",
"id": 8753634,
"node_id": "MDQ6VXNlcjg3NTM2MzQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/8753634?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/frances720",
"html_url": "https://github.com/frances720",
"followers_url": "https://api.github.com/users/frances720/followers",
"following_url": "https://api.github.com/users/frances720/following{/other_user}",
"gists_url": "https://api.github.com/users/frances720/gists{/gist_id}",
"starred_url": "https://api.github.com/users/frances720/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/frances720/subscriptions",
"organizations_url": "https://api.github.com/users/frances720/orgs",
"repos_url": "https://api.github.com/users/frances720/repos",
"events_url": "https://api.github.com/users/frances720/events{/privacy}",
"received_events_url": "https://api.github.com/users/frances720/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2024-11-05T19:44:45 | 2024-11-15T23:41:09 | 2024-11-11T03:04:24 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/7517",
"html_url": "https://github.com/ollama/ollama/pull/7517",
"diff_url": "https://github.com/ollama/ollama/pull/7517.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7517.patch",
"merged_at": "2024-11-11T03:04:24"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7517/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7517/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/757 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/757/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/757/comments | https://api.github.com/repos/ollama/ollama/issues/757/events | https://github.com/ollama/ollama/pull/757 | 1,938,405,377 | PR_kwDOJ0Z1Ps5cjJ1f | 757 | cleanup format time | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2023-10-11T18:06:06 | 2023-10-11T18:12:30 | 2023-10-11T18:12:29 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/757",
"html_url": "https://github.com/ollama/ollama/pull/757",
"diff_url": "https://github.com/ollama/ollama/pull/757.diff",
"patch_url": "https://github.com/ollama/ollama/pull/757.patch",
"merged_at": "2023-10-11T18:12:29"
} | only `HumanTime` is actually being used | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/757/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/757/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5870 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5870/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5870/comments | https://api.github.com/repos/ollama/ollama/issues/5870/events | https://github.com/ollama/ollama/issues/5870 | 2,424,725,170 | I_kwDOJ0Z1Ps6Qhl6y | 5,870 | The embeddings api interface is not working properly. | {
"login": "xldistance",
"id": 29418474,
"node_id": "MDQ6VXNlcjI5NDE4NDc0",
"avatar_url": "https://avatars.githubusercontent.com/u/29418474?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/xldistance",
"html_url": "https://github.com/xldistance",
"followers_url": "https://api.github.com/users/xldistance/followers",
"following_url": "https://api.github.com/users/xldistance/following{/other_user}",
"gists_url": "https://api.github.com/users/xldistance/gists{/gist_id}",
"starred_url": "https://api.github.com/users/xldistance/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/xldistance/subscriptions",
"organizations_url": "https://api.github.com/users/xldistance/orgs",
"repos_url": "https://api.github.com/users/xldistance/repos",
"events_url": "https://api.github.com/users/xldistance/events{/privacy}",
"received_events_url": "https://api.github.com/users/xldistance/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 11 | 2024-07-23T09:33:21 | 2025-01-04T10:50:19 | 2024-07-30T17:55:02 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I use the bge-m3 model in graphrag with the following parameters
```
embeddings:
## parallelization: override the global parallelization settings for embeddings
async_mode: asyncio
llm:
api_key:
type: openai_embedding # or azure_openai_embedding
model: chatfire/bge-m3:q8_0
api_base: http://localhost:11434/api
```
The following error is returned
```
17:11:30,126 httpx INFO HTTP Request: POST http://localhost:11434/api/embeddings "HTTP/1.1 200 OK"
17:11:30,129 datashaper.workflow.workflow ERROR Error executing verb "text_embed" in create_final_entities: 'NoneType' object is not iterable
Traceback (most recent call last):
File "E:\Langchain-Chatchat\glut\lib\site-packages\datashaper\workflow\workflow.py", line 415, in _execute_verb
result = await result
File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\index\verbs\text\embed\text_embed.py", line 105, in text_embed
return await _text_embed_in_memory(
File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\index\verbs\text\embed\text_embed.py", line 130, in _text_embed_in_memory
result = await strategy_exec(texts, callbacks, cache, strategy_args)
File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\index\verbs\text\embed\strategies\openai.py", line 61, in run
embeddings = await _execute(llm, text_batches, ticker, semaphore)
File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\index\verbs\text\embed\strategies\openai.py", line 105, in _execute
results = await asyncio.gather(*futures)
File "E:\Langchain-Chatchat\glut\lib\asyncio\tasks.py", line 304, in __wakeup
future.result()
File "E:\Langchain-Chatchat\glut\lib\asyncio\tasks.py", line 232, in __step
result = coro.send(None)
File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\index\verbs\text\embed\strategies\openai.py", line 99, in embed
chunk_embeddings = await llm(chunk)
File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\base\caching_llm.py", line 104, in __call__
result = await self._delegate(input, **kwargs)
File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\base\rate_limiting_llm.py", line 177, in __call__
result, start = await execute_with_retry()
File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\base\rate_limiting_llm.py", line 159, in execute_with_retry
async for attempt in retryer:
File "E:\Langchain-Chatchat\glut\lib\site-packages\tenacity\_asyncio.py", line 71, in __anext__
do = self.iter(retry_state=self._retry_state)
File "E:\Langchain-Chatchat\glut\lib\site-packages\tenacity\__init__.py", line 314, in iter
return fut.result()
File "E:\Langchain-Chatchat\glut\lib\concurrent\futures\_base.py", line 451, in result
return self.__get_result()
File "E:\Langchain-Chatchat\glut\lib\concurrent\futures\_base.py", line 403, in __get_result
raise self._exception
File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\base\rate_limiting_llm.py", line 165, in execute_with_retry
return await do_attempt(), start
File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\base\rate_limiting_llm.py", line 147, in do_attempt
return await self._delegate(input, **kwargs)
File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\base\base_llm.py", line 49, in __call__
return await self._invoke(input, **kwargs)
File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\base\base_llm.py", line 53, in _invoke
output = await self._execute_llm(input, **kwargs)
File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\openai\openai_embeddings_llm.py", line 36, in _execute_llm
embedding = await self.client.embeddings.create(
File "E:\Langchain-Chatchat\glut\lib\site-packages\openai\resources\embeddings.py", line 215, in create
return await self._post(
File "E:\Langchain-Chatchat\glut\lib\site-packages\openai\_base_client.py", line 1826, in post
return await self.request(cast_to, opts, stream=stream, stream_cls=stream_cls)
File "E:\Langchain-Chatchat\glut\lib\site-packages\openai\_base_client.py", line 1519, in request
return await self._request(
File "E:\Langchain-Chatchat\glut\lib\site-packages\openai\_base_client.py", line 1622, in _request
return await self._process_response(
File "E:\Langchain-Chatchat\glut\lib\site-packages\openai\_base_client.py", line 1714, in _process_response
return await api_response.parse()
File "E:\Langchain-Chatchat\glut\lib\site-packages\openai\_response.py", line 419, in parse
parsed = self._options.post_parser(parsed)
File "E:\Langchain-Chatchat\glut\lib\site-packages\openai\resources\embeddings.py", line 203, in parser
for embedding in obj.data:
TypeError: 'NoneType' object is not iterable
17:11:30,131 graphrag.index.reporting.file_workflow_callbacks INFO Error executing verb "text_embed" in create_final_entities: 'NoneType' object is not iterable details=None
17:11:30,142 graphrag.index.run ERROR error running workflow create_final_entities
```
### OS
Windows
### GPU
Nvidia
### CPU
AMD
### Ollama version
0.2.8 | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5870/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5870/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5291 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5291/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5291/comments | https://api.github.com/repos/ollama/ollama/issues/5291/events | https://github.com/ollama/ollama/issues/5291 | 2,374,355,746 | I_kwDOJ0Z1Ps6Nhcsi | 5,291 | 请上架cogvlm2 | {
"login": "enryteam",
"id": 20081090,
"node_id": "MDQ6VXNlcjIwMDgxMDkw",
"avatar_url": "https://avatars.githubusercontent.com/u/20081090?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/enryteam",
"html_url": "https://github.com/enryteam",
"followers_url": "https://api.github.com/users/enryteam/followers",
"following_url": "https://api.github.com/users/enryteam/following{/other_user}",
"gists_url": "https://api.github.com/users/enryteam/gists{/gist_id}",
"starred_url": "https://api.github.com/users/enryteam/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/enryteam/subscriptions",
"organizations_url": "https://api.github.com/users/enryteam/orgs",
"repos_url": "https://api.github.com/users/enryteam/repos",
"events_url": "https://api.github.com/users/enryteam/events{/privacy}",
"received_events_url": "https://api.github.com/users/enryteam/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] | open | false | null | [] | null | 1 | 2024-06-26T05:57:32 | 2024-06-26T12:12:42 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | https://huggingface.co/THUDM/cogvlm2-llama3-chinese-chat-19B
thanks 谢谢
ollama0.1.43
error format not yet support!
错误格式尚不支持!
错误格式尚不支持!
搞了多次 均报错 | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5291/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5291/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/3544 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3544/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3544/comments | https://api.github.com/repos/ollama/ollama/issues/3544/events | https://github.com/ollama/ollama/issues/3544 | 2,232,444,606 | I_kwDOJ0Z1Ps6FEGa- | 3,544 | ollama 0.1.31 Segmentation fault (core dumped) | {
"login": "zhqfdn",
"id": 25156863,
"node_id": "MDQ6VXNlcjI1MTU2ODYz",
"avatar_url": "https://avatars.githubusercontent.com/u/25156863?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/zhqfdn",
"html_url": "https://github.com/zhqfdn",
"followers_url": "https://api.github.com/users/zhqfdn/followers",
"following_url": "https://api.github.com/users/zhqfdn/following{/other_user}",
"gists_url": "https://api.github.com/users/zhqfdn/gists{/gist_id}",
"starred_url": "https://api.github.com/users/zhqfdn/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/zhqfdn/subscriptions",
"organizations_url": "https://api.github.com/users/zhqfdn/orgs",
"repos_url": "https://api.github.com/users/zhqfdn/repos",
"events_url": "https://api.github.com/users/zhqfdn/events{/privacy}",
"received_events_url": "https://api.github.com/users/zhqfdn/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 3 | 2024-04-09T01:36:09 | 2024-05-01T16:43:13 | 2024-05-01T16:43:13 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
[root@localhost ~]# cat /etc/redhat-release
AlmaLinux release 9.3 (Shamrock Pampas Cat)
[root@localhost ~]# ollama -v
Warning: could not connect to a running Ollama instance
Warning: client version is 0.1.30
[root@localhost ~]# ./ollama -v
Segmentation fault (core dumped)
[root@localhost ~]#
### What did you expect to see?
[root@localhost ~]# ./ollama -v
Segmentation fault (core dumped)
### Steps to reproduce
[root@localhost ~]# ./ollama -v
Segmentation fault (core dumped)
### Are there any recent changes that introduced the issue?
[root@localhost ~]# ./ollama -v
Segmentation fault (core dumped)
### OS
Linux
### Architecture
amd64
### Platform
_No response_
### Ollama version
0.1.31
### GPU
Nvidia
### GPU info
[root@localhost ~]# nvidia-smi
Tue Apr 9 09:31:38 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 550.54.15 Driver Version: 550.54.15 CUDA Version: 12.4 |
|-----------------------------------------+------------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+========================+======================|
| 0 Tesla T4 Off | 00000000:06:00.0 Off | 0 |
| N/A 34C P0 27W / 70W | 0MiB / 15360MiB | 0% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
| 1 Tesla T4 Off | 00000000:86:00.0 Off | 0 |
| N/A 34C P0 26W / 70W | 0MiB / 15360MiB | 0% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
+-----------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=========================================================================================|
| No running processes found |
+-----------------------------------------------------------------------------------------+
### CPU
Intel
### Other software
AlmaLinux release 9.3 (Shamrock Pampas Cat) | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3544/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3544/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5712 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5712/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5712/comments | https://api.github.com/repos/ollama/ollama/issues/5712/events | https://github.com/ollama/ollama/pull/5712 | 2,409,800,448 | PR_kwDOJ0Z1Ps51cb6s | 5,712 | Add Windows arm64 support to official builds | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 14 | 2024-07-15T23:18:15 | 2024-09-20T20:09:41 | 2024-09-20T20:09:38 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5712",
"html_url": "https://github.com/ollama/ollama/pull/5712",
"diff_url": "https://github.com/ollama/ollama/pull/5712.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5712.patch",
"merged_at": "2024-09-20T20:09:38"
} | Wire up CI and build rigging to generate a unified Windows installer with x64 and arm64 payloads. At install time, the correct binaries will be installed for the platform.
I was unable to find a combination of hand-picked msvc redist DLLs manually that yielded a working setup on a pristine Windows 11 install, but running the vc_redist installer works reliably, so for arm64, we run the nested installer conditionally. If it is already installed, that step will be skipped.
Fixes #2589
Note: I've tested most of the CI steps in the PR, but signing isn't yet verified and might require minor fixes on the first release after this merges.
Resulting build artifacts: (Note: current OllamaSetup.exe with only x64 binaries is 273MB)
```
% ls -lh dist/
total 932M
-rw-r--r-- 1 daniel 197609 12K Jul 17 09:24 ollama_welcome.ps1
-rwxr-xr-x 1 daniel 197609 291M Jul 17 09:27 OllamaSetup.exe*
-rw-r--r-- 1 daniel 197609 649M Jul 17 09:27 ollama-windows-amd64.zip
-rw-r--r-- 1 daniel 197609 20M Jul 19 15:41 ollama-windows-arm64.zip
drwxr-xr-x 1 daniel 197609 0 Jul 17 09:24 windows-amd64/
-rwxr-xr-x 1 daniel 197609 5.9M Jul 17 09:24 windows-amd64-app.exe*
drwxr-xr-x 1 daniel 197609 0 Jul 16 15:53 windows-arm64/
-rwxr-xr-x 1 daniel 197609 5.5M Jul 16 16:12 windows-arm64-app.exe*
% du -sh dist/windows-a*64
2.1G dist/windows-amd64
37M dist/windows-arm64
```
On a Snapdragon X 12-core laptop:
```
> ollama run --verbose llama3 why is the sky blue
...
total duration: 23.6819409s
load duration: 4.738127s
prompt eval count: 16 token(s)
prompt eval duration: 430.297ms
prompt eval rate: 37.18 tokens/s
eval count: 348 token(s)
eval duration: 18.513796s
eval rate: 18.80 tokens/s
```
| {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5712/reactions",
"total_count": 3,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5712/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/2120 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2120/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2120/comments | https://api.github.com/repos/ollama/ollama/issues/2120/events | https://github.com/ollama/ollama/issues/2120 | 2,092,519,133 | I_kwDOJ0Z1Ps58uU7d | 2,120 | How to install libnvidia-ml.so? | {
"login": "silverwind63",
"id": 104142549,
"node_id": "U_kgDOBjUW1Q",
"avatar_url": "https://avatars.githubusercontent.com/u/104142549?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/silverwind63",
"html_url": "https://github.com/silverwind63",
"followers_url": "https://api.github.com/users/silverwind63/followers",
"following_url": "https://api.github.com/users/silverwind63/following{/other_user}",
"gists_url": "https://api.github.com/users/silverwind63/gists{/gist_id}",
"starred_url": "https://api.github.com/users/silverwind63/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/silverwind63/subscriptions",
"organizations_url": "https://api.github.com/users/silverwind63/orgs",
"repos_url": "https://api.github.com/users/silverwind63/repos",
"events_url": "https://api.github.com/users/silverwind63/events{/privacy}",
"received_events_url": "https://api.github.com/users/silverwind63/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 4 | 2024-01-21T10:16:16 | 2024-01-27T11:25:08 | 2024-01-26T21:06:50 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi guys! I have been using ollama with ollama webui this month.However,it output
```
WARNING:
You should always run with libnvidia-ml.so that is installed with your
NVIDIA Display Driver. By default it's installed in /usr/lib and /usr/lib64.
libnvidia-ml.so in GDK package is a stub library that is attached only for
build purposes (e.g. machine that you build your application doesn't have
to have Display Driver installed).
```
And whenever I want to run any model(which is capable to load it and the speed is about 5 tokens/s) it will always run into cuda memory error.
My system:
RAM:16GB
GPU:3060ti 8GB
SYSTEM:archlinux
Kernel:6.7.0-arch3-1
Nvidia GPU Driver:nvidia-dkms 545.29.06-1
I have also installed following package which is related to nvidia:
```
lib32-nvidia-utils 545.29.06-1
libnvidia-container 1.14.3-1
libnvidia-container-tools 1.14.3-1
libva-nvidia-driver-git 0.0.11.r1.gea6d862-1
nvidia-container-toolkit 1.14.3-9
nvidia-docker-compose 0.1.6-1
nvidia-utils 545.29.06-1
``` | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2120/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2120/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2639 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2639/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2639/comments | https://api.github.com/repos/ollama/ollama/issues/2639/events | https://github.com/ollama/ollama/issues/2639 | 2,147,069,563 | I_kwDOJ0Z1Ps5_-a57 | 2,639 | History via up arrow and down arrow not working on windows using `ollama run` | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg",
"url": "https://api.github.com/repos/ollama/ollama/labels/windows",
"name": "windows",
"color": "0052CC",
"default": false,
"description": ""
}
] | closed | false | null | [] | null | 0 | 2024-02-21T15:39:35 | 2024-03-26T22:21:57 | 2024-03-26T22:21:57 | MEMBER | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2639/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2639/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/612 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/612/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/612/comments | https://api.github.com/repos/ollama/ollama/issues/612/events | https://github.com/ollama/ollama/pull/612 | 1,914,484,120 | PR_kwDOJ0Z1Ps5bSJCL | 612 | prune empty directories | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2023-09-27T00:40:27 | 2023-09-29T18:23:41 | 2023-09-29T18:23:40 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/612",
"html_url": "https://github.com/ollama/ollama/pull/612",
"diff_url": "https://github.com/ollama/ollama/pull/612.diff",
"patch_url": "https://github.com/ollama/ollama/pull/612.patch",
"merged_at": "2023-09-29T18:23:40"
} | Resolves #270 | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/612/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/612/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1033 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1033/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1033/comments | https://api.github.com/repos/ollama/ollama/issues/1033/events | https://github.com/ollama/ollama/issues/1033 | 1,981,678,318 | I_kwDOJ0Z1Ps52HgLu | 1,033 | Are these system specs good enough for any models? | {
"login": "simoovara",
"id": 100516318,
"node_id": "U_kgDOBf3B3g",
"avatar_url": "https://avatars.githubusercontent.com/u/100516318?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/simoovara",
"html_url": "https://github.com/simoovara",
"followers_url": "https://api.github.com/users/simoovara/followers",
"following_url": "https://api.github.com/users/simoovara/following{/other_user}",
"gists_url": "https://api.github.com/users/simoovara/gists{/gist_id}",
"starred_url": "https://api.github.com/users/simoovara/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/simoovara/subscriptions",
"organizations_url": "https://api.github.com/users/simoovara/orgs",
"repos_url": "https://api.github.com/users/simoovara/repos",
"events_url": "https://api.github.com/users/simoovara/events{/privacy}",
"received_events_url": "https://api.github.com/users/simoovara/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 6 | 2023-11-07T15:50:36 | 2023-11-07T21:32:31 | 2023-11-07T21:32:14 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Just a question, I have an old laptop that i turned into a server with Ubuntu LTS. It has an AMD E1-6015 APU and 8gb of ram. I would like to know if that's enough to run any of these models, thank you! | {
"login": "simoovara",
"id": 100516318,
"node_id": "U_kgDOBf3B3g",
"avatar_url": "https://avatars.githubusercontent.com/u/100516318?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/simoovara",
"html_url": "https://github.com/simoovara",
"followers_url": "https://api.github.com/users/simoovara/followers",
"following_url": "https://api.github.com/users/simoovara/following{/other_user}",
"gists_url": "https://api.github.com/users/simoovara/gists{/gist_id}",
"starred_url": "https://api.github.com/users/simoovara/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/simoovara/subscriptions",
"organizations_url": "https://api.github.com/users/simoovara/orgs",
"repos_url": "https://api.github.com/users/simoovara/repos",
"events_url": "https://api.github.com/users/simoovara/events{/privacy}",
"received_events_url": "https://api.github.com/users/simoovara/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1033/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1033/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5907 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5907/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5907/comments | https://api.github.com/repos/ollama/ollama/issues/5907/events | https://github.com/ollama/ollama/issues/5907 | 2,427,317,383 | I_kwDOJ0Z1Ps6QreyH | 5,907 | Support token embeddings for `v1/embeddings` | {
"login": "WoJiaoFuXiaoYun",
"id": 30924105,
"node_id": "MDQ6VXNlcjMwOTI0MTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/30924105?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/WoJiaoFuXiaoYun",
"html_url": "https://github.com/WoJiaoFuXiaoYun",
"followers_url": "https://api.github.com/users/WoJiaoFuXiaoYun/followers",
"following_url": "https://api.github.com/users/WoJiaoFuXiaoYun/following{/other_user}",
"gists_url": "https://api.github.com/users/WoJiaoFuXiaoYun/gists{/gist_id}",
"starred_url": "https://api.github.com/users/WoJiaoFuXiaoYun/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/WoJiaoFuXiaoYun/subscriptions",
"organizations_url": "https://api.github.com/users/WoJiaoFuXiaoYun/orgs",
"repos_url": "https://api.github.com/users/WoJiaoFuXiaoYun/repos",
"events_url": "https://api.github.com/users/WoJiaoFuXiaoYun/events{/privacy}",
"received_events_url": "https://api.github.com/users/WoJiaoFuXiaoYun/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 7706482389,
"node_id": "LA_kwDOJ0Z1Ps8AAAABy1eW1Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/api",
"name": "api",
"color": "bfdadc",
"default": false,
"description": ""
}
] | open | false | null | [] | null | 3 | 2024-07-24T11:17:36 | 2024-11-06T01:00:33 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
When encoding with `tiktoken`, the interface is no longer compatible
```
tiktoken.get_encoding("cl100k_base").encode(text)
```
```json
{
"input": [30624,99849,64479,51392,31809,29207,233,45829],
"model": "nomic-embed-text"
}
```
```
{
"error": {
"message": "invalid input type",
"type": "api_error",
"param": null,
"code": null
}
}
```
I hope it can be compatible, thank you
### OS
_No response_
### GPU
_No response_
### CPU
_No response_
### Ollama version
ollama version is 0.2.7 | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5907/reactions",
"total_count": 2,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/ollama/ollama/issues/5907/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/5107 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5107/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5107/comments | https://api.github.com/repos/ollama/ollama/issues/5107/events | https://github.com/ollama/ollama/issues/5107 | 2,358,707,183 | I_kwDOJ0Z1Ps6MlwPv | 5,107 | ollama 模型授权 | {
"login": "yawzhe",
"id": 127652671,
"node_id": "U_kgDOB5vTPw",
"avatar_url": "https://avatars.githubusercontent.com/u/127652671?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yawzhe",
"html_url": "https://github.com/yawzhe",
"followers_url": "https://api.github.com/users/yawzhe/followers",
"following_url": "https://api.github.com/users/yawzhe/following{/other_user}",
"gists_url": "https://api.github.com/users/yawzhe/gists{/gist_id}",
"starred_url": "https://api.github.com/users/yawzhe/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yawzhe/subscriptions",
"organizations_url": "https://api.github.com/users/yawzhe/orgs",
"repos_url": "https://api.github.com/users/yawzhe/repos",
"events_url": "https://api.github.com/users/yawzhe/events{/privacy}",
"received_events_url": "https://api.github.com/users/yawzhe/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2024-06-18T02:16:01 | 2024-06-18T11:28:38 | 2024-06-18T11:28:37 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
我想问一下ollama怎么自定义设置KEY_传参, 每个模型定义不同的key,2.ollama是否支持模型授权,加密模型之类的,
### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
最新的 | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5107/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5107/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4621 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4621/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4621/comments | https://api.github.com/repos/ollama/ollama/issues/4621/events | https://github.com/ollama/ollama/issues/4621 | 2,316,267,571 | I_kwDOJ0Z1Ps6KD3Az | 4,621 | phi3-medium-128k wrong number of tensors | {
"login": "EthanGraber",
"id": 18070053,
"node_id": "MDQ6VXNlcjE4MDcwMDUz",
"avatar_url": "https://avatars.githubusercontent.com/u/18070053?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/EthanGraber",
"html_url": "https://github.com/EthanGraber",
"followers_url": "https://api.github.com/users/EthanGraber/followers",
"following_url": "https://api.github.com/users/EthanGraber/following{/other_user}",
"gists_url": "https://api.github.com/users/EthanGraber/gists{/gist_id}",
"starred_url": "https://api.github.com/users/EthanGraber/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/EthanGraber/subscriptions",
"organizations_url": "https://api.github.com/users/EthanGraber/orgs",
"repos_url": "https://api.github.com/users/EthanGraber/repos",
"events_url": "https://api.github.com/users/EthanGraber/events{/privacy}",
"received_events_url": "https://api.github.com/users/EthanGraber/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 2 | 2024-05-24T21:11:10 | 2024-05-24T22:13:06 | 2024-05-24T22:13:05 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I'm getting the following error when testing the new 128k versions of phi3-medium:
```sh
$ ollama run phi3:14b-medium-128k-instruct-q4_0
Error: llama runner process has terminated: signal: abort trap error:done_getting_tensors: wrong number of tensors; expected 245, got 243
```
```sh
$ ollama -v
ollama version is 0.1.38
```
### OS
macOS
### GPU
Apple
### CPU
Apple
### Ollama version
0.1.38 | {
"login": "EthanGraber",
"id": 18070053,
"node_id": "MDQ6VXNlcjE4MDcwMDUz",
"avatar_url": "https://avatars.githubusercontent.com/u/18070053?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/EthanGraber",
"html_url": "https://github.com/EthanGraber",
"followers_url": "https://api.github.com/users/EthanGraber/followers",
"following_url": "https://api.github.com/users/EthanGraber/following{/other_user}",
"gists_url": "https://api.github.com/users/EthanGraber/gists{/gist_id}",
"starred_url": "https://api.github.com/users/EthanGraber/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/EthanGraber/subscriptions",
"organizations_url": "https://api.github.com/users/EthanGraber/orgs",
"repos_url": "https://api.github.com/users/EthanGraber/repos",
"events_url": "https://api.github.com/users/EthanGraber/events{/privacy}",
"received_events_url": "https://api.github.com/users/EthanGraber/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4621/reactions",
"total_count": 2,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/ollama/ollama/issues/4621/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/4226 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4226/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4226/comments | https://api.github.com/repos/ollama/ollama/issues/4226/events | https://github.com/ollama/ollama/issues/4226 | 2,282,922,640 | I_kwDOJ0Z1Ps6IEqKQ | 4,226 | run llama3-70B-q8_0 error | {
"login": "leoHostProject",
"id": 87935281,
"node_id": "MDQ6VXNlcjg3OTM1Mjgx",
"avatar_url": "https://avatars.githubusercontent.com/u/87935281?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/leoHostProject",
"html_url": "https://github.com/leoHostProject",
"followers_url": "https://api.github.com/users/leoHostProject/followers",
"following_url": "https://api.github.com/users/leoHostProject/following{/other_user}",
"gists_url": "https://api.github.com/users/leoHostProject/gists{/gist_id}",
"starred_url": "https://api.github.com/users/leoHostProject/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/leoHostProject/subscriptions",
"organizations_url": "https://api.github.com/users/leoHostProject/orgs",
"repos_url": "https://api.github.com/users/leoHostProject/repos",
"events_url": "https://api.github.com/users/leoHostProject/events{/privacy}",
"received_events_url": "https://api.github.com/users/leoHostProject/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg",
"url": "https://api.github.com/repos/ollama/ollama/labels/nvidia",
"name": "nvidia",
"color": "8CDB00",
"default": false,
"description": "Issues relating to Nvidia GPUs and CUDA"
},
{
"id": 6677745918,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgZQ_g",
"url": "https://api.github.com/repos/ollama/ollama/labels/gpu",
"name": "gpu",
"color": "76C49E",
"default": false,
"description": ""
}
] | closed | false | null | [] | null | 1 | 2024-05-07T10:44:08 | 2024-07-25T18:53:03 | 2024-07-25T18:53:02 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
api call error
message:
{"error":{"message":"timed out waiting for llama runner to start:CUDA error:uncorrectable ECC error encountered\
n current device:0,in function ggml cuda_compute_forward at /go/src/github.com/ollama/ollama/11m/1lama.cpp/ggml
-cuda.cu:2300\n err\nGGML_ASSERT:/go/src/github.com/ollama/ollama/11m/1lama.cpp/ggml-cuda.cu:60:!\"CUDA error\"
""type":"api_error","param"null,"code":null}}
### OS
Linux, Docker
### GPU
Nvidia
### CPU
Intel
### Ollama version
_No response_ | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4226/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4226/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5750 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5750/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5750/comments | https://api.github.com/repos/ollama/ollama/issues/5750/events | https://github.com/ollama/ollama/pull/5750 | 2,414,156,448 | PR_kwDOJ0Z1Ps51q_iz | 5,750 | stub response | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2024-07-17T17:28:46 | 2024-07-17T17:39:25 | 2024-07-17T17:39:22 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5750",
"html_url": "https://github.com/ollama/ollama/pull/5750",
"diff_url": "https://github.com/ollama/ollama/pull/5750.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5750.patch",
"merged_at": "2024-07-17T17:39:22"
} | for compatibility, `{{ .Response }}` cannot be in any template control flow structures. therefore any template execution should set an empty Response if one should not be rendered otherwise the output will contain `<no value>` in place of `{{ .Response }}` | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5750/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5750/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4183 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4183/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4183/comments | https://api.github.com/repos/ollama/ollama/issues/4183/events | https://github.com/ollama/ollama/issues/4183 | 2,279,709,927 | I_kwDOJ0Z1Ps6H4Zzn | 4,183 | pull orca2:7b-fp16 Error: EOF | {
"login": "MarkWard0110",
"id": 90335263,
"node_id": "MDQ6VXNlcjkwMzM1MjYz",
"avatar_url": "https://avatars.githubusercontent.com/u/90335263?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/MarkWard0110",
"html_url": "https://github.com/MarkWard0110",
"followers_url": "https://api.github.com/users/MarkWard0110/followers",
"following_url": "https://api.github.com/users/MarkWard0110/following{/other_user}",
"gists_url": "https://api.github.com/users/MarkWard0110/gists{/gist_id}",
"starred_url": "https://api.github.com/users/MarkWard0110/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/MarkWard0110/subscriptions",
"organizations_url": "https://api.github.com/users/MarkWard0110/orgs",
"repos_url": "https://api.github.com/users/MarkWard0110/repos",
"events_url": "https://api.github.com/users/MarkWard0110/events{/privacy}",
"received_events_url": "https://api.github.com/users/MarkWard0110/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 1 | 2024-05-05T20:14:27 | 2024-05-05T20:17:17 | 2024-05-05T20:17:17 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
`ollama pull orca2:7b-fp16` results in `Error: EOF`
### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.1.33 | {
"login": "MarkWard0110",
"id": 90335263,
"node_id": "MDQ6VXNlcjkwMzM1MjYz",
"avatar_url": "https://avatars.githubusercontent.com/u/90335263?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/MarkWard0110",
"html_url": "https://github.com/MarkWard0110",
"followers_url": "https://api.github.com/users/MarkWard0110/followers",
"following_url": "https://api.github.com/users/MarkWard0110/following{/other_user}",
"gists_url": "https://api.github.com/users/MarkWard0110/gists{/gist_id}",
"starred_url": "https://api.github.com/users/MarkWard0110/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/MarkWard0110/subscriptions",
"organizations_url": "https://api.github.com/users/MarkWard0110/orgs",
"repos_url": "https://api.github.com/users/MarkWard0110/repos",
"events_url": "https://api.github.com/users/MarkWard0110/events{/privacy}",
"received_events_url": "https://api.github.com/users/MarkWard0110/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4183/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4183/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3397 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3397/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3397/comments | https://api.github.com/repos/ollama/ollama/issues/3397/events | https://github.com/ollama/ollama/pull/3397 | 2,214,232,629 | PR_kwDOJ0Z1Ps5rG7Q3 | 3,397 | Parallel requests | {
"login": "0x77dev",
"id": 46429701,
"node_id": "MDQ6VXNlcjQ2NDI5NzAx",
"avatar_url": "https://avatars.githubusercontent.com/u/46429701?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/0x77dev",
"html_url": "https://github.com/0x77dev",
"followers_url": "https://api.github.com/users/0x77dev/followers",
"following_url": "https://api.github.com/users/0x77dev/following{/other_user}",
"gists_url": "https://api.github.com/users/0x77dev/gists{/gist_id}",
"starred_url": "https://api.github.com/users/0x77dev/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/0x77dev/subscriptions",
"organizations_url": "https://api.github.com/users/0x77dev/orgs",
"repos_url": "https://api.github.com/users/0x77dev/repos",
"events_url": "https://api.github.com/users/0x77dev/events{/privacy}",
"received_events_url": "https://api.github.com/users/0x77dev/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 2 | 2024-03-28T21:54:46 | 2024-03-30T22:41:18 | 2024-03-30T22:41:18 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | true | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3397",
"html_url": "https://github.com/ollama/ollama/pull/3397",
"diff_url": "https://github.com/ollama/ollama/pull/3397.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3397.patch",
"merged_at": null
} | Stage: PoC
Related issue: #358
- loaded.mu.{Lock,Unlock}() is not implemented correctly in this change
- sparams.n_parallel is hardcoded to 4
| {
"login": "0x77dev",
"id": 46429701,
"node_id": "MDQ6VXNlcjQ2NDI5NzAx",
"avatar_url": "https://avatars.githubusercontent.com/u/46429701?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/0x77dev",
"html_url": "https://github.com/0x77dev",
"followers_url": "https://api.github.com/users/0x77dev/followers",
"following_url": "https://api.github.com/users/0x77dev/following{/other_user}",
"gists_url": "https://api.github.com/users/0x77dev/gists{/gist_id}",
"starred_url": "https://api.github.com/users/0x77dev/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/0x77dev/subscriptions",
"organizations_url": "https://api.github.com/users/0x77dev/orgs",
"repos_url": "https://api.github.com/users/0x77dev/repos",
"events_url": "https://api.github.com/users/0x77dev/events{/privacy}",
"received_events_url": "https://api.github.com/users/0x77dev/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3397/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3397/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1158 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1158/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1158/comments | https://api.github.com/repos/ollama/ollama/issues/1158/events | https://github.com/ollama/ollama/issues/1158 | 1,997,987,411 | I_kwDOJ0Z1Ps53Ft5T | 1,158 | max retries exceeded: unexpected EOF | {
"login": "priamai",
"id": 57333254,
"node_id": "MDQ6VXNlcjU3MzMzMjU0",
"avatar_url": "https://avatars.githubusercontent.com/u/57333254?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/priamai",
"html_url": "https://github.com/priamai",
"followers_url": "https://api.github.com/users/priamai/followers",
"following_url": "https://api.github.com/users/priamai/following{/other_user}",
"gists_url": "https://api.github.com/users/priamai/gists{/gist_id}",
"starred_url": "https://api.github.com/users/priamai/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/priamai/subscriptions",
"organizations_url": "https://api.github.com/users/priamai/orgs",
"repos_url": "https://api.github.com/users/priamai/repos",
"events_url": "https://api.github.com/users/priamai/events{/privacy}",
"received_events_url": "https://api.github.com/users/priamai/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 15 | 2023-11-16T23:47:19 | 2025-01-28T16:11:44 | 2024-03-11T18:25:00 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | Hi there,
I am not sure if this is related to your file service, but I am getting this connection drops out very often.

Maybe there is a way to throttle requests? | {
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1158/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1158/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/216 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/216/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/216/comments | https://api.github.com/repos/ollama/ollama/issues/216/events | https://github.com/ollama/ollama/issues/216 | 1,822,180,868 | I_kwDOJ0Z1Ps5snEYE | 216 | Something might still be wrong with K-Quant | {
"login": "nkoehring",
"id": 246402,
"node_id": "MDQ6VXNlcjI0NjQwMg==",
"avatar_url": "https://avatars.githubusercontent.com/u/246402?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nkoehring",
"html_url": "https://github.com/nkoehring",
"followers_url": "https://api.github.com/users/nkoehring/followers",
"following_url": "https://api.github.com/users/nkoehring/following{/other_user}",
"gists_url": "https://api.github.com/users/nkoehring/gists{/gist_id}",
"starred_url": "https://api.github.com/users/nkoehring/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/nkoehring/subscriptions",
"organizations_url": "https://api.github.com/users/nkoehring/orgs",
"repos_url": "https://api.github.com/users/nkoehring/repos",
"events_url": "https://api.github.com/users/nkoehring/events{/privacy}",
"received_events_url": "https://api.github.com/users/nkoehring/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 3 | 2023-07-26T11:17:46 | 2023-08-02T19:03:27 | 2023-08-02T19:03:27 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | When I run a 30B model (in this case upstage-llama-30b-instruct-2048.ggmlv3.q5_K_M.bin) the debug output in ollama talks about a 13B model size:

when running the same model with llama.cpp it outputs the correct size:

I tested with a 13B model and the output seems correct. Both models seem to work (as in generating output). | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/216/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/216/timeline | null | not_planned | false |
https://api.github.com/repos/ollama/ollama/issues/4880 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4880/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4880/comments | https://api.github.com/repos/ollama/ollama/issues/4880/events | https://github.com/ollama/ollama/issues/4880 | 2,339,156,469 | I_kwDOJ0Z1Ps6LbLH1 | 4,880 | Extend ollama show command | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] | closed | false | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 0 | 2024-06-06T21:07:21 | 2024-06-26T17:31:00 | 2024-06-26T17:31:00 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | In reference to #3570 | {
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4880/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4880/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/7355 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/7355/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/7355/comments | https://api.github.com/repos/ollama/ollama/issues/7355/events | https://github.com/ollama/ollama/issues/7355 | 2,613,814,322 | I_kwDOJ0Z1Ps6by6Qy | 7,355 | Released binaries have High severity CVEs due to Go version 1.22.5 | {
"login": "pivotal-marcela-campo",
"id": 20945140,
"node_id": "MDQ6VXNlcjIwOTQ1MTQw",
"avatar_url": "https://avatars.githubusercontent.com/u/20945140?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pivotal-marcela-campo",
"html_url": "https://github.com/pivotal-marcela-campo",
"followers_url": "https://api.github.com/users/pivotal-marcela-campo/followers",
"following_url": "https://api.github.com/users/pivotal-marcela-campo/following{/other_user}",
"gists_url": "https://api.github.com/users/pivotal-marcela-campo/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pivotal-marcela-campo/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pivotal-marcela-campo/subscriptions",
"organizations_url": "https://api.github.com/users/pivotal-marcela-campo/orgs",
"repos_url": "https://api.github.com/users/pivotal-marcela-campo/repos",
"events_url": "https://api.github.com/users/pivotal-marcela-campo/events{/privacy}",
"received_events_url": "https://api.github.com/users/pivotal-marcela-campo/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 1 | 2024-10-25T11:17:12 | 2024-10-27T00:03:38 | 2024-10-27T00:03:38 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
Scanning linux binary with `grype` yields the following report

Upgrading to 1.22.7+ for building would fix this issue:
https://github.com/ollama/ollama/blob/3085c47bea508e638c0b9f15a2f4c00afd83b66d/Dockerfile#L1
https://github.com/ollama/ollama/blob/3085c47bea508e638c0b9f15a2f4c00afd83b66d/llama/Dockerfile#L2
https://github.com/ollama/ollama/blob/3085c47bea508e638c0b9f15a2f4c00afd83b66d/go.mod#L3
### OS
Linux, macOS, Docker
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.3.14 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/7355/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/7355/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/5004 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5004/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5004/comments | https://api.github.com/repos/ollama/ollama/issues/5004/events | https://github.com/ollama/ollama/pull/5004 | 2,349,490,461 | PR_kwDOJ0Z1Ps5yRrGr | 5,004 | fix: multiple templates when creating from model | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2024-06-12T19:00:13 | 2024-06-12T21:39:29 | 2024-06-12T21:39:29 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5004",
"html_url": "https://github.com/ollama/ollama/pull/5004",
"diff_url": "https://github.com/ollama/ollama/pull/5004.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5004.patch",
"merged_at": "2024-06-12T21:39:29"
} | multiple templates may appear in a model if a model is created from another model that 1) has an autodetected template and 2) defines a custom template
this fixes the bug by not detecting chat template when inheriting from another model | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5004/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5004/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6676 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6676/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6676/comments | https://api.github.com/repos/ollama/ollama/issues/6676/events | https://github.com/ollama/ollama/issues/6676 | 2,510,163,293 | I_kwDOJ0Z1Ps6Vng1d | 6,676 | on ollama.com , the centrate new profile picture page , looked on andro chrome canary , out of bound | {
"login": "fxmbsw7",
"id": 39368685,
"node_id": "MDQ6VXNlcjM5MzY4Njg1",
"avatar_url": "https://avatars.githubusercontent.com/u/39368685?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/fxmbsw7",
"html_url": "https://github.com/fxmbsw7",
"followers_url": "https://api.github.com/users/fxmbsw7/followers",
"following_url": "https://api.github.com/users/fxmbsw7/following{/other_user}",
"gists_url": "https://api.github.com/users/fxmbsw7/gists{/gist_id}",
"starred_url": "https://api.github.com/users/fxmbsw7/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/fxmbsw7/subscriptions",
"organizations_url": "https://api.github.com/users/fxmbsw7/orgs",
"repos_url": "https://api.github.com/users/fxmbsw7/repos",
"events_url": "https://api.github.com/users/fxmbsw7/events{/privacy}",
"received_events_url": "https://api.github.com/users/fxmbsw7/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6573197867,
"node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw",
"url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com",
"name": "ollama.com",
"color": "ffffff",
"default": false,
"description": ""
}
] | open | false | {
"login": "hoyyeva",
"id": 63033505,
"node_id": "MDQ6VXNlcjYzMDMzNTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoyyeva",
"html_url": "https://github.com/hoyyeva",
"followers_url": "https://api.github.com/users/hoyyeva/followers",
"following_url": "https://api.github.com/users/hoyyeva/following{/other_user}",
"gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}",
"starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions",
"organizations_url": "https://api.github.com/users/hoyyeva/orgs",
"repos_url": "https://api.github.com/users/hoyyeva/repos",
"events_url": "https://api.github.com/users/hoyyeva/events{/privacy}",
"received_events_url": "https://api.github.com/users/hoyyeva/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"login": "hoyyeva",
"id": 63033505,
"node_id": "MDQ6VXNlcjYzMDMzNTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoyyeva",
"html_url": "https://github.com/hoyyeva",
"followers_url": "https://api.github.com/users/hoyyeva/followers",
"following_url": "https://api.github.com/users/hoyyeva/following{/other_user}",
"gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}",
"starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions",
"organizations_url": "https://api.github.com/users/hoyyeva/orgs",
"repos_url": "https://api.github.com/users/hoyyeva/repos",
"events_url": "https://api.github.com/users/hoyyeva/events{/privacy}",
"received_events_url": "https://api.github.com/users/hoyyeva/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 0 | 2024-09-06T10:55:18 | 2024-09-10T21:07:41 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
when upload profile pic , a page to move the pic into o round place .. the canvas doesnt fit into the display , of andro chrome canary

### OS
_No response_
### GPU
_No response_
### CPU
_No response_
### Ollama version
_No response_ | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6676/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6676/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/4902 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4902/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4902/comments | https://api.github.com/repos/ollama/ollama/issues/4902/events | https://github.com/ollama/ollama/issues/4902 | 2,339,983,408 | I_kwDOJ0Z1Ps6LeVAw | 4,902 | Performance issue with CPU only inference start 0.1.39 - to latest version of todate. | {
"login": "raymond-infinitecode",
"id": 4714784,
"node_id": "MDQ6VXNlcjQ3MTQ3ODQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/4714784?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/raymond-infinitecode",
"html_url": "https://github.com/raymond-infinitecode",
"followers_url": "https://api.github.com/users/raymond-infinitecode/followers",
"following_url": "https://api.github.com/users/raymond-infinitecode/following{/other_user}",
"gists_url": "https://api.github.com/users/raymond-infinitecode/gists{/gist_id}",
"starred_url": "https://api.github.com/users/raymond-infinitecode/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/raymond-infinitecode/subscriptions",
"organizations_url": "https://api.github.com/users/raymond-infinitecode/orgs",
"repos_url": "https://api.github.com/users/raymond-infinitecode/repos",
"events_url": "https://api.github.com/users/raymond-infinitecode/events{/privacy}",
"received_events_url": "https://api.github.com/users/raymond-infinitecode/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5808482718,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWjZpng",
"url": "https://api.github.com/repos/ollama/ollama/labels/performance",
"name": "performance",
"color": "A5B5C6",
"default": false,
"description": ""
}
] | closed | false | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 7 | 2024-06-07T09:17:48 | 2024-07-03T23:34:02 | 2024-07-03T23:34:02 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
I am running the ollama on intel xeon 32 processors (CPU only) previously which high token generation count using version 0.1.38
However, once I migrate to the latest ollama version 0.1.41, I found that the inference speed for even a model like phi3 on pure CPU slow to a halt.
I retest the version and reproducing the slowness start with 0.1.39.
Unable to provide log details as there is no error just pure slowness.
Didn't change any model nor configuration.
Revert back to ver 0.1.38, the performance turn high speed again.
Using centos 8 linux
Xeon gold processor 32 core
### OS
Linux
### GPU
Other
### CPU
Intel
### Ollama version
0.1.39 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4902/reactions",
"total_count": 3,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4902/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6115 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6115/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6115/comments | https://api.github.com/repos/ollama/ollama/issues/6115/events | https://github.com/ollama/ollama/pull/6115 | 2,441,859,247 | PR_kwDOJ0Z1Ps53Fqs2 | 6,115 | Fix context in /api/generate grows too much (#5980). | {
"login": "slouffka",
"id": 8129,
"node_id": "MDQ6VXNlcjgxMjk=",
"avatar_url": "https://avatars.githubusercontent.com/u/8129?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/slouffka",
"html_url": "https://github.com/slouffka",
"followers_url": "https://api.github.com/users/slouffka/followers",
"following_url": "https://api.github.com/users/slouffka/following{/other_user}",
"gists_url": "https://api.github.com/users/slouffka/gists{/gist_id}",
"starred_url": "https://api.github.com/users/slouffka/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/slouffka/subscriptions",
"organizations_url": "https://api.github.com/users/slouffka/orgs",
"repos_url": "https://api.github.com/users/slouffka/repos",
"events_url": "https://api.github.com/users/slouffka/events{/privacy}",
"received_events_url": "https://api.github.com/users/slouffka/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 6 | 2024-08-01T08:47:37 | 2024-08-01T22:14:00 | 2024-08-01T22:13:59 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/6115",
"html_url": "https://github.com/ollama/ollama/pull/6115",
"diff_url": "https://github.com/ollama/ollama/pull/6115.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6115.patch",
"merged_at": "2024-08-01T22:13:59"
} | This PR fixes [Context in /api/generate response grows too big. #5980
](https://github.com/ollama/ollama/issues/5980) | {
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6115/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6115/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/6094 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6094/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6094/comments | https://api.github.com/repos/ollama/ollama/issues/6094/events | https://github.com/ollama/ollama/issues/6094 | 2,439,598,593 | I_kwDOJ0Z1Ps6RaVIB | 6,094 | "embedding generation failed: do embedding request: Post \"http://127.0.0.1:33967/embedding\": EOF" | {
"login": "yeexiangzhen1001",
"id": 70881071,
"node_id": "MDQ6VXNlcjcwODgxMDcx",
"avatar_url": "https://avatars.githubusercontent.com/u/70881071?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yeexiangzhen1001",
"html_url": "https://github.com/yeexiangzhen1001",
"followers_url": "https://api.github.com/users/yeexiangzhen1001/followers",
"following_url": "https://api.github.com/users/yeexiangzhen1001/following{/other_user}",
"gists_url": "https://api.github.com/users/yeexiangzhen1001/gists{/gist_id}",
"starred_url": "https://api.github.com/users/yeexiangzhen1001/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yeexiangzhen1001/subscriptions",
"organizations_url": "https://api.github.com/users/yeexiangzhen1001/orgs",
"repos_url": "https://api.github.com/users/yeexiangzhen1001/repos",
"events_url": "https://api.github.com/users/yeexiangzhen1001/events{/privacy}",
"received_events_url": "https://api.github.com/users/yeexiangzhen1001/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 19 | 2024-07-31T09:39:08 | 2025-01-10T08:14:20 | 2024-09-02T23:36:51 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | ### What is the issue?
2024/07/31 09:18:15 routes.go:1099: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:2562047h47m16.854775807s OLLAMA_LLM_LIBRARY: OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://*] OLLAMA_RUNNERS_DIR: OLLAMA_SCHED_SPREAD:false OLLAMA_TMPDIR: ROCR_VISIBLE_DEVICES:]"
time=2024-07-31T09:18:16.095Z level=INFO source=images.go:786 msg="total blobs: 2"
time=2024-07-31T09:18:16.095Z level=INFO source=images.go:793 msg="total unused blobs removed: 0"
time=2024-07-31T09:18:16.095Z level=INFO source=routes.go:1146 msg="Listening on [::]:11434 (version 0.3.1)"
time=2024-07-31T09:18:16.095Z level=INFO source=payload.go:30 msg="extracting embedded files" dir=/tmp/ollama37639419/runners
time=2024-07-31T09:18:18.739Z level=INFO source=payload.go:44 msg="Dynamic LLM libraries [rocm_v60102 cpu cpu_avx cpu_avx2 cuda_v11]"
time=2024-07-31T09:18:18.739Z level=INFO source=gpu.go:205 msg="looking for compatible GPUs"
time=2024-07-31T09:18:18.808Z level=INFO source=types.go:105 msg="inference compute" id=GPU-31fa3c8c-f42e-bade-72ec-f936eb48ac45 library=cuda compute=8.6 driver=12.2 name="NVIDIA GeForce RTX 3090 Ti" total="23.7 GiB" available="17.2 GiB"
time=2024-07-31T09:20:14.214Z level=INFO source=sched.go:701 msg="new model will fit in available VRAM in single GPU, loading" model=/root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 gpu=GPU-31fa3c8c-f42e-bade-72ec-f936eb48ac45 parallel=4 available=18469158912 required="737.9 MiB"
time=2024-07-31T09:20:14.214Z level=INFO source=memory.go:309 msg="offload to cuda" layers.requested=-1 layers.model=13 layers.offload=13 layers.split="" memory.available="[17.2 GiB]" memory.required.full="737.9 MiB" memory.required.partial="737.9 MiB" memory.required.kv="24.0 MiB" memory.required.allocations="[737.9 MiB]" memory.weights.total="186.5 MiB" memory.weights.repeating="155.5 MiB" memory.weights.nonrepeating="30.9 MiB" memory.graph.full="48.0 MiB" memory.graph.partial="48.0 MiB"
time=2024-07-31T09:20:14.214Z level=INFO source=server.go:384 msg="starting llama server" cmd="/tmp/ollama37639419/runners/cuda_v11/ollama_llama_server --model /root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 --ctx-size 8192 --batch-size 512 --embedding --log-disable --n-gpu-layers 13 --parallel 4 --port 44985"
time=2024-07-31T09:20:14.214Z level=INFO source=sched.go:437 msg="loaded runners" count=1
time=2024-07-31T09:20:14.214Z level=INFO source=server.go:584 msg="waiting for llama runner to start responding"
time=2024-07-31T09:20:14.214Z level=INFO source=server.go:618 msg="waiting for server to become available" status="llm server error"
INFO [main] build info | build=1 commit="6eeaeba" tid="127422522179584" timestamp=1722417614
INFO [main] system info | n_threads=8 n_threads_batch=-1 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 0 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="127422522179584" timestamp=1722417614 total_threads=16
INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="15" port="44985" tid="127422522179584" timestamp=1722417614
llama_model_loader: loaded meta data with 22 key-value pairs and 197 tensors from /root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 (version GGUF V3 (latest))
llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.
llama_model_loader: - kv 0: general.architecture str = bert
llama_model_loader: - kv 1: general.name str = Dmeta-embedding-zh
llama_model_loader: - kv 2: bert.block_count u32 = 12
llama_model_loader: - kv 3: bert.context_length u32 = 1024
llama_model_loader: - kv 4: bert.embedding_length u32 = 768
llama_model_loader: - kv 5: bert.feed_forward_length u32 = 3072
llama_model_loader: - kv 6: bert.attention.head_count u32 = 12
llama_model_loader: - kv 7: bert.attention.layer_norm_epsilon f32 = 0.000000
llama_model_loader: - kv 8: general.file_type u32 = 1
llama_model_loader: - kv 9: bert.attention.causal bool = false
llama_model_loader: - kv 10: bert.pooling_type u32 = 2
llama_model_loader: - kv 11: tokenizer.ggml.token_type_count u32 = 2
llama_model_loader: - kv 12: tokenizer.ggml.model str = bert
llama_model_loader: - kv 13: tokenizer.ggml.tokens arr[str,21128] = ["[PAD]", "[unused1]", "[unused2]", "...
llama_model_loader: - kv 14: tokenizer.ggml.token_type arr[i32,21128] = [3, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ...
llama_model_loader: - kv 15: tokenizer.ggml.unknown_token_id u32 = 100
llama_model_loader: - kv 16: tokenizer.ggml.seperator_token_id u32 = 102
llama_model_loader: - kv 17: tokenizer.ggml.padding_token_id u32 = 0
llama_model_loader: - kv 18: tokenizer.ggml.cls_token_id u32 = 101
llama_model_loader: - kv 19: tokenizer.ggml.mask_token_id u32 = 103
llama_model_loader: - kv 20: tokenizer.ggml.bos_token_id u32 = 0
llama_model_loader: - kv 21:
tokenizer.ggml.eos_token_id u32 = 2
llama_model_loader: - type f32: 123 tensors
llama_model_loader: - type f16: 74 tensors
llm_load_vocab: special tokens cache size = 5
llm_load_vocab: token to piece cache size = 0.0769 MB
llm_load_print_meta: format = GGUF V3 (latest)
llm_load_print_meta: arch = bert
llm_load_print_meta: vocab type = WPM
llm_load_print_meta: n_vocab = 21128
llm_load_print_meta: n_merges = 0
llm_load_print_meta: vocab_only = 0
llm_load_print_meta: n_ctx_train = 1024
llm_load_print_meta: n_embd = 768
llm_load_print_meta: n_layer = 12
llm_load_print_meta: n_head = 12
llm_load_print_meta: n_head_kv = 12
llm_load_print_meta: n_rot = 64
llm_load_print_meta: n_swa = 0
llm_load_print_meta: n_embd_head_k = 64
llm_load_print_meta: n_embd_head_v = 64
llm_load_print_meta: n_gqa = 1
llm_load_print_meta: n_embd_k_gqa = 768
llm_load_print_meta: n_embd_v_gqa = 768
llm_load_print_meta: f_norm_eps = 1.0e-12
llm_load_print_meta: f_norm_rms_eps = 0.0e+00
llm_load_print_meta: f_clamp_kqv = 0.0e+00
llm_load_print_meta: f_max_alibi_bias = 0.0e+00
llm_load_print_meta: f_logit_scale = 0.0e+00
llm_load_print_meta: n_ff = 3072
llm_load_print_meta: n_expert = 0
llm_load_print_meta: n_expert_used = 0
llm_load_print_meta: causal attn = 0
llm_load_print_meta: pooling type = 2
llm_load_print_meta: rope type = 2
llm_load_print_meta: rope scaling = linear
llm_load_print_meta: freq_base_train = 10000.0
llm_load_print_meta: freq_scale_train = 1
llm_load_print_meta: n_ctx_orig_yarn = 1024
llm_load_print_meta: rope_finetuned = unknown
llm_load_print_meta: ssm_d_conv = 0
llm_load_print_meta: ssm_d_inner = 0
llm_load_print_meta: ssm_d_state = 0
llm_load_print_meta: ssm_dt_rank = 0
llm_load_print_meta: model type = 109M
llm_load_print_meta: model ftype = F16
llm_load_print_meta: model params = 102.07 M
llm_load_print_meta: model size = 194.92 MiB (16.02 BPW)
llm_load_print_meta: general.name = Dmeta-embedding-zh
llm_load_print_meta: BOS token = 0 '[PAD]'
llm_load_print_meta: EOS token = 2 '[unused2]'
llm_load_print_meta: UNK token = 100 '[UNK]'
llm_load_print_meta: SEP token = 102 '[SEP]'
llm_load_print_meta: PAD token = 0 '[PAD]'
llm_load_print_meta: CLS token = 101 '[CLS]'
llm_load_print_meta: MASK token = 103 '[MASK]'
llm_load_print_meta: LF token = 0 '[PAD]'
llm_load_print_meta: max token length = 48
ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 CUDA devices:
Device 0: NVIDIA GeForce RTX 3090 Ti, compute capability 8.6, VMM: yes
llm_load_tensors: ggml ctx size = 0.16 MiB
llm_load_tensors: offloading 12 repeating layers to GPU
llm_load_tensors: offloading non-repeating layers to GPU
llm_load_tensors: offloaded 13/13 layers to GPU
llm_load_tensors: CPU buffer size = 32.46 MiB
llm_load_tensors: CUDA0 buffer size = 162.46 MiB
llama_new_context_with_model: n_ctx = 8192
llama_new_context_with_model: n_batch = 512
llama_new_context_with_model: n_ubatch = 512
llama_new_context_with_model: flash_attn = 0
llama_new_context_with_model: freq_base = 10000.0
llama_new_context_with_model: freq_scale = 1
llama_kv_cache_init: CUDA0 KV buffer size = 288.00 MiB
llama_new_context_with_model: KV self size = 288.00 MiB, K (f16): 144.00 MiB, V (f16): 144.00 MiB
llama_new_context_with_model: CPU output buffer size = 0.00 MiB
llama_new_context_with_model: CUDA0 compute buffer size = 19.00 MiB
llama_new_context_with_model: CUDA_Host compute buffer size = 4.00 MiB
llama_new_context_with_model: graph nodes = 429
llama_new_context_with_model: graph splits = 2
time=2024-07-31T09:20:14.465Z level=INFO source=server.go:618 msg="waiting for server to become available" status="llm server loading model"
INFO [main] model loaded | tid="127422522179584" timestamp=1722417614
time=2024-07-31T09:20:14.966Z level=INFO source=server.go:623 msg="llama runner started in 0.75 seconds"
[GIN] 2024/07/31 - 09:20:15 | 200 | 862.184786ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:20:15 | 200 | 91.260258ms | 10.234.218.0 | POST "/api/embeddings"
time=2024-07-31T09:20:15.383Z level=INFO source=routes.go:426 msg="embedding generation failed: do embedding request: Post \"http://127.0.0.1:44985/embedding\": EOF"
[GIN] 2024/07/31 - 09:20:15 | 500 | 140.114654ms | 10.234.218.0 | POST "/api/embeddings"
time=2024-07-31T09:23:45.923Z level=WARN source=server.go:503 msg="llama runner process no longer running" sys=139 string="signal: segmentation fault (core dumped)"
time=2024-07-31T09:23:50.993Z level=WARN source=sched.go:634 msg="gpu VRAM usage didn't recover within timeout" seconds=5.069197565 model=/root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527
time=2024-07-31T09:23:51.075Z level=INFO source=sched.go:701 msg="new model will fit in available VRAM in single GPU, loading" model=/root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 gpu=GPU-31fa3c8c-f42e-bade-72ec-f936eb48ac45 parallel=4 available=18469158912 required="737.9 MiB"
time=2024-07-31T09:23:51.075Z level=INFO source=memory.go:309 msg="offload to cuda" layers.requested=-1 layers.model=13 layers.offload=13 layers.split="" memory.available="[17.2 GiB]" memory.required.full="737.9 MiB" memory.required.partial="737.9 MiB" memory.required.kv="24.0 MiB" memory.required.allocations="[737.9 MiB]" memory.weights.total="186.5 MiB" memory.weights.repeating="155.5 MiB" memory.weights.nonrepeating="30.9 MiB" memory.graph.full="48.0 MiB" memory.graph.partial="48.0 MiB"
time=2024-07-31T09:23:51.075Z level=INFO source=server.go:384 msg="starting llama server" cmd="/tmp/ollama37639419/runners/cuda_v11/ollama_llama_server --model /root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 --ctx-size 8192 --batch-size 512 --embedding --log-disable --n-gpu-layers 13 --parallel 4 --port 42155"
time=2024-07-31T09:23:51.075Z level=INFO source=sched.go:437 msg="loaded runners" count=1
time=2024-07-31T09:23:51.075Z level=INFO source=server.go:584 msg="waiting for llama runner to start responding"
time=2024-07-31T09:23:51.076Z level=INFO source=server.go:618 msg="waiting for server to become available" status="llm server error"
INFO [main] build info | build=1 commit="6eeaeba" tid="131709034942464" timestamp=1722417831
INFO [main] system info | n_threads=8 n_threads_batch=-1 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 0 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="131709034942464" timestamp=1722417831 total_threads=16
INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="15" port="42155" tid="131709034942464" timestamp=1722417831
llama_model_loader: loaded meta data with 22 key-value pairs and 197 tensors from /root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 (version GGUF V3 (latest))
llama_model_loader: D
umping metadata keys/values. Note: KV overrides do not apply in this output.
llama_model_loader: - kv 0: general.architecture str = bert
llama_model_loader: - kv 1: general.name str = Dmeta-embedding-zh
llama_model_loader: - kv 2: bert.block_count u32 = 12
llama_model_loader: - kv 3: bert.context_length u32 = 1024
llama_model_loader: - kv 4: bert.embedding_length u32 = 768
llama_model_loader: - kv 5: bert.feed_forward_length u32 = 3072
llama_model_loader: - kv 6: bert.attention.head_count u32 = 12
llama_model_loader: - kv 7: bert.attention.layer_norm_epsilon f32 = 0.000000
llama_model_loader: - kv 8: general.file_type u32 = 1
llama_model_loader: - kv 9: bert.attention.causal bool = false
llama_model_loader: - kv 10: bert.pooling_type u32 = 2
llama_model_loader: - kv 11: tokenizer.ggml.token_type_count u32 = 2
llama_model_loader: - kv 12: tokenizer.ggml.model str = bert
llama_model_loader: - kv 13: tokenizer.ggml.tokens arr[str,21128] = ["[PAD]", "[unused1]", "[unused2]", "...
llama_model_loader: - kv 14: tokenizer.ggml.token_type arr[i32,21128] = [3, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ...
llama_model_loader: - kv 15: tokenizer.ggml.unknown_token_id u32 = 100
llama_model_loader: - kv 16: tokenizer.ggml.seperator_token_id u32 = 102
llama_model_loader: - kv 17: tokenizer.ggml.padding_token_id u32 = 0
llama_model_loader: - kv 18: tokenizer.ggml.cls_token_id u32 = 101
llama_model_loader: - kv 19: tokenizer.ggml.mask_token_id u32 = 103
llama_model_loader: - kv 20: tokenizer.ggml.bos_token_id u32 = 0
llama_model_loader: - kv 21: tokenizer.ggml.eos_token_id u32 = 2
llama_model_loader: - type f32: 123 tensors
llama_model_loader: - type f16: 74 tensors
llm_load_vocab: special tokens cache size = 5
llm_load_vocab: token to piece cache size = 0.0769 MB
llm_load_print_meta: format = GGUF V3 (latest)
llm_load_print_meta: arch = bert
llm_load_print_meta: vocab type = WPM
llm_load_print_meta: n_vocab = 21128
llm_load_print_meta: n_merges = 0
llm_load_print_meta: vocab_only = 0
llm_load_print_meta: n_ctx_train = 1024
llm_load_print_meta: n_embd = 768
llm_load_print_meta: n_layer = 12
llm_load_print_meta: n_head = 12
llm_load_print_meta: n_head_kv = 12
llm_load_print_meta: n_rot = 64
llm_load_print_meta: n_swa = 0
llm_load_print_meta: n_embd_head_k = 64
llm_load_print_meta: n_embd_head_v = 64
llm_load_print_meta: n_gqa = 1
llm_load_print_meta: n_embd_k_gqa = 768
llm_load_print_meta: n_embd_v_gqa = 768
llm_load_print_meta: f_norm_eps = 1.0e-12
llm_load_print_meta: f_norm_rms_eps = 0.0e+00
llm_load_print_meta: f_clamp_kqv = 0.0e+00
llm_load_print_meta: f_max_alibi_bias = 0.0e+00
llm_load_print_meta: f_logit_scale = 0.0e+00
llm_load_print_meta: n_ff = 3072
llm_load_print_meta: n_expert = 0
llm_load_print_meta: n_expert_used = 0
llm_load_print_meta: causal attn = 0
llm_load_print_meta: pooling type = 2
llm_load_print_meta: rope type = 2
llm_load_print_meta: rope scaling = linear
llm_load_print_meta: freq_base_train = 10000.0
llm_load_print_meta: freq_scale_train = 1
llm_load_print_meta: n_ctx_orig_yarn = 1024
llm_load_print_meta: rope_finetuned = unknown
llm_load_print_meta: ssm_d_conv = 0
llm_load_print_meta: ssm_d_inner = 0
llm_load_print_meta: ssm_d_state = 0
llm_load_print_meta: ssm_dt_rank = 0
llm_load_print_meta: model type = 109M
llm_load_print_meta: model ftype = F16
llm_load_print_meta: model params = 102.07 M
llm_load_print_meta: model size = 194.92 MiB (16.02 BPW)
llm_load_print_meta: general.name = Dmeta-embedding-zh
llm_load_print_meta: BOS token = 0 '[PAD]'
llm_load_print_meta: EOS token = 2 '[unused2]'
llm_load_print_meta: UNK token = 100 '[UNK]'
llm_load_print_meta: SEP token = 102 '[SEP]'
llm_load_print_meta: PAD token = 0 '[PAD]'
llm_load_print_meta: CLS token = 101 '[CLS]'
llm_load_print_meta: MASK token = 103 '[MASK]'
llm_load_print_meta: LF token = 0 '[PAD]'
llm_load_print_meta: max token length = 48
ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 CUDA devices:
Device 0: NVIDIA GeForce RTX 3090 Ti, compute capability 8.6, VMM: yes
llm_load_tensors: ggml ctx size = 0.16 MiB
llm_load_tensors: offloading 12 repeating layers to GPU
llm_load_tensors: offloading non-repeating layers to GPU
llm_load_tensors: offloaded 13/13 layers to GPU
llm_load_tensors: CPU buffer size = 32.46 MiB
llm_load_tensors: CUDA0 buffer size = 162.46 MiB
llama_new_context_with_model: n_ctx = 8192
llama_new_context_with_model: n_batch = 512
llama_new_context_with_model: n_ubatch = 512
llama_new_context_with_model: flash_attn = 0
llama_new_context_with_model: freq_base = 10000.0
llama_new_context_with_model: freq_scale = 1
llama_kv_cache_init: CUDA0 KV buffer size = 288.00 MiB
llama_new_context_with_model: KV self size = 288.00 MiB, K (f16): 144.00 MiB, V (f16): 144.00 MiB
llama_new_context_with_model: CPU output buffer size = 0.00 MiB
llama_new_context_with_model: CUDA0 compute buffer size = 19.00 MiB
llama_new_context_with_model: CUDA_Host compute buffer size = 4.00 MiB
llama_new_context_with_model: graph nodes = 429
llama_new_context_with_model: graph splits = 2
time=2024-07-31T09:23:51.243Z level=WARN source=sched.go:634 msg="gpu VRAM usage didn't recover within timeout" seconds=5.319657234 model=/root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527
time=2024-07-31T09:23:51.327Z level=INFO source=server.go:618 msg="waiting for server to become available" status="llm server loading model"
INFO [main] model loaded | tid="131709034942464" timestamp=1722417831
time=2024-07-31T09:23:51.829Z level=INFO source=server.go:623 msg="llama runner started in 0.75 seconds"
[GIN] 2024/07/31 - 09:23:51 | 200 | 5.954027368s | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:51 | 200 | 5.997875851s | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:51 | 200 | 6.001301156s | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:51 | 200 | 6.05401596s | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 6.093406397s | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 6.093515843s | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 141.106871ms | 10.234.218.0 | POST "/api/embeddings"
INFO [update_slots] input truncated | n_ctx=2048 n_erase=1989 n_keep=0 n_left=2048 n_shift=1024 tid="131709034942464" timestamp=1722417832
[GIN] 2024/07/31 - 09:23:52 | 200 | 156.396038ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 159.160468ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 155.371305ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 150.237024ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 161.78585ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 158.374292ms | 10.234.218.0 | POST "/api/embeddings"
INFO [update_slots] input truncated | n_ctx=2048 n_erase=1517 n_keep=0 n_left=2048 n_shift=1024 tid="131709034942464" timestamp=1722417832
[GIN] 2024/07/31 - 09:23:52 | 200 | 144.427285ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 192.549717ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 131.371235ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 185.844931ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 151.950066ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 141.888776ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 171.173954ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 130.251712ms | 10.234.218.0 | POST "/api/embeddings"
INFO [update_slots] input truncated | n_ctx=2048 n_erase=1709 n_keep=0 n_left=2048 n_shift=1024 tid="131709034942464" timestamp=1722417832
[GIN] 2024/07/31 - 09:23:52 | 200 | 140.112505ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 171.12123ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 227.184409ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 264.346952ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 189.302007ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 183.643992ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 165.703255ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 200 | 229.741451ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:52 | 500 | 303.282026ms | 10.234.218.0 | POST "/api/embeddings"
time=2024-07-31T09:23:52.825Z level=INFO source=routes.go:426 msg="embedding generation failed: do embedding request: Post \"http://127.0.0.1:42155/embedding\": EOF"
time=2024-07-31T09:23:57.889Z level=WARN source=sched.go:634 msg="gpu VRAM usage didn't recover within timeout" seconds=5.063724982 model=/root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527
time=2024-07-31T09:23:57.975Z level=INFO source=sched.go:701 msg="new model will fit in available VRAM in single GPU, loading" model=/root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 gpu=GPU-31fa3c8c-f42e-bade-72ec-f936eb48ac45 parallel=4 available=18469158912 required="737.9 MiB"
time=2024-07-31T09:23:57.975Z level=INFO source=memory.go:309 msg="offload to cuda" layers.requested=-1 layers.model=13 layers.offload=13 layers.split="" memory.available="[17.2 GiB]" memory.required.full="737.9 MiB" memory.required.partial="737.9 MiB" memory.required.kv="24.0 MiB" memory.required.allocations="[737.9 MiB]" memory.weights.total="186.5 MiB" memory.weights.repeating="155.5 MiB" memory.weights.nonrepeating="30.9 MiB" memory.graph.full="48.0 MiB" memory.graph.partial="48.0 MiB"
time=2024-07-31T09:23:57.975Z level=INFO source=server.go:384 msg="starting llama server" cmd="/tmp/ollama37639419/runners/cuda_v11/ollama_llama_server --model /root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 --ctx-size 8192 --batch-size 512 --embedding --log-disable --n-gpu-layers 1
3 --parallel 4 --port 33967"
time=2024-07-31T09:23:57.976Z level=INFO source=sched.go:437 msg="loaded runners" count=1
time=2024-07-31T09:23:57.976Z level=INFO source=server.go:584 msg="waiting for llama runner to start responding"
time=2024-07-31T09:23:57.976Z level=INFO source=server.go:618 msg="waiting for server to become available" status="llm server error"
INFO [main] build info | build=1 commit="6eeaeba" tid="125558191894528" timestamp=1722417837
INFO [main] system info | n_threads=8 n_threads_batch=-1 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 0 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="125558191894528" timestamp=1722417837 total_threads=16
INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="15" port="33967" tid="125558191894528" timestamp=1722417837
llama_model_loader: loaded meta data with 22 key-value pairs and 197 tensors from /root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 (version GGUF V3 (latest))
llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.
llama_model_loader: - kv 0: general.architecture str = bert
llama_model_loader: - kv 1: general.name str = Dmeta-embedding-zh
llama_model_loader: - kv 2: bert.block_count u32 = 12
llama_model_loader: - kv 3: bert.context_length u32 = 1024
llama_model_loader: - kv 4: bert.embedding_length u32 = 768
llama_model_loader: - kv 5: bert.feed_forward_length u32 = 3072
llama_model_loader: - kv 6: bert.attention.head_count u32 = 12
llama_model_loader: - kv 7: bert.attention.layer_norm
_epsilon f32 = 0.000000
llama_model_loader: - kv 8: general.file_type u32 = 1
llama_model_loader: - kv 9: bert.attention.causal bool = false
llama_model_loader: - kv 10: bert.pooling_type u32 = 2
llama_model_loader: - kv 11: tokenizer.ggml.token_type_count u32 = 2
llama_model_loader: - kv 12: tokenizer.ggml.model str = bert
llama_model_loader: - kv 13: tokenizer.ggml.tokens arr[str,21128] = ["[PAD]", "[unused1]", "[unused2]", "...
llama_model_loader: - kv 14: tokenizer.ggml.token_type arr[i32,21128] = [3, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ...
llama_model_loader: - kv 15: tokenizer.ggml.unknown_token_id u32 = 100
llama_model_loader: - kv 16: tokenizer.ggml.seperator_token_id u32 = 102
llama_model_loader: - kv 17: tokenizer.ggml.padding_token_id u32 = 0
llama_model_loader: - kv 18: tokenizer.ggml.cls_token_id u32 = 101
llama_model_loader: - kv 19: tokenizer.ggml.mask_token_id u32 = 103
llama_model_loader: - kv 20: tokenizer.ggml.bos_token_id u32 = 0
llama_model_loader: - kv 21: tokenizer.ggml.eos_token_id u32 = 2
llama_model_loader: - type f32: 123 tensors
llama_model_loader: - type f16: 74 tensors
llm_load_vocab: special tokens cache size = 5
llm_load_vocab: token to piece cache size = 0.0769 MB
llm_load_print_meta: format = GGUF V3 (latest)
llm_load_print_meta: arch = bert
llm_load_print_meta: vocab type = WPM
llm_load_print_meta: n_vocab = 21128
llm_load_print_meta: n_merges = 0
llm_load_print_meta: vocab_only = 0
llm_load_print_meta: n_ctx_train = 1024
llm_load_print_meta: n_embd = 768
llm_load_print_meta: n_layer = 12
llm_load_print_meta: n_head = 12
llm_load_print_meta: n_head_kv = 12
llm_load_print_meta: n_rot = 64
llm_load_print_meta: n_swa = 0
llm_load_print_meta: n_embd_head_k = 64
llm_load_print_meta: n_embd_head_v = 64
llm_load_print_meta: n_gqa = 1
llm_load_print_meta: n_embd_k_gqa = 768
llm_load_print_meta: n_embd_v_gqa = 768
llm_load_print_meta: f_norm_eps = 1.0e-12
llm_load_print_meta: f_norm_rms_eps = 0.0e+00
llm_load_print_meta: f_clamp_kqv = 0.0e+00
llm_load_print_meta: f_max_alibi_bias = 0.0e+00
llm_load_print_meta: f_logit_scale = 0.0e+00
llm_load_print_meta: n_ff = 3072
llm_load_print_meta: n_expert = 0
llm_load_print_meta: n_expert_used = 0
llm_load_print_meta: causal attn = 0
llm_load_print_meta: pooling type = 2
llm_load_print_meta: rope type = 2
llm_load_print_meta: rope scaling = linear
llm_load_print_meta: freq_base_train = 10000.0
llm_load_print_meta: freq_scale_train = 1
llm_load_print_meta: n_ctx_orig_yarn = 1024
llm_load_print_meta: rope_finetuned = unknown
llm_load_print_meta: ssm_d_conv = 0
llm_load_print_meta: ssm_d_inner = 0
llm_load_print_meta: ssm_d_state = 0
llm_load_print_meta: ssm_dt_rank = 0
llm_load_print_meta: model type = 109M
llm_load_print_meta: model ftype = F16
llm_load_print_meta: model params = 102.07 M
llm_load_print_meta: model size = 194.92 MiB (16.02 BPW)
llm_load_print_meta: general.name = Dmeta-embedding-zh
llm_load_print_meta: BOS token = 0 '[PAD]'
llm_load_print_meta: EOS token = 2 '[unused2]'
llm_load_print_meta: UNK token = 100 '[UNK]'
llm_load_print_meta: SEP token = 102 '[SEP]'
llm_load_print_meta: PAD token = 0 '[PAD]'
llm_load_print_meta: CLS token = 101 '[CLS]'
llm_load_print_meta: MASK token = 103 '[MASK]'
llm_load_print_meta: LF token = 0 '[PAD]'
llm_load_print_meta: max token length = 48
ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 CUDA devices:
Device 0: NVIDIA GeForce RTX 3090 Ti, compute capability 8.6, VMM: yes
llm_load_tensors: ggml ctx size = 0.16 MiB
llm_load_tensors: offloading 12 repeating layers to GPU
llm_load_tensors: offloading non-repeating layers to GPU
llm_load_tensors: offloaded 13/13 layers to GPU
llm_load_tensors: CPU buffer size = 32.46 MiB
llm_load_tensors: CUDA0 buffer size = 162.46 MiB
llama_new_context_with_model: n_ctx = 8192
llama_new_context_with_model: n_batch = 512
llama_new_context_with_model: n_ubatch = 512
llama_new_context_with_model: flash_attn = 0
llama_new_context_with_model: freq_base = 10000.0
llama_new_context_with_model: freq_scale = 1
llama_kv_cache_init: CUDA0 KV buffer size = 288.00 MiB
llama_new_context_with_model: KV self size = 288.00 MiB, K (f16): 144.00 MiB, V (f16): 144.00 MiB
llama_new_context_with_model: CPU output buffer size = 0.00 MiB
llama_new_context_with_model: CUDA0 compute buffer size = 19.00 MiB
llama_new_context_with_model: CUDA_Host compute buffer size = 4.00 MiB
llama_new_context_with_model: graph nodes = 429
llama_new_context_with_model: graph splits = 2
time=2024-07-31T09:23:58.139Z level=WARN source=sched.go:634 msg="gpu VRAM usage didn't recover within timeout" seconds=5.312995606 model=/root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527
time=2024-07-31T09:23:58.226Z level=INFO source=server.go:618 msg="waiting for server to become available" status="llm server loading model"
INFO [main] model loaded | tid="125558191894528" timestamp=1722417838
time=2024-07-31T09:23:58.729Z level=INFO source=server.go:623 msg="llama runner started in 0.75 seconds"
[GIN] 2024/07/31 - 09:23:58 | 200 | 6.175518609s | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:58 | 200 | 6.173129645s | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/
31 - 09:23:58 | 200 | 6.181901759s | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:58 | 200 | 6.217999442s | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:58 | 200 | 6.128390115s | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:58 | 200 | 139.275881ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:58 | 200 | 141.805964ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:58 | 200 | 147.553231ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:59 | 200 | 147.626781ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:59 | 200 | 90.649859ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:59 | 200 | 134.183906ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:59 | 200 | 100.703301ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:59 | 200 | 76.093064ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:59 | 200 | 139.579148ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:59 | 200 | 195.963998ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:59 | 200 | 184.951077ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:59 | 200 | 204.863879ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:59 | 200 | 93.607337ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:59 | 200 | 92.691741ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:59 | 200 | 122.460956ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:23:59 | 200 | 164.876363ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:26:50 | 200 | 93.430143ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:26:50 | 200 | 51.56662ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:26:
50 | 200 | 139.845262ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:26:50 | 200 | 48.229681ms | 10.234.218.0 | POST "/api/embeddings"
INFO [update_slots] input truncated | n_ctx=2048 n_erase=1522 n_keep=0 n_left=2048 n_shift=1024 tid="125558191894528" timestamp=1722418010
[GIN] 2024/07/31 - 09:26:50 | 200 | 103.527766ms | 10.234.218.0 | POST "/api/embeddings"
[GIN] 2024/07/31 - 09:26:50 | 500 | 138.709641ms | 10.234.218.0 | POST "/api/embeddings"
time=2024-07-31T09:26:50.849Z level=INFO source=routes.go:426 msg="embedding generation failed: do embedding request: Post \"http://127.0.0.1:33967/embedding\": EOF"
[GIN] 2024/07/31 - 09:37:35 | 200 | 19.4µs | 127.0.0.1 | GET "/api/version"
### OS
Docker
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.3.1 | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6094/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 1,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6094/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/6743 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/6743/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/6743/comments | https://api.github.com/repos/ollama/ollama/issues/6743/events | https://github.com/ollama/ollama/pull/6743 | 2,518,666,601 | PR_kwDOJ0Z1Ps57GpZP | 6,743 | Fixed no redirect URL scenario when downloading blobs | {
"login": "JingWoo",
"id": 21989093,
"node_id": "MDQ6VXNlcjIxOTg5MDkz",
"avatar_url": "https://avatars.githubusercontent.com/u/21989093?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/JingWoo",
"html_url": "https://github.com/JingWoo",
"followers_url": "https://api.github.com/users/JingWoo/followers",
"following_url": "https://api.github.com/users/JingWoo/following{/other_user}",
"gists_url": "https://api.github.com/users/JingWoo/gists{/gist_id}",
"starred_url": "https://api.github.com/users/JingWoo/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/JingWoo/subscriptions",
"organizations_url": "https://api.github.com/users/JingWoo/orgs",
"repos_url": "https://api.github.com/users/JingWoo/repos",
"events_url": "https://api.github.com/users/JingWoo/events{/privacy}",
"received_events_url": "https://api.github.com/users/JingWoo/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | open | false | null | [] | null | 1 | 2024-09-11T06:34:01 | 2024-09-30T09:08:37 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/6743",
"html_url": "https://github.com/ollama/ollama/pull/6743",
"diff_url": "https://github.com/ollama/ollama/pull/6743.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6743.patch",
"merged_at": null
} | null | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/6743/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/6743/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/1312 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/1312/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/1312/comments | https://api.github.com/repos/ollama/ollama/issues/1312/events | https://github.com/ollama/ollama/issues/1312 | 2,016,449,283 | I_kwDOJ0Z1Ps54MJMD | 1,312 | trouble with deepseek-coder | {
"login": "niknoproblems",
"id": 3484515,
"node_id": "MDQ6VXNlcjM0ODQ1MTU=",
"avatar_url": "https://avatars.githubusercontent.com/u/3484515?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/niknoproblems",
"html_url": "https://github.com/niknoproblems",
"followers_url": "https://api.github.com/users/niknoproblems/followers",
"following_url": "https://api.github.com/users/niknoproblems/following{/other_user}",
"gists_url": "https://api.github.com/users/niknoproblems/gists{/gist_id}",
"starred_url": "https://api.github.com/users/niknoproblems/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/niknoproblems/subscriptions",
"organizations_url": "https://api.github.com/users/niknoproblems/orgs",
"repos_url": "https://api.github.com/users/niknoproblems/repos",
"events_url": "https://api.github.com/users/niknoproblems/events{/privacy}",
"received_events_url": "https://api.github.com/users/niknoproblems/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 4 | 2023-11-29T12:10:21 | 2024-03-12T01:15:06 | 2024-03-12T01:15:05 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | I'm having trouble getting this model to run on mac m1 16gb ram:
ollama run deepseek-coder:6.7b-base-q8_0
but this model work without any troubles:
ollama run neural-chat:7b-v3.1-q8_0
it has more weights and bigger file size. | {
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/1312/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/1312/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/2529 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/2529/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/2529/comments | https://api.github.com/repos/ollama/ollama/issues/2529/events | https://github.com/ollama/ollama/issues/2529 | 2,137,653,184 | I_kwDOJ0Z1Ps5_af_A | 2,529 | Ollama Windows is much slower at inference than Ollama on WSL2 | {
"login": "devinprater",
"id": 15256014,
"node_id": "MDQ6VXNlcjE1MjU2MDE0",
"avatar_url": "https://avatars.githubusercontent.com/u/15256014?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/devinprater",
"html_url": "https://github.com/devinprater",
"followers_url": "https://api.github.com/users/devinprater/followers",
"following_url": "https://api.github.com/users/devinprater/following{/other_user}",
"gists_url": "https://api.github.com/users/devinprater/gists{/gist_id}",
"starred_url": "https://api.github.com/users/devinprater/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/devinprater/subscriptions",
"organizations_url": "https://api.github.com/users/devinprater/orgs",
"repos_url": "https://api.github.com/users/devinprater/repos",
"events_url": "https://api.github.com/users/devinprater/events{/privacy}",
"received_events_url": "https://api.github.com/users/devinprater/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg",
"url": "https://api.github.com/repos/ollama/ollama/labels/windows",
"name": "windows",
"color": "0052CC",
"default": false,
"description": ""
}
] | closed | false | null | [] | null | 6 | 2024-02-16T00:18:17 | 2024-02-21T09:25:21 | 2024-02-19T21:23:33 | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | CPU: AMD 5500U with Radion internal GPU. Ollama runs on CPU mode on both WSL2 and Windows. Attached are the logs from Windows, and Linux.
[server.log](https://github.com/ollama/ollama/files/14303692/server.log)
[ollama-log-linux.log](https://github.com/ollama/ollama/files/14303696/ollama-log-linux.log) | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/2529/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/2529/timeline | null | completed | false |
https://api.github.com/repos/ollama/ollama/issues/3068 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/3068/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/3068/comments | https://api.github.com/repos/ollama/ollama/issues/3068/events | https://github.com/ollama/ollama/pull/3068 | 2,180,371,586 | PR_kwDOJ0Z1Ps5pT6Cw | 3,068 | Use stdin for term discovery on windows | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2024-03-11T22:28:55 | 2024-03-14T18:55:22 | 2024-03-14T18:55:19 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/3068",
"html_url": "https://github.com/ollama/ollama/pull/3068",
"diff_url": "https://github.com/ollama/ollama/pull/3068.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3068.patch",
"merged_at": "2024-03-14T18:55:19"
} | When you feed input to the cmd via a pipe it no longer reports a warning
Before:
```
> echo "what is the captial of australia" | .\ollama.exe run phi
failed to get console mode for stdin: The handle is invalid.
The capital of Australia is Canberra. It's located in the Australian Capital Territory, about 120 kilometers
northwest of Sydney and 130 kilometers southwest of Melbourne.
```
After fix:
```
> echo "what is the captial of australia" | .\ollama.exe run phi
The capital city of Australia is Canberra. It is located in the Australian Capital Territory (ACT) and is home to important government buildings, such as Parliament House and the National Museum of Australia
```
Fixes #2698 | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/3068/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/3068/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/5146 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/5146/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/5146/comments | https://api.github.com/repos/ollama/ollama/issues/5146/events | https://github.com/ollama/ollama/pull/5146 | 2,362,711,736 | PR_kwDOJ0Z1Ps5y-1ig | 5,146 | Put back temporary intel GPU env var | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2024-06-19T15:58:50 | 2024-06-19T16:12:48 | 2024-06-19T16:12:45 | COLLABORATOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/5146",
"html_url": "https://github.com/ollama/ollama/pull/5146",
"diff_url": "https://github.com/ollama/ollama/pull/5146.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5146.patch",
"merged_at": "2024-06-19T16:12:45"
} | Until we merge #4876 lets keep the opt-in env var to avoid confusion in the binary releases if we discover an Intel GPU but don't actually have the runner built in.
This reverts commit 755b4e4fc291366595ed7bfb37c2a91ff5834df8. | {
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/5146/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/5146/timeline | null | null | true |
https://api.github.com/repos/ollama/ollama/issues/4219 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/4219/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/4219/comments | https://api.github.com/repos/ollama/ollama/issues/4219/events | https://github.com/ollama/ollama/issues/4219 | 2,282,184,189 | I_kwDOJ0Z1Ps6IB139 | 4,219 | 模型整理 - Categorize models on ollama.com | {
"login": "syssbs",
"id": 129733386,
"node_id": "U_kgDOB7uTCg",
"avatar_url": "https://avatars.githubusercontent.com/u/129733386?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/syssbs",
"html_url": "https://github.com/syssbs",
"followers_url": "https://api.github.com/users/syssbs/followers",
"following_url": "https://api.github.com/users/syssbs/following{/other_user}",
"gists_url": "https://api.github.com/users/syssbs/gists{/gist_id}",
"starred_url": "https://api.github.com/users/syssbs/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/syssbs/subscriptions",
"organizations_url": "https://api.github.com/users/syssbs/orgs",
"repos_url": "https://api.github.com/users/syssbs/repos",
"events_url": "https://api.github.com/users/syssbs/events{/privacy}",
"received_events_url": "https://api.github.com/users/syssbs/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 6573197867,
"node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw",
"url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com",
"name": "ollama.com",
"color": "ffffff",
"default": false,
"description": ""
}
] | open | false | null | [] | null | 3 | 2024-05-07T03:16:01 | 2024-07-25T18:15:27 | null | NONE | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | null | null | 麻烦把官网的模型进行分类整理下吧,模型太多了以后也会越来越多,官网的模型列表感觉会很乱 | null | {
"url": "https://api.github.com/repos/ollama/ollama/issues/4219/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/4219/timeline | null | null | false |
https://api.github.com/repos/ollama/ollama/issues/8307 | https://api.github.com/repos/ollama/ollama | https://api.github.com/repos/ollama/ollama/issues/8307/labels{/name} | https://api.github.com/repos/ollama/ollama/issues/8307/comments | https://api.github.com/repos/ollama/ollama/issues/8307/events | https://github.com/ollama/ollama/pull/8307 | 2,769,004,347 | PR_kwDOJ0Z1Ps6GvOdm | 8,307 | fix: correct endpoint URL to avoid 404 error | {
"login": "ubaldus",
"id": 660076,
"node_id": "MDQ6VXNlcjY2MDA3Ng==",
"avatar_url": "https://avatars.githubusercontent.com/u/660076?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ubaldus",
"html_url": "https://github.com/ubaldus",
"followers_url": "https://api.github.com/users/ubaldus/followers",
"following_url": "https://api.github.com/users/ubaldus/following{/other_user}",
"gists_url": "https://api.github.com/users/ubaldus/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ubaldus/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ubaldus/subscriptions",
"organizations_url": "https://api.github.com/users/ubaldus/orgs",
"repos_url": "https://api.github.com/users/ubaldus/repos",
"events_url": "https://api.github.com/users/ubaldus/events{/privacy}",
"received_events_url": "https://api.github.com/users/ubaldus/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | [] | closed | false | null | [] | null | 0 | 2025-01-04T21:14:22 | 2025-01-04T23:45:16 | 2025-01-04T23:45:16 | CONTRIBUTOR | {
"total": 0,
"completed": 0,
"percent_completed": 0
} | null | false | {
"url": "https://api.github.com/repos/ollama/ollama/pulls/8307",
"html_url": "https://github.com/ollama/ollama/pull/8307",
"diff_url": "https://github.com/ollama/ollama/pull/8307.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8307.patch",
"merged_at": "2025-01-04T23:45:16"
} | null | {
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
} | {
"url": "https://api.github.com/repos/ollama/ollama/issues/8307/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/ollama/ollama/issues/8307/timeline | null | null | true |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.