url
stringlengths
51
54
repository_url
stringclasses
1 value
labels_url
stringlengths
65
68
comments_url
stringlengths
60
63
events_url
stringlengths
58
61
html_url
stringlengths
39
44
id
int64
1.78B
2.82B
node_id
stringlengths
18
19
number
int64
1
8.69k
title
stringlengths
1
382
user
dict
labels
listlengths
0
5
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
2
milestone
null
comments
int64
0
323
created_at
timestamp[s]
updated_at
timestamp[s]
closed_at
timestamp[s]
author_association
stringclasses
4 values
sub_issues_summary
dict
active_lock_reason
null
draft
bool
2 classes
pull_request
dict
body
stringlengths
2
118k
closed_by
dict
reactions
dict
timeline_url
stringlengths
60
63
performed_via_github_app
null
state_reason
stringclasses
4 values
is_pull_request
bool
2 classes
https://api.github.com/repos/ollama/ollama/issues/6525
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6525/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6525/comments
https://api.github.com/repos/ollama/ollama/issues/6525/events
https://github.com/ollama/ollama/issues/6525
2,488,468,101
I_kwDOJ0Z1Ps6UUwKF
6,525
ollama collapses CPU
{ "login": "Hyphaed", "id": 19622367, "node_id": "MDQ6VXNlcjE5NjIyMzY3", "avatar_url": "https://avatars.githubusercontent.com/u/19622367?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Hyphaed", "html_url": "https://github.com/Hyphaed", "followers_url": "https://api.github.com/users/Hyphaed/followers", "following_url": "https://api.github.com/users/Hyphaed/following{/other_user}", "gists_url": "https://api.github.com/users/Hyphaed/gists{/gist_id}", "starred_url": "https://api.github.com/users/Hyphaed/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Hyphaed/subscriptions", "organizations_url": "https://api.github.com/users/Hyphaed/orgs", "repos_url": "https://api.github.com/users/Hyphaed/repos", "events_url": "https://api.github.com/users/Hyphaed/events{/privacy}", "received_events_url": "https://api.github.com/users/Hyphaed/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
9
2024-08-27T07:00:05
2024-09-16T20:41:39
2024-08-27T15:12:18
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? ollama collapses CPU even when I stop the server the CPU still stuck from 75% to 90% even when I do have an RTX 3070 and terminal is showind that is using the GPU there is no error in terminal I have no verbose since I forcelly shutedown the workstation ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.3.6
{ "login": "Hyphaed", "id": 19622367, "node_id": "MDQ6VXNlcjE5NjIyMzY3", "avatar_url": "https://avatars.githubusercontent.com/u/19622367?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Hyphaed", "html_url": "https://github.com/Hyphaed", "followers_url": "https://api.github.com/users/Hyphaed/followers", "following_url": "https://api.github.com/users/Hyphaed/following{/other_user}", "gists_url": "https://api.github.com/users/Hyphaed/gists{/gist_id}", "starred_url": "https://api.github.com/users/Hyphaed/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Hyphaed/subscriptions", "organizations_url": "https://api.github.com/users/Hyphaed/orgs", "repos_url": "https://api.github.com/users/Hyphaed/repos", "events_url": "https://api.github.com/users/Hyphaed/events{/privacy}", "received_events_url": "https://api.github.com/users/Hyphaed/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6525/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6525/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3972
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3972/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3972/comments
https://api.github.com/repos/ollama/ollama/issues/3972/events
https://github.com/ollama/ollama/pull/3972
2,266,851,869
PR_kwDOJ0Z1Ps5t5oCp
3,972
Add support for building on Windows ARM64
{ "login": "hmartinez82", "id": 1100440, "node_id": "MDQ6VXNlcjExMDA0NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/1100440?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hmartinez82", "html_url": "https://github.com/hmartinez82", "followers_url": "https://api.github.com/users/hmartinez82/followers", "following_url": "https://api.github.com/users/hmartinez82/following{/other_user}", "gists_url": "https://api.github.com/users/hmartinez82/gists{/gist_id}", "starred_url": "https://api.github.com/users/hmartinez82/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hmartinez82/subscriptions", "organizations_url": "https://api.github.com/users/hmartinez82/orgs", "repos_url": "https://api.github.com/users/hmartinez82/repos", "events_url": "https://api.github.com/users/hmartinez82/events{/privacy}", "received_events_url": "https://api.github.com/users/hmartinez82/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
7
2024-04-27T05:38:44
2024-05-08T04:05:14
2024-04-28T21:52:59
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3972", "html_url": "https://github.com/ollama/ollama/pull/3972", "diff_url": "https://github.com/ollama/ollama/pull/3972.diff", "patch_url": "https://github.com/ollama/ollama/pull/3972.patch", "merged_at": "2024-04-28T21:52:58" }
Part of #2589 - Builds only the cpu runner for ARM64 Also, the existing CMake recipe already enables NEON and Armv8.2 extensions when ARM64 is detected. - I'll create another PR with build instructions. The main trick is that MSY2 has the CLANGARM64 environment that provides gcc aliases to Clang. Maintainer changes are enabled. Feel free to adjust the changes to your liking.
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3972/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3972/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5552
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5552/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5552/comments
https://api.github.com/repos/ollama/ollama/issues/5552/events
https://github.com/ollama/ollama/pull/5552
2,396,671,216
PR_kwDOJ0Z1Ps50wDvt
5,552
docs
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-07-08T22:17:57
2024-07-25T23:26:21
2024-07-25T23:26:19
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5552", "html_url": "https://github.com/ollama/ollama/pull/5552", "diff_url": "https://github.com/ollama/ollama/pull/5552.diff", "patch_url": "https://github.com/ollama/ollama/pull/5552.patch", "merged_at": "2024-07-25T23:26:19" }
part of #5216 part of #5284 part of #5207
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5552/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5552/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/422
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/422/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/422/comments
https://api.github.com/repos/ollama/ollama/issues/422/events
https://github.com/ollama/ollama/issues/422
1,868,154,838
I_kwDOJ0Z1Ps5vWcfW
422
`Error: Post "http://localhost:11434/api/generate": EOF` with long propmts with phind-codellama
{ "login": "tomduncalf", "id": 5458070, "node_id": "MDQ6VXNlcjU0NTgwNzA=", "avatar_url": "https://avatars.githubusercontent.com/u/5458070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tomduncalf", "html_url": "https://github.com/tomduncalf", "followers_url": "https://api.github.com/users/tomduncalf/followers", "following_url": "https://api.github.com/users/tomduncalf/following{/other_user}", "gists_url": "https://api.github.com/users/tomduncalf/gists{/gist_id}", "starred_url": "https://api.github.com/users/tomduncalf/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tomduncalf/subscriptions", "organizations_url": "https://api.github.com/users/tomduncalf/orgs", "repos_url": "https://api.github.com/users/tomduncalf/repos", "events_url": "https://api.github.com/users/tomduncalf/events{/privacy}", "received_events_url": "https://api.github.com/users/tomduncalf/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
20
2023-08-26T16:00:33
2024-04-22T09:12:40
2023-09-07T11:08:40
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
It seems like if you provide a long prompt (I was using one of 1,000ish tokens according to OpenAI tokenizer) with this model, you get an error `Error: Post "http://localhost:11434/api/generate": EOF`. It may or may not relate to the contents of the prompt as well as the length
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/422/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/422/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7405
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7405/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7405/comments
https://api.github.com/repos/ollama/ollama/issues/7405/events
https://github.com/ollama/ollama/issues/7405
2,619,529,082
I_kwDOJ0Z1Ps6cItd6
7,405
Feature request: Add CLI argument to specify a system prompt
{ "login": "Kerrick", "id": 552093, "node_id": "MDQ6VXNlcjU1MjA5Mw==", "avatar_url": "https://avatars.githubusercontent.com/u/552093?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Kerrick", "html_url": "https://github.com/Kerrick", "followers_url": "https://api.github.com/users/Kerrick/followers", "following_url": "https://api.github.com/users/Kerrick/following{/other_user}", "gists_url": "https://api.github.com/users/Kerrick/gists{/gist_id}", "starred_url": "https://api.github.com/users/Kerrick/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Kerrick/subscriptions", "organizations_url": "https://api.github.com/users/Kerrick/orgs", "repos_url": "https://api.github.com/users/Kerrick/repos", "events_url": "https://api.github.com/users/Kerrick/events{/privacy}", "received_events_url": "https://api.github.com/users/Kerrick/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
1
2024-10-28T20:54:54
2024-10-29T21:48:49
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I'd like to be able to set the system prompt from the call to `ollama` in my shell, rather than in the conversation. For example: ```sh ollama run llama3.1 --system="Your nickname is 'Grass' now" ``` ...or... ```sh ollama run llama3.1 -s "system" "Your nickname is 'Grass' now" ``` With this ability, I could set up aliases in my shell profile so that I can run system-prompt-customized versions of a model with a single command. ```sh alias grass="ollama run llama3.1 --system=\"Your nickname is 'Grass' now\"" ``` It'd be even better if I could also specify a path to a text file with a system prompt (e.g. `ollama run llama3.1 --system-file="~/system_prompts/grass.txt"`, but that wouldn't be necessary. This is unique from #807, whose fix only works in the conversation.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7405/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7405/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/7620
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7620/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7620/comments
https://api.github.com/repos/ollama/ollama/issues/7620/events
https://github.com/ollama/ollama/pull/7620
2,649,452,239
PR_kwDOJ0Z1Ps6BhB1y
7,620
api: fix typo in Golang API types docs
{ "login": "neomantra", "id": 26842, "node_id": "MDQ6VXNlcjI2ODQy", "avatar_url": "https://avatars.githubusercontent.com/u/26842?v=4", "gravatar_id": "", "url": "https://api.github.com/users/neomantra", "html_url": "https://github.com/neomantra", "followers_url": "https://api.github.com/users/neomantra/followers", "following_url": "https://api.github.com/users/neomantra/following{/other_user}", "gists_url": "https://api.github.com/users/neomantra/gists{/gist_id}", "starred_url": "https://api.github.com/users/neomantra/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/neomantra/subscriptions", "organizations_url": "https://api.github.com/users/neomantra/orgs", "repos_url": "https://api.github.com/users/neomantra/repos", "events_url": "https://api.github.com/users/neomantra/events{/privacy}", "received_events_url": "https://api.github.com/users/neomantra/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-11-11T14:10:40
2024-12-08T17:32:55
2024-11-12T00:21:58
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7620", "html_url": "https://github.com/ollama/ollama/pull/7620", "diff_url": "https://github.com/ollama/ollama/pull/7620.diff", "patch_url": "https://github.com/ollama/ollama/pull/7620.patch", "merged_at": "2024-11-12T00:21:58" }
Fixes minor typos and grammar in `api/types.go` I had only reviewed `client.go` in my commit yesterday, sorry I didn't check this one too. Somehow the last PR had Python in the title, but it and this affect Golang.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7620/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7620/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8485
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8485/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8485/comments
https://api.github.com/repos/ollama/ollama/issues/8485/events
https://github.com/ollama/ollama/issues/8485
2,797,657,711
I_kwDOJ0Z1Ps6mwN5v
8,485
[0.5.7] small models are loaded to GPU, but inference is slow and using a lot of CPU
{ "login": "kha84", "id": 110789576, "node_id": "U_kgDOBpqDyA", "avatar_url": "https://avatars.githubusercontent.com/u/110789576?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kha84", "html_url": "https://github.com/kha84", "followers_url": "https://api.github.com/users/kha84/followers", "following_url": "https://api.github.com/users/kha84/following{/other_user}", "gists_url": "https://api.github.com/users/kha84/gists{/gist_id}", "starred_url": "https://api.github.com/users/kha84/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kha84/subscriptions", "organizations_url": "https://api.github.com/users/kha84/orgs", "repos_url": "https://api.github.com/users/kha84/repos", "events_url": "https://api.github.com/users/kha84/events{/privacy}", "received_events_url": "https://api.github.com/users/kha84/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
6
2025-01-19T14:33:04
2025-01-20T09:50:51
2025-01-20T09:50:51
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Hello there. Just upgraded from ollama 0.4.x version to the latest one (0.5.7) and immediately noticed that inference of all models (even small ones, like llama 3.2 3B) become very slow. Like orders of magnitude slow. I can see that during inference CPU is being used intensively, even though the model itself is loaded into VRAM of GPU and there's a lot of VRAM is still free (as per nvtop). OS Ubuntu 22 LTS RTX 4090 ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.5.7
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8485/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8485/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4145
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4145/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4145/comments
https://api.github.com/repos/ollama/ollama/issues/4145/events
https://github.com/ollama/ollama/pull/4145
2,278,580,561
PR_kwDOJ0Z1Ps5uhXWg
4,145
Fix lint warnings
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-05-03T23:44:35
2024-05-03T23:53:20
2024-05-03T23:53:17
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4145", "html_url": "https://github.com/ollama/ollama/pull/4145", "diff_url": "https://github.com/ollama/ollama/pull/4145.diff", "patch_url": "https://github.com/ollama/ollama/pull/4145.patch", "merged_at": "2024-05-03T23:53:17" }
null
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4145/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4145/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/290
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/290/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/290/comments
https://api.github.com/repos/ollama/ollama/issues/290/events
https://github.com/ollama/ollama/pull/290
1,837,466,485
PR_kwDOJ0Z1Ps5XPFzC
290
implement loading ggml lora adapters through the modelfile
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-08-05T00:21:45
2023-08-11T00:23:03
2023-08-11T00:23:01
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/290", "html_url": "https://github.com/ollama/ollama/pull/290", "diff_url": "https://github.com/ollama/ollama/pull/290.diff", "patch_url": "https://github.com/ollama/ollama/pull/290.patch", "merged_at": "2023-08-11T00:23:01" }
LoRA adapters can be added to Ollama models through the Modelfile and automatically applied when the model is loaded: ``` FROM llama2:13b TEMPLATE {{ .Prompt }} ADAPTER ./llama2-13b-storywriter-lora.ggml.bin ``` A few caveats: * LoRA adapters must be GGML. If the adapter isn't GGML, it can be converted with the `convert-lora-to-ggml.py` script in https://github.com/ggerganov/llama.cpp * Using adapters with quantized weights might not produce good results * Using adapters disables mmap * It's possible to apply multiple adapters but 1. Ordering is important 2. There may be unintended side effects since most adapters are not intended to be layered on other adapters 3. Performance may degrade with more adapters
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/290/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/290/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6651
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6651/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6651/comments
https://api.github.com/repos/ollama/ollama/issues/6651/events
https://github.com/ollama/ollama/issues/6651
2,507,064,853
I_kwDOJ0Z1Ps6VbsYV
6,651
The speed of using embedded models is much slower compared to xinference
{ "login": "yushengliao", "id": 29765903, "node_id": "MDQ6VXNlcjI5NzY1OTAz", "avatar_url": "https://avatars.githubusercontent.com/u/29765903?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yushengliao", "html_url": "https://github.com/yushengliao", "followers_url": "https://api.github.com/users/yushengliao/followers", "following_url": "https://api.github.com/users/yushengliao/following{/other_user}", "gists_url": "https://api.github.com/users/yushengliao/gists{/gist_id}", "starred_url": "https://api.github.com/users/yushengliao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yushengliao/subscriptions", "organizations_url": "https://api.github.com/users/yushengliao/orgs", "repos_url": "https://api.github.com/users/yushengliao/repos", "events_url": "https://api.github.com/users/yushengliao/events{/privacy}", "received_events_url": "https://api.github.com/users/yushengliao/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 5808482718, "node_id": "LA_kwDOJ0Z1Ps8AAAABWjZpng", "url": "https://api.github.com/repos/ollama/ollama/labels/performance", "name": "performance", "color": "A5B5C6", "default": false, "description": "" } ]
open
false
null
[]
null
0
2024-09-05T07:58:41
2024-09-05T16:17:42
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I use the BGE-M3 model and send the same request, especially with xinference taking about 10 seconds and ollama taking about 200 seconds. I'm sure they all use GPUs. I found that xinference allocates more video memory, while ollama's video memory usage remains basically unchanged. Perhaps this is the reason for the speed difference?
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6651/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/ollama/ollama/issues/6651/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/891
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/891/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/891/comments
https://api.github.com/repos/ollama/ollama/issues/891/events
https://github.com/ollama/ollama/issues/891
1,959,768,778
I_kwDOJ0Z1Ps50z7LK
891
Support remote `ollama create`
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
0
2023-10-24T17:47:30
2023-11-16T00:41:14
2023-11-16T00:41:14
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
`ollama create` should support remote instances of Ollama with `OLLAMA_HOST` ``` OLLAMA_HOST=my-test-host:11434 ollama create my-model -f ./Modelfile ```
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/891/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/891/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3323
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3323/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3323/comments
https://api.github.com/repos/ollama/ollama/issues/3323/events
https://github.com/ollama/ollama/issues/3323
2,204,255,976
I_kwDOJ0Z1Ps6DYkbo
3,323
Feat req: Add "Last updated" sorting to the hub
{ "login": "knoopx", "id": 100993, "node_id": "MDQ6VXNlcjEwMDk5Mw==", "avatar_url": "https://avatars.githubusercontent.com/u/100993?v=4", "gravatar_id": "", "url": "https://api.github.com/users/knoopx", "html_url": "https://github.com/knoopx", "followers_url": "https://api.github.com/users/knoopx/followers", "following_url": "https://api.github.com/users/knoopx/following{/other_user}", "gists_url": "https://api.github.com/users/knoopx/gists{/gist_id}", "starred_url": "https://api.github.com/users/knoopx/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/knoopx/subscriptions", "organizations_url": "https://api.github.com/users/knoopx/orgs", "repos_url": "https://api.github.com/users/knoopx/repos", "events_url": "https://api.github.com/users/knoopx/events{/privacy}", "received_events_url": "https://api.github.com/users/knoopx/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 6573197867, "node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw", "url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com", "name": "ollama.com", "color": "ffffff", "default": false, "description": "" } ]
closed
false
null
[]
null
1
2024-03-24T09:09:12
2024-07-18T19:04:28
2024-07-18T19:04:28
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What are you trying to do? There's no way to discover recent updates for existing models. ### How should we solve this? Add "Last updated" sort choice ### What is the impact of not solving this? No way to find out recently updated models ### Anything else? _No response_
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3323/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3323/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7820
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7820/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7820/comments
https://api.github.com/repos/ollama/ollama/issues/7820/events
https://github.com/ollama/ollama/issues/7820
2,687,973,667
I_kwDOJ0Z1Ps6gNzkj
7,820
Instant closure when using shell input with piped output.
{ "login": "WyvernDotRed", "id": 41121402, "node_id": "MDQ6VXNlcjQxMTIxNDAy", "avatar_url": "https://avatars.githubusercontent.com/u/41121402?v=4", "gravatar_id": "", "url": "https://api.github.com/users/WyvernDotRed", "html_url": "https://github.com/WyvernDotRed", "followers_url": "https://api.github.com/users/WyvernDotRed/followers", "following_url": "https://api.github.com/users/WyvernDotRed/following{/other_user}", "gists_url": "https://api.github.com/users/WyvernDotRed/gists{/gist_id}", "starred_url": "https://api.github.com/users/WyvernDotRed/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/WyvernDotRed/subscriptions", "organizations_url": "https://api.github.com/users/WyvernDotRed/orgs", "repos_url": "https://api.github.com/users/WyvernDotRed/repos", "events_url": "https://api.github.com/users/WyvernDotRed/events{/privacy}", "received_events_url": "https://api.github.com/users/WyvernDotRed/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" } ]
open
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
6
2024-11-24T16:36:36
2024-12-10T21:07:24
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? When running `ollama run [model] | cat` or `ollama run [model] > [file]`, ollama now closes immediately and does not accept any manual input. `ollama run [model]` still functions as expected. While `cat | ollama run [model] ...` seems to be the workaround, this requires entering ^d to have the input processed. Thus closing the input stream and also ollama as a result, expected in this case since input was explicitly closed. Additionally, thanks to [pull 416](https://github.com/ollama/ollama/pull/416), this removes the ability to queue or follow up with further prompts. This regression seems specific to 0.4.4. After downgrading to 0.4.3 or 0.4.2, ollama functions as I expect. I have not tested older versions and will stick to version 0.4.3 for the time being. The purpose of the above syntax being to use the ability of the chat functionality for to enter multiple prompts while processing the output by a further script. Though this will be covered in a following feature request and is only tangentially related to this regression. Thank you for your time, have a nice rest of your day! ### OS Linux ### GPU AMD ### CPU AMD ### Ollama version 0.4.4
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7820/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7820/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/7239
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7239/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7239/comments
https://api.github.com/repos/ollama/ollama/issues/7239/events
https://github.com/ollama/ollama/issues/7239
2,594,416,445
I_kwDOJ0Z1Ps6ao6c9
7,239
Add Tab-Enabled Autocomplete for Local Model Parameters in Ollama CLI
{ "login": "lucianoayres", "id": 20209393, "node_id": "MDQ6VXNlcjIwMjA5Mzkz", "avatar_url": "https://avatars.githubusercontent.com/u/20209393?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lucianoayres", "html_url": "https://github.com/lucianoayres", "followers_url": "https://api.github.com/users/lucianoayres/followers", "following_url": "https://api.github.com/users/lucianoayres/following{/other_user}", "gists_url": "https://api.github.com/users/lucianoayres/gists{/gist_id}", "starred_url": "https://api.github.com/users/lucianoayres/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lucianoayres/subscriptions", "organizations_url": "https://api.github.com/users/lucianoayres/orgs", "repos_url": "https://api.github.com/users/lucianoayres/repos", "events_url": "https://api.github.com/users/lucianoayres/events{/privacy}", "received_events_url": "https://api.github.com/users/lucianoayres/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
3
2024-10-17T11:20:14
2025-01-13T00:46:55
2025-01-13T00:46:55
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
It would greatly enhance usability if the Ollama CLI supported tab-autocomplete for model names when using commands like `run`, `show`, `list`, etc. For example: ```bash # This would autocomplete to something like `llama3.2`, based on the locally available models. ollama run lla<TAB> ``` Implementing this across all commands where model names are parameters would improve efficiency and user experience. Thank you!
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7239/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7239/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2138
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2138/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2138/comments
https://api.github.com/repos/ollama/ollama/issues/2138/events
https://github.com/ollama/ollama/pull/2138
2,094,374,620
PR_kwDOJ0Z1Ps5kvv-F
2,138
Update README.md - Community Integrations - Obsidian Local GPT plugin
{ "login": "pfrankov", "id": 584632, "node_id": "MDQ6VXNlcjU4NDYzMg==", "avatar_url": "https://avatars.githubusercontent.com/u/584632?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pfrankov", "html_url": "https://github.com/pfrankov", "followers_url": "https://api.github.com/users/pfrankov/followers", "following_url": "https://api.github.com/users/pfrankov/following{/other_user}", "gists_url": "https://api.github.com/users/pfrankov/gists{/gist_id}", "starred_url": "https://api.github.com/users/pfrankov/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pfrankov/subscriptions", "organizations_url": "https://api.github.com/users/pfrankov/orgs", "repos_url": "https://api.github.com/users/pfrankov/repos", "events_url": "https://api.github.com/users/pfrankov/events{/privacy}", "received_events_url": "https://api.github.com/users/pfrankov/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-01-22T17:11:44
2024-02-22T15:52:36
2024-02-22T15:52:36
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2138", "html_url": "https://github.com/ollama/ollama/pull/2138", "diff_url": "https://github.com/ollama/ollama/pull/2138.diff", "patch_url": "https://github.com/ollama/ollama/pull/2138.patch", "merged_at": "2024-02-22T15:52:36" }
Local GPT plugin for Obsidian mainly relies on Ollama provider ![image](https://github.com/pfrankov/obsidian-local-gpt/assets/584632/724d4399-cb6c-4531-9f04-a1e5df2e3dad) Also works with images <img width="400" src="https://github.com/pfrankov/obsidian-local-gpt/assets/584632/a05d68fa-5419-4386-ac43-82b9513999ad"> <img width="598" alt="Settings" src="https://github.com/pfrankov/obsidian-local-gpt/assets/584632/6ab2d802-13ed-42be-aab1-6a3f689b18a0"> I'd say that Local GPT plugin is enhanced version of [Obsidian Ollama plugin](https://github.com/hinterdupfinger/obsidian-ollama) in every way. Tried to resolve merge conflicts in https://github.com/jmorganca/ollama/pull/1662 but accidentally closed it.
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2138/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2138/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5783
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5783/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5783/comments
https://api.github.com/repos/ollama/ollama/issues/5783/events
https://github.com/ollama/ollama/issues/5783
2,417,335,104
I_kwDOJ0Z1Ps6QFZtA
5,783
erorr loading models x3 7900 XTX #5708
{ "login": "darwinvelez58", "id": 118543481, "node_id": "U_kgDOBxDUeQ", "avatar_url": "https://avatars.githubusercontent.com/u/118543481?v=4", "gravatar_id": "", "url": "https://api.github.com/users/darwinvelez58", "html_url": "https://github.com/darwinvelez58", "followers_url": "https://api.github.com/users/darwinvelez58/followers", "following_url": "https://api.github.com/users/darwinvelez58/following{/other_user}", "gists_url": "https://api.github.com/users/darwinvelez58/gists{/gist_id}", "starred_url": "https://api.github.com/users/darwinvelez58/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/darwinvelez58/subscriptions", "organizations_url": "https://api.github.com/users/darwinvelez58/orgs", "repos_url": "https://api.github.com/users/darwinvelez58/repos", "events_url": "https://api.github.com/users/darwinvelez58/events{/privacy}", "received_events_url": "https://api.github.com/users/darwinvelez58/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-07-18T20:28:02
2024-07-22T23:07:52
2024-07-22T23:07:52
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Few Days ago I report this error #5708, #5710 suppose to fix the issue but I still have the same error. ### OS Linux ### GPU AMD ### CPU AMD ### Ollama version 0.2.6-rocm
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5783/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5783/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1528
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1528/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1528/comments
https://api.github.com/repos/ollama/ollama/issues/1528/events
https://github.com/ollama/ollama/pull/1528
2,042,606,696
PR_kwDOJ0Z1Ps5iDRhN
1,528
Add unit test of API routes
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-12-14T22:35:10
2023-12-15T00:47:41
2023-12-15T00:47:40
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1528", "html_url": "https://github.com/ollama/ollama/pull/1528", "diff_url": "https://github.com/ollama/ollama/pull/1528.diff", "patch_url": "https://github.com/ollama/ollama/pull/1528.patch", "merged_at": "2023-12-15T00:47:40" }
This change modifies the base server to allow it to be more easily unit tested. It also adds in a simple unit test to "/api/version" to demonstrate how to add unit tests in the future.
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1528/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1528/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5830
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5830/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5830/comments
https://api.github.com/repos/ollama/ollama/issues/5830/events
https://github.com/ollama/ollama/issues/5830
2,421,426,851
I_kwDOJ0Z1Ps6QVAqj
5,830
OpenAI endpoint gives 404
{ "login": "defaultsecurity", "id": 34036534, "node_id": "MDQ6VXNlcjM0MDM2NTM0", "avatar_url": "https://avatars.githubusercontent.com/u/34036534?v=4", "gravatar_id": "", "url": "https://api.github.com/users/defaultsecurity", "html_url": "https://github.com/defaultsecurity", "followers_url": "https://api.github.com/users/defaultsecurity/followers", "following_url": "https://api.github.com/users/defaultsecurity/following{/other_user}", "gists_url": "https://api.github.com/users/defaultsecurity/gists{/gist_id}", "starred_url": "https://api.github.com/users/defaultsecurity/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/defaultsecurity/subscriptions", "organizations_url": "https://api.github.com/users/defaultsecurity/orgs", "repos_url": "https://api.github.com/users/defaultsecurity/repos", "events_url": "https://api.github.com/users/defaultsecurity/events{/privacy}", "received_events_url": "https://api.github.com/users/defaultsecurity/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-07-21T14:20:07
2024-07-22T06:18:03
2024-07-22T06:18:03
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? - http://localhost:11434/v1/chat/completions (gives 404) - http://localhost:11434 (shows ollama is running) Otherwise Ollama is working. I'm not sure what to do. ### OS Linux ### GPU Nvidia ### CPU AMD ### Ollama version 0.2.7
{ "login": "defaultsecurity", "id": 34036534, "node_id": "MDQ6VXNlcjM0MDM2NTM0", "avatar_url": "https://avatars.githubusercontent.com/u/34036534?v=4", "gravatar_id": "", "url": "https://api.github.com/users/defaultsecurity", "html_url": "https://github.com/defaultsecurity", "followers_url": "https://api.github.com/users/defaultsecurity/followers", "following_url": "https://api.github.com/users/defaultsecurity/following{/other_user}", "gists_url": "https://api.github.com/users/defaultsecurity/gists{/gist_id}", "starred_url": "https://api.github.com/users/defaultsecurity/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/defaultsecurity/subscriptions", "organizations_url": "https://api.github.com/users/defaultsecurity/orgs", "repos_url": "https://api.github.com/users/defaultsecurity/repos", "events_url": "https://api.github.com/users/defaultsecurity/events{/privacy}", "received_events_url": "https://api.github.com/users/defaultsecurity/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5830/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5830/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2920
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2920/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2920/comments
https://api.github.com/repos/ollama/ollama/issues/2920/events
https://github.com/ollama/ollama/issues/2920
2,167,396,829
I_kwDOJ0Z1Ps6BL9nd
2,920
ollama call failed with status code 500 llama 2
{ "login": "sabahatza", "id": 135341585, "node_id": "U_kgDOCBEmEQ", "avatar_url": "https://avatars.githubusercontent.com/u/135341585?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sabahatza", "html_url": "https://github.com/sabahatza", "followers_url": "https://api.github.com/users/sabahatza/followers", "following_url": "https://api.github.com/users/sabahatza/following{/other_user}", "gists_url": "https://api.github.com/users/sabahatza/gists{/gist_id}", "starred_url": "https://api.github.com/users/sabahatza/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sabahatza/subscriptions", "organizations_url": "https://api.github.com/users/sabahatza/orgs", "repos_url": "https://api.github.com/users/sabahatza/repos", "events_url": "https://api.github.com/users/sabahatza/events{/privacy}", "received_events_url": "https://api.github.com/users/sabahatza/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
3
2024-03-04T17:44:32
2024-03-04T17:57:26
2024-03-04T17:48:17
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi team, I am trying to run the llama2 model locally ( I was doing it previously for the last couple of weeks without any problems), but now I face the following error when I am trying to -> ollama run llama2 `Error: error loading model /Users/S_Z/.ollama/models/blobs/sha256:8934d96d3f08982e95922b2b7a2c626a1fe873d7c3b06e8e56d7bc0a1` No ollama models are running for that matter. for context, I have an M1 Max Macbook. Is there any fix to this?
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2920/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2920/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/601
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/601/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/601/comments
https://api.github.com/repos/ollama/ollama/issues/601/events
https://github.com/ollama/ollama/pull/601
1,912,778,743
PR_kwDOJ0Z1Ps5bMTBe
601
Update README.md for linux + cleanup
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-09-26T06:30:46
2023-09-26T06:44:54
2023-09-26T06:44:53
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/601", "html_url": "https://github.com/ollama/ollama/pull/601", "diff_url": "https://github.com/ollama/ollama/pull/601.diff", "patch_url": "https://github.com/ollama/ollama/pull/601.patch", "merged_at": "2023-09-26T06:44:53" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/601/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/601/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5208
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5208/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5208/comments
https://api.github.com/repos/ollama/ollama/issues/5208/events
https://github.com/ollama/ollama/pull/5208
2,367,386,322
PR_kwDOJ0Z1Ps5zOuoI
5,208
Support image input for OpenAI chat compatibility
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
8
2024-06-22T00:31:49
2024-07-30T20:10:05
2024-07-14T05:07:45
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5208", "html_url": "https://github.com/ollama/ollama/pull/5208", "diff_url": "https://github.com/ollama/ollama/pull/5208.diff", "patch_url": "https://github.com/ollama/ollama/pull/5208.patch", "merged_at": "2024-07-14T05:07:45" }
Supports passing in base64 encoded image into image_url. E.g. ``` curl http://localhost:11434/v1/chat/completions \ -H "Content-Type: application/json" \ -d '{ "model": "llava", "messages": [ { "role": "user", "content": [ { "type": "text", "text": "What'\''s in this image?" }, { "type": "image_url", "image_url": { "url": "'$image'" } } ] } ], "max_tokens": 300 }' | jq ``` ``` { "id": "chatcmpl-659", "object": "chat.completion", "created": 1719016156, "model": "llava", "system_fingerprint": "fp_ollama", "choices": [ { "index": 0, "message": { "role": "assistant", "content": " The image shows a cute cartoon of an animal. It appears to be a dog or similar creature, styled with exaggerated features typical in internet memes. The character has big eyes, a round face, and its arms are raised in the air, as if waving or giving a thumbs-up gesture. There's also some motion blur that gives the impression of movement, suggesting the animal might be jumping or dancing. This kind of image is often used in digital communication to convey emotions or add a playful element to text messages. " }, "finish_reason": "stop" } ], "usage": { "prompt_tokens": 1, "completion_tokens": 112, "total_tokens": 113 } } ``` Resolves #3690, #5304
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5208/reactions", "total_count": 6, "+1": 0, "-1": 0, "laugh": 0, "hooray": 6, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5208/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1640
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1640/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1640/comments
https://api.github.com/repos/ollama/ollama/issues/1640/events
https://github.com/ollama/ollama/pull/1640
2,051,142,771
PR_kwDOJ0Z1Ps5igMrN
1,640
added logprobs (`n_probs`)
{ "login": "janpf", "id": 9437600, "node_id": "MDQ6VXNlcjk0Mzc2MDA=", "avatar_url": "https://avatars.githubusercontent.com/u/9437600?v=4", "gravatar_id": "", "url": "https://api.github.com/users/janpf", "html_url": "https://github.com/janpf", "followers_url": "https://api.github.com/users/janpf/followers", "following_url": "https://api.github.com/users/janpf/following{/other_user}", "gists_url": "https://api.github.com/users/janpf/gists{/gist_id}", "starred_url": "https://api.github.com/users/janpf/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/janpf/subscriptions", "organizations_url": "https://api.github.com/users/janpf/orgs", "repos_url": "https://api.github.com/users/janpf/repos", "events_url": "https://api.github.com/users/janpf/events{/privacy}", "received_events_url": "https://api.github.com/users/janpf/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
42
2023-12-20T19:26:27
2025-01-07T19:25:56
2025-01-07T19:25:56
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1640", "html_url": "https://github.com/ollama/ollama/pull/1640", "diff_url": "https://github.com/ollama/ollama/pull/1640.diff", "patch_url": "https://github.com/ollama/ollama/pull/1640.patch", "merged_at": null }
As discussed on discord I implemented the feature. It just passes through the probs from the llamacpp server. Sorry, first time writing Go, might have missed something. https://discord.com/channels/1128867683291627614/1128867684130508875/1187028494228664340
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1640/reactions", "total_count": 38, "+1": 17, "-1": 0, "laugh": 0, "hooray": 11, "confused": 0, "heart": 10, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1640/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3065
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3065/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3065/comments
https://api.github.com/repos/ollama/ollama/issues/3065/events
https://github.com/ollama/ollama/pull/3065
2,180,154,843
PR_kwDOJ0Z1Ps5pTJW3
3,065
relay load model errors to the client
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-03-11T20:25:07
2024-03-11T20:48:28
2024-03-11T20:48:27
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3065", "html_url": "https://github.com/ollama/ollama/pull/3065", "diff_url": "https://github.com/ollama/ollama/pull/3065.diff", "patch_url": "https://github.com/ollama/ollama/pull/3065.patch", "merged_at": "2024-03-11T20:48:27" }
Relay errors on model load, this is needed to help people troubleshoot the specific problem they are experiencing when running a model. This function is a bottle-neck where many different errors can occur. As seen in #2753, there are many issues when the generic "failed to load model" error being reported. In order to effectively identify and fix issues we need to relay more information to the client. This change: - Re-throws load errors, so that the client can output the problem This change does not: - Provide enhanced information to user's about how to resolve their problem. We will have to add these as we see the issues people face, if they can't be fixed on the Ollama side. Errors will be relayed like this: ``` ollama run bad-model Error: exception invalid model dimensions ```
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3065/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3065/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/661
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/661/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/661/comments
https://api.github.com/repos/ollama/ollama/issues/661/events
https://github.com/ollama/ollama/pull/661
1,920,414,250
PR_kwDOJ0Z1Ps5bmHz9
661
Documenting Docker Hub image and OpenAI compatibility
{ "login": "jamesbraza", "id": 8990777, "node_id": "MDQ6VXNlcjg5OTA3Nzc=", "avatar_url": "https://avatars.githubusercontent.com/u/8990777?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jamesbraza", "html_url": "https://github.com/jamesbraza", "followers_url": "https://api.github.com/users/jamesbraza/followers", "following_url": "https://api.github.com/users/jamesbraza/following{/other_user}", "gists_url": "https://api.github.com/users/jamesbraza/gists{/gist_id}", "starred_url": "https://api.github.com/users/jamesbraza/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jamesbraza/subscriptions", "organizations_url": "https://api.github.com/users/jamesbraza/orgs", "repos_url": "https://api.github.com/users/jamesbraza/repos", "events_url": "https://api.github.com/users/jamesbraza/events{/privacy}", "received_events_url": "https://api.github.com/users/jamesbraza/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2023-09-30T22:09:47
2023-10-25T20:18:44
2023-10-24T23:15:30
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/661", "html_url": "https://github.com/ollama/ollama/pull/661", "diff_url": "https://github.com/ollama/ollama/pull/661.diff", "patch_url": "https://github.com/ollama/ollama/pull/661.patch", "merged_at": null }
- Closes https://github.com/jmorganca/ollama/issues/538 - Upstreams some knowledge from https://github.com/jmorganca/ollama/issues/546 - Simplifies `brew install` to one line
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/661/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/661/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5233
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5233/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5233/comments
https://api.github.com/repos/ollama/ollama/issues/5233/events
https://github.com/ollama/ollama/issues/5233
2,368,103,862
I_kwDOJ0Z1Ps6NJmW2
5,233
filtering library models based on tags?
{ "login": "itsPreto", "id": 45348368, "node_id": "MDQ6VXNlcjQ1MzQ4MzY4", "avatar_url": "https://avatars.githubusercontent.com/u/45348368?v=4", "gravatar_id": "", "url": "https://api.github.com/users/itsPreto", "html_url": "https://github.com/itsPreto", "followers_url": "https://api.github.com/users/itsPreto/followers", "following_url": "https://api.github.com/users/itsPreto/following{/other_user}", "gists_url": "https://api.github.com/users/itsPreto/gists{/gist_id}", "starred_url": "https://api.github.com/users/itsPreto/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/itsPreto/subscriptions", "organizations_url": "https://api.github.com/users/itsPreto/orgs", "repos_url": "https://api.github.com/users/itsPreto/repos", "events_url": "https://api.github.com/users/itsPreto/events{/privacy}", "received_events_url": "https://api.github.com/users/itsPreto/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 6573197867, "node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw", "url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com", "name": "ollama.com", "color": "ffffff", "default": false, "description": "" } ]
open
false
null
[]
null
0
2024-06-23T01:02:38
2024-07-08T17:17:16
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
can we get something like this for the models library? would be reallyyyy nice! <img width="387" alt="Screenshot 2024-06-22 at 9 01 25 PM" src="https://github.com/ollama/ollama/assets/45348368/bd142627-9426-41af-b451-67dc82c427df">
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5233/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5233/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/5806
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5806/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5806/comments
https://api.github.com/repos/ollama/ollama/issues/5806/events
https://github.com/ollama/ollama/issues/5806
2,420,556,455
I_kwDOJ0Z1Ps6QRsKn
5,806
allowing ollama 3 to access local txt files for a larger memory?
{ "login": "boba1234567890", "id": 107341969, "node_id": "U_kgDOBmXokQ", "avatar_url": "https://avatars.githubusercontent.com/u/107341969?v=4", "gravatar_id": "", "url": "https://api.github.com/users/boba1234567890", "html_url": "https://github.com/boba1234567890", "followers_url": "https://api.github.com/users/boba1234567890/followers", "following_url": "https://api.github.com/users/boba1234567890/following{/other_user}", "gists_url": "https://api.github.com/users/boba1234567890/gists{/gist_id}", "starred_url": "https://api.github.com/users/boba1234567890/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/boba1234567890/subscriptions", "organizations_url": "https://api.github.com/users/boba1234567890/orgs", "repos_url": "https://api.github.com/users/boba1234567890/repos", "events_url": "https://api.github.com/users/boba1234567890/events{/privacy}", "received_events_url": "https://api.github.com/users/boba1234567890/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
3
2024-07-20T04:34:36
2024-09-04T04:29:27
2024-09-04T04:29:26
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
is there a way to allow ollama 3 to use local txt files for a larger memory and maybe other stuff?
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5806/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5806/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5878
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5878/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5878/comments
https://api.github.com/repos/ollama/ollama/issues/5878/events
https://github.com/ollama/ollama/issues/5878
2,425,437,900
I_kwDOJ0Z1Ps6QkT7M
5,878
Apple LLM -> DCLM-7B
{ "login": "dvelez3815", "id": 40648189, "node_id": "MDQ6VXNlcjQwNjQ4MTg5", "avatar_url": "https://avatars.githubusercontent.com/u/40648189?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dvelez3815", "html_url": "https://github.com/dvelez3815", "followers_url": "https://api.github.com/users/dvelez3815/followers", "following_url": "https://api.github.com/users/dvelez3815/following{/other_user}", "gists_url": "https://api.github.com/users/dvelez3815/gists{/gist_id}", "starred_url": "https://api.github.com/users/dvelez3815/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dvelez3815/subscriptions", "organizations_url": "https://api.github.com/users/dvelez3815/orgs", "repos_url": "https://api.github.com/users/dvelez3815/repos", "events_url": "https://api.github.com/users/dvelez3815/events{/privacy}", "received_events_url": "https://api.github.com/users/dvelez3815/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
0
2024-07-23T14:59:28
2024-07-23T15:00:50
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
apple llm https://huggingface.co/apple/DCLM-7B ![image](https://github.com/user-attachments/assets/39d07483-f4e0-4884-9ee4-8149efc45b79)
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5878/reactions", "total_count": 5, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 5, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5878/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/7300
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7300/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7300/comments
https://api.github.com/repos/ollama/ollama/issues/7300/events
https://github.com/ollama/ollama/issues/7300
2,603,110,739
I_kwDOJ0Z1Ps6bKFFT
7,300
Llama3.2-vision Run Error
{ "login": "mruckman1", "id": 10116867, "node_id": "MDQ6VXNlcjEwMTE2ODY3", "avatar_url": "https://avatars.githubusercontent.com/u/10116867?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mruckman1", "html_url": "https://github.com/mruckman1", "followers_url": "https://api.github.com/users/mruckman1/followers", "following_url": "https://api.github.com/users/mruckman1/following{/other_user}", "gists_url": "https://api.github.com/users/mruckman1/gists{/gist_id}", "starred_url": "https://api.github.com/users/mruckman1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mruckman1/subscriptions", "organizations_url": "https://api.github.com/users/mruckman1/orgs", "repos_url": "https://api.github.com/users/mruckman1/repos", "events_url": "https://api.github.com/users/mruckman1/events{/privacy}", "received_events_url": "https://api.github.com/users/mruckman1/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
21
2024-10-21T16:40:09
2024-11-05T16:16:29
2024-10-23T01:29:12
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? 1. Updated Ollama this morning. 2. Entered `ollama run x/llama3.2-vision` on macbook 3. Got below output: > pulling manifest > pulling 652e85aa1e14... 100% ▕████████████████▏ 6.0 GB > pulling 622429e8d318... 100% ▕████████████████▏ 1.9 GB > pulling 962e0f69a367... 100% ▕████████████████▏ 163 B > pulling dc49c86b8ebb... 100% ▕████████████████▏ 30 B > pulling 6a50468ba2a8... 100% ▕████████████████▏ 498 B > verifying sha256 digest > writing manifest > success **> Error: llama runner process has terminated: error:Missing required key: clip.has_text_encoder** Expected: Ollama download without error. ### OS macOS ### GPU Apple ### CPU Apple ### Ollama version 0.3.14
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7300/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7300/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7437
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7437/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7437/comments
https://api.github.com/repos/ollama/ollama/issues/7437/events
https://github.com/ollama/ollama/pull/7437
2,625,863,096
PR_kwDOJ0Z1Ps6AeYjb
7,437
Give unicode test more time to run
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-10-31T02:58:26
2024-10-31T20:35:33
2024-10-31T20:35:31
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7437", "html_url": "https://github.com/ollama/ollama/pull/7437", "diff_url": "https://github.com/ollama/ollama/pull/7437.diff", "patch_url": "https://github.com/ollama/ollama/pull/7437.patch", "merged_at": "2024-10-31T20:35:31" }
Some slower GPUs (or partial CPU/GPU loads) can take more than the default 30s to complete this test
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7437/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7437/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/800
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/800/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/800/comments
https://api.github.com/repos/ollama/ollama/issues/800/events
https://github.com/ollama/ollama/pull/800
1,944,775,176
PR_kwDOJ0Z1Ps5c4Ey-
800
API docs link fix
{ "login": "richawo", "id": 35015261, "node_id": "MDQ6VXNlcjM1MDE1MjYx", "avatar_url": "https://avatars.githubusercontent.com/u/35015261?v=4", "gravatar_id": "", "url": "https://api.github.com/users/richawo", "html_url": "https://github.com/richawo", "followers_url": "https://api.github.com/users/richawo/followers", "following_url": "https://api.github.com/users/richawo/following{/other_user}", "gists_url": "https://api.github.com/users/richawo/gists{/gist_id}", "starred_url": "https://api.github.com/users/richawo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/richawo/subscriptions", "organizations_url": "https://api.github.com/users/richawo/orgs", "repos_url": "https://api.github.com/users/richawo/repos", "events_url": "https://api.github.com/users/richawo/events{/privacy}", "received_events_url": "https://api.github.com/users/richawo/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
3
2023-10-16T09:28:16
2023-10-21T16:00:22
2023-10-21T16:00:21
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/800", "html_url": "https://github.com/ollama/ollama/pull/800", "diff_url": "https://github.com/ollama/ollama/pull/800.diff", "patch_url": "https://github.com/ollama/ollama/pull/800.patch", "merged_at": null }
For some reason, the relative API docs link is broken (api is a particular path in Github). Replaced the API docs link in README.md with the absolute path. Fixes issue #802.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/800/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/800/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4174
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4174/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4174/comments
https://api.github.com/repos/ollama/ollama/issues/4174/events
https://github.com/ollama/ollama/pull/4174
2,279,618,845
PR_kwDOJ0Z1Ps5ukkak
4,174
update libraries for langchain_community + llama3 changed from llama2
{ "login": "Drlordbasil", "id": 126736516, "node_id": "U_kgDOB43YhA", "avatar_url": "https://avatars.githubusercontent.com/u/126736516?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Drlordbasil", "html_url": "https://github.com/Drlordbasil", "followers_url": "https://api.github.com/users/Drlordbasil/followers", "following_url": "https://api.github.com/users/Drlordbasil/following{/other_user}", "gists_url": "https://api.github.com/users/Drlordbasil/gists{/gist_id}", "starred_url": "https://api.github.com/users/Drlordbasil/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Drlordbasil/subscriptions", "organizations_url": "https://api.github.com/users/Drlordbasil/orgs", "repos_url": "https://api.github.com/users/Drlordbasil/repos", "events_url": "https://api.github.com/users/Drlordbasil/events{/privacy}", "received_events_url": "https://api.github.com/users/Drlordbasil/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-05-05T16:46:34
2024-05-06T02:06:32
2024-05-05T23:07:04
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4174", "html_url": "https://github.com/ollama/ollama/pull/4174", "diff_url": "https://github.com/ollama/ollama/pull/4174.diff", "patch_url": "https://github.com/ollama/ollama/pull/4174.patch", "merged_at": "2024-05-05T23:07:04" }
Changed: - run ->invoke for updated lib - updated langchain libraries for non-depreciated - updated llama2 to llama3
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4174/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4174/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3782
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3782/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3782/comments
https://api.github.com/repos/ollama/ollama/issues/3782/events
https://github.com/ollama/ollama/pull/3782
2,254,683,690
PR_kwDOJ0Z1Ps5tQTYN
3,782
Critical fix from llama.cpp JSON grammar to forbid un-escaped escape characters in JSON strings
{ "login": "hughescr", "id": 46348, "node_id": "MDQ6VXNlcjQ2MzQ4", "avatar_url": "https://avatars.githubusercontent.com/u/46348?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hughescr", "html_url": "https://github.com/hughescr", "followers_url": "https://api.github.com/users/hughescr/followers", "following_url": "https://api.github.com/users/hughescr/following{/other_user}", "gists_url": "https://api.github.com/users/hughescr/gists{/gist_id}", "starred_url": "https://api.github.com/users/hughescr/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hughescr/subscriptions", "organizations_url": "https://api.github.com/users/hughescr/orgs", "repos_url": "https://api.github.com/users/hughescr/repos", "events_url": "https://api.github.com/users/hughescr/events{/privacy}", "received_events_url": "https://api.github.com/users/hughescr/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-04-20T19:18:56
2024-06-10T01:53:52
2024-06-09T17:57:09
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3782", "html_url": "https://github.com/ollama/ollama/pull/3782", "diff_url": "https://github.com/ollama/ollama/pull/3782.diff", "patch_url": "https://github.com/ollama/ollama/pull/3782.patch", "merged_at": "2024-06-09T17:57:09" }
JSON generation is broken, as models can insert control characters inside strings, which violates JSON. For example, with the current JSON grammar, models could generate: ``` { "key": "value broken" } ``` This is incorrect, and if a linebreak is wanted in the middle of the string there, it should be: ``` { "key": "value\nbroken" } ``` The former breaks at least the nodejs JSON parser, and likely many many others, since it's not compliant JSON. This PR injects the grammar directly from llama.cpp upstream, which prohibits `\x00` through `\x1f` inside JSON strings, fixing the problem.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3782/reactions", "total_count": 3, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 2, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3782/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7608
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7608/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7608/comments
https://api.github.com/repos/ollama/ollama/issues/7608/events
https://github.com/ollama/ollama/issues/7608
2,648,022,235
I_kwDOJ0Z1Ps6d1Zzb
7,608
pulling manifest error
{ "login": "the-nine-nation", "id": 103977945, "node_id": "U_kgDOBjKT2Q", "avatar_url": "https://avatars.githubusercontent.com/u/103977945?v=4", "gravatar_id": "", "url": "https://api.github.com/users/the-nine-nation", "html_url": "https://github.com/the-nine-nation", "followers_url": "https://api.github.com/users/the-nine-nation/followers", "following_url": "https://api.github.com/users/the-nine-nation/following{/other_user}", "gists_url": "https://api.github.com/users/the-nine-nation/gists{/gist_id}", "starred_url": "https://api.github.com/users/the-nine-nation/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/the-nine-nation/subscriptions", "organizations_url": "https://api.github.com/users/the-nine-nation/orgs", "repos_url": "https://api.github.com/users/the-nine-nation/repos", "events_url": "https://api.github.com/users/the-nine-nation/events{/privacy}", "received_events_url": "https://api.github.com/users/the-nine-nation/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-11-11T03:32:24
2024-11-11T03:34:34
2024-11-11T03:34:34
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? How to skip pulling manifest ? because my mechine without internet. ### OS Linux, Docker ### GPU _No response_ ### CPU Intel ### Ollama version newest
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7608/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7608/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3118
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3118/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3118/comments
https://api.github.com/repos/ollama/ollama/issues/3118/events
https://github.com/ollama/ollama/issues/3118
2,184,554,272
I_kwDOJ0Z1Ps6CNacg
3,118
ollama RAM use on orangepi 5
{ "login": "parzzd", "id": 103915075, "node_id": "U_kgDOBjGeQw", "avatar_url": "https://avatars.githubusercontent.com/u/103915075?v=4", "gravatar_id": "", "url": "https://api.github.com/users/parzzd", "html_url": "https://github.com/parzzd", "followers_url": "https://api.github.com/users/parzzd/followers", "following_url": "https://api.github.com/users/parzzd/following{/other_user}", "gists_url": "https://api.github.com/users/parzzd/gists{/gist_id}", "starred_url": "https://api.github.com/users/parzzd/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/parzzd/subscriptions", "organizations_url": "https://api.github.com/users/parzzd/orgs", "repos_url": "https://api.github.com/users/parzzd/repos", "events_url": "https://api.github.com/users/parzzd/events{/privacy}", "received_events_url": "https://api.github.com/users/parzzd/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-03-13T17:35:48
2024-03-13T18:21:56
2024-03-13T18:15:01
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Im trying model on my SBC, it has 16gb RAM, but the execution just uses 1.6gb, making the model take so much time to process. Is there any parameter or configuration to allow it. im new on Ollama, so any answer would be appreciated. ![scr_proceso](https://github.com/ollama/ollama/assets/103915075/5e43931c-9cb0-40d3-9f5f-975308db87aa) ![times](https://github.com/ollama/ollama/assets/103915075/57d9987a-0b99-4a6c-850b-5066b181619a)
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3118/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3118/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/393
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/393/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/393/comments
https://api.github.com/repos/ollama/ollama/issues/393/events
https://github.com/ollama/ollama/pull/393
1,860,413,586
PR_kwDOJ0Z1Ps5YcUmI
393
use url.URL
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-08-22T01:56:35
2023-08-22T22:51:34
2023-08-22T22:51:33
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/393", "html_url": "https://github.com/ollama/ollama/pull/393", "diff_url": "https://github.com/ollama/ollama/pull/393.diff", "patch_url": "https://github.com/ollama/ollama/pull/393.patch", "merged_at": "2023-08-22T22:51:33" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/393/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/393/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/961
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/961/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/961/comments
https://api.github.com/repos/ollama/ollama/issues/961/events
https://github.com/ollama/ollama/issues/961
1,972,308,192
I_kwDOJ0Z1Ps51jwjg
961
garbage output on small models spread to many GPUs
{ "login": "chymian", "id": 1899961, "node_id": "MDQ6VXNlcjE4OTk5NjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1899961?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chymian", "html_url": "https://github.com/chymian", "followers_url": "https://api.github.com/users/chymian/followers", "following_url": "https://api.github.com/users/chymian/following{/other_user}", "gists_url": "https://api.github.com/users/chymian/gists{/gist_id}", "starred_url": "https://api.github.com/users/chymian/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/chymian/subscriptions", "organizations_url": "https://api.github.com/users/chymian/orgs", "repos_url": "https://api.github.com/users/chymian/repos", "events_url": "https://api.github.com/users/chymian/events{/privacy}", "received_events_url": "https://api.github.com/users/chymian/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
7
2023-11-01T12:43:10
2024-04-23T15:31:40
2024-04-23T15:31:40
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
when loading a small model on multiple GPUs, it produces garbage. the machine has 4 x 3070 (8GB) and an older i5-7400, UBU 22.04, Cuda 11.8 ### How to reproduce starting the server by hand ```bash ollama serve ``` ```bash ollama run zephyr >>> why is the sky blue? acia##############################################################################^C ``` nvidia-smi shows it loads 4 x 1,2 GB ### Workaround Till we get some flags to control the loading to a discrete GPU start the server with `CUDA_VISIBLE_DEVICES` ```bash CUDA_VISIBLE_DEVICES="3" ollama serve # loads the model on GPU No. 4 ``` There seems to be a "memory-low-limit-point", under which this error occures.
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/961/reactions", "total_count": 3, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 2, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/961/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1978
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1978/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1978/comments
https://api.github.com/repos/ollama/ollama/issues/1978/events
https://github.com/ollama/ollama/issues/1978
2,080,398,998
I_kwDOJ0Z1Ps58AF6W
1,978
Error "unknown architecture MistralModel" during quantization
{ "login": "philippgille", "id": 170670, "node_id": "MDQ6VXNlcjE3MDY3MA==", "avatar_url": "https://avatars.githubusercontent.com/u/170670?v=4", "gravatar_id": "", "url": "https://api.github.com/users/philippgille", "html_url": "https://github.com/philippgille", "followers_url": "https://api.github.com/users/philippgille/followers", "following_url": "https://api.github.com/users/philippgille/following{/other_user}", "gists_url": "https://api.github.com/users/philippgille/gists{/gist_id}", "starred_url": "https://api.github.com/users/philippgille/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/philippgille/subscriptions", "organizations_url": "https://api.github.com/users/philippgille/orgs", "repos_url": "https://api.github.com/users/philippgille/repos", "events_url": "https://api.github.com/users/philippgille/events{/privacy}", "received_events_url": "https://api.github.com/users/philippgille/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-01-13T17:16:28
2024-05-07T00:08:53
2024-05-06T23:48:21
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hello :wave: , First of all thank you very much for creating and maintaining ollama! It's so simple to use :+1: Now I wanted to use ollama for creating embeddings, and saw https://huggingface.co/intfloat/e5-mistral-7b-instruct performing very well on the [embeddings benchmark](https://huggingface.co/spaces/mteb/leaderboard). The official ollama model library doesn't contain it yet, so I wanted to create and upload it myself. But during the quantization step (`docker run --rm -v .:/model:Z ollama/quantize -q q4_0 /model`) I get the error: > unknown architecture MistralModel As Mistral is supported by ollama, I'm wondering about this error. The E5 model is based on the Mistral instruct v0.1 one, so I assume it's the same architecture. Right? Is maybe just the `ollama/quantize` image not updated with the support yet?
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1978/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1978/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4861
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4861/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4861/comments
https://api.github.com/repos/ollama/ollama/issues/4861/events
https://github.com/ollama/ollama/issues/4861
2,338,549,595
I_kwDOJ0Z1Ps6LY29b
4,861
Jetson - "ollama run" command loads until timeout
{ "login": "Vassar-HARPER-Project", "id": 171359116, "node_id": "U_kgDOCja7jA", "avatar_url": "https://avatars.githubusercontent.com/u/171359116?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Vassar-HARPER-Project", "html_url": "https://github.com/Vassar-HARPER-Project", "followers_url": "https://api.github.com/users/Vassar-HARPER-Project/followers", "following_url": "https://api.github.com/users/Vassar-HARPER-Project/following{/other_user}", "gists_url": "https://api.github.com/users/Vassar-HARPER-Project/gists{/gist_id}", "starred_url": "https://api.github.com/users/Vassar-HARPER-Project/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Vassar-HARPER-Project/subscriptions", "organizations_url": "https://api.github.com/users/Vassar-HARPER-Project/orgs", "repos_url": "https://api.github.com/users/Vassar-HARPER-Project/repos", "events_url": "https://api.github.com/users/Vassar-HARPER-Project/events{/privacy}", "received_events_url": "https://api.github.com/users/Vassar-HARPER-Project/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
11
2024-06-06T15:34:26
2024-11-12T18:31:55
2024-11-12T18:31:55
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Upon running "ollama run gemma:2b" (though this happens for all tested models: llama3, phi, tinyllama), the loading animation appears and after ~5 minutes (estimate, untimed), the response / result of the command is: `Error: timed out waiting for llama runner to start - progress 1.00 - ` the server shows this log for this command: ``` 2024/06/06 11:21:53 routes.go:1007: INFO server config env="map[OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_HOST: OLLAMA_KEEP_ALIVE: OLLAMA_LLM_LIBRARY: OLLAMA_MAX_LOADED_MODELS:1 OLLAMA_MAX_QUEUE:512 OLLAMA_MAX_VRAM:0 OLLAMA_MODELS: OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:*] OLLAMA_RUNNERS_DIR: OLLAMA_TMPDIR:]" time=2024-06-06T11:21:53.848-04:00 level=INFO source=images.go:729 msg="total blobs: 11" time=2024-06-06T11:21:53.849-04:00 level=INFO source=images.go:736 msg="total unused blobs removed: 0" time=2024-06-06T11:21:53.849-04:00 level=INFO source=routes.go:1053 msg="Listening on 127.0.0.1:11434 (version 0.1.41)" time=2024-06-06T11:21:53.850-04:00 level=INFO source=payload.go:30 msg="extracting embedded files" dir=/tmp/ollama3794080172/runners time=2024-06-06T11:21:58.984-04:00 level=INFO source=payload.go:44 msg="Dynamic LLM libraries [cpu cuda_v11]" time=2024-06-06T11:21:59.082-04:00 level=INFO source=types.go:71 msg="inference compute" id=GPU-42638932-6929-58db-a006-34d50a6799c1 library=cuda compute=8.7 driver=11.4 name=Orin total="29.9 GiB" available="21.7 GiB" [GIN] 2024/06/06 - 11:22:14 | 200 | 64.512µs | 127.0.0.1 | HEAD "/" [GIN] 2024/06/06 - 11:22:14 | 200 | 1.232036ms | 127.0.0.1 | POST "/api/show" [GIN] 2024/06/06 - 11:22:14 | 200 | 717.058µs | 127.0.0.1 | POST "/api/show" time=2024-06-06T11:22:16.239-04:00 level=INFO source=memory.go:133 msg="offload to gpu" layers.requested=-1 layers.real=19 memory.available="21.7 GiB" memory.required.full="2.6 GiB" memory.required.partial="2.6 GiB" memory.required.kv="36.0 MiB" memory.weights.total="1.6 GiB" memory.weights.repeating="1.0 GiB" memory.weights.nonrepeating="531.5 MiB" memory.graph.full="504.2 MiB" memory.graph.partial="918.6 MiB" time=2024-06-06T11:22:16.239-04:00 level=INFO source=memory.go:133 msg="offload to gpu" layers.requested=-1 layers.real=19 memory.available="21.7 GiB" memory.required.full="2.6 GiB" memory.required.partial="2.6 GiB" memory.required.kv="36.0 MiB" memory.weights.total="1.6 GiB" memory.weights.repeating="1.0 GiB" memory.weights.nonrepeating="531.5 MiB" memory.graph.full="504.2 MiB" memory.graph.partial="918.6 MiB" time=2024-06-06T11:22:16.240-04:00 level=INFO source=server.go:341 msg="starting llama server" cmd="/tmp/ollama3794080172/runners/cuda_v11/ollama_llama_server --model /home/harper/.ollama/models/blobs/sha256-c1864a5eb19305c40519da12cc543519e48a0697ecd30e15d5ac228644957d12 --ctx-size 2048 --batch-size 512 --embedding --log-disable --n-gpu-layers 19 --parallel 1 --port 42781" time=2024-06-06T11:22:16.240-04:00 level=INFO source=sched.go:338 msg="loaded runners" count=1 time=2024-06-06T11:22:16.240-04:00 level=INFO source=server.go:529 msg="waiting for llama runner to start responding" time=2024-06-06T11:22:16.241-04:00 level=INFO source=server.go:567 msg="waiting for server to become available" status="llm server error" INFO [main] build info | build=1 commit="5921b8f" tid="281473278327040" timestamp=1717687336 INFO [main] system info | n_threads=8 n_threads_batch=-1 system_info="AVX = 0 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 0 | NEON = 1 | SVE = 0 | ARM_FMA = 1 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 0 | SSSE3 = 0 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="281473278327040" timestamp=1717687336 total_threads=8 INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="7" port="42781" tid="281473278327040" timestamp=1717687336 llama_model_loader: loaded meta data with 21 key-value pairs and 164 tensors from /home/harper/.ollama/models/blobs/sha256-c1864a5eb19305c40519da12cc543519e48a0697ecd30e15d5ac228644957d12 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = gemma llama_model_loader: - kv 1: general.name str = gemma-2b-it llama_model_loader: - kv 2: gemma.context_length u32 = 8192 llama_model_loader: - kv 3: gemma.block_count u32 = 18 llama_model_loader: - kv 4: gemma.embedding_length u32 = 2048 llama_model_loader: - kv 5: gemma.feed_forward_length u32 = 16384 llama_model_loader: - kv 6: gemma.attention.head_count u32 = 8 llama_model_loader: - kv 7: gemma.attention.head_count_kv u32 = 1 llama_model_loader: - kv 8: gemma.attention.key_length u32 = 256 llama_model_loader: - kv 9: gemma.attention.value_length u32 = 256 llama_model_loader: - kv 10: gemma.attention.layer_norm_rms_epsilon f32 = 0.000001 llama_model_loader: - kv 11: tokenizer.ggml.model str = llama llama_model_loader: - kv 12: tokenizer.ggml.bos_token_id u32 = 2 llama_model_loader: - kv 13: tokenizer.ggml.eos_token_id u32 = 1 llama_model_loader: - kv 14: tokenizer.ggml.padding_token_id u32 = 0 llama_model_loader: - kv 15: tokenizer.ggml.unknown_token_id u32 = 3 llama_model_loader: - kv 16: tokenizer.ggml.tokens arr[str,256128] = ["<pad>", "<eos>", "<bos>", "<unk>", ... time=2024-06-06T11:22:16.493-04:00 level=INFO source=server.go:567 msg="waiting for server to become available" status="llm server loading model" llama_model_loader: - kv 17: tokenizer.ggml.scores arr[f32,256128] = [0.000000, 0.000000, 0.000000, 0.0000... llama_model_loader: - kv 18: tokenizer.ggml.token_type arr[i32,256128] = [3, 3, 3, 2, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 19: general.quantization_version u32 = 2 llama_model_loader: - kv 20: general.file_type u32 = 2 llama_model_loader: - type f32: 37 tensors llama_model_loader: - type q4_0: 126 tensors llama_model_loader: - type q8_0: 1 tensors llm_load_vocab: special tokens cache size = 388 llm_load_vocab: token to piece cache size = 3.2028 MB llm_load_print_meta: format = GGUF V3 (latest) llm_load_print_meta: arch = gemma llm_load_print_meta: vocab type = SPM llm_load_print_meta: n_vocab = 256128 llm_load_print_meta: n_merges = 0 llm_load_print_meta: n_ctx_train = 8192 llm_load_print_meta: n_embd = 2048 llm_load_print_meta: n_head = 8 llm_load_print_meta: n_head_kv = 1 llm_load_print_meta: n_layer = 18 llm_load_print_meta: n_rot = 256 llm_load_print_meta: n_embd_head_k = 256 llm_load_print_meta: n_embd_head_v = 256 llm_load_print_meta: n_gqa = 8 llm_load_print_meta: n_embd_k_gqa = 256 llm_load_print_meta: n_embd_v_gqa = 256 llm_load_print_meta: f_norm_eps = 0.0e+00 llm_load_print_meta: f_norm_rms_eps = 1.0e-06 llm_load_print_meta: f_clamp_kqv = 0.0e+00 llm_load_print_meta: f_max_alibi_bias = 0.0e+00 llm_load_print_meta: f_logit_scale = 0.0e+00 llm_load_print_meta: n_ff = 16384 llm_load_print_meta: n_expert = 0 llm_load_print_meta: n_expert_used = 0 llm_load_print_meta: causal attn = 1 llm_load_print_meta: pooling type = 0 llm_load_print_meta: rope type = 2 llm_load_print_meta: rope scaling = linear llm_load_print_meta: freq_base_train = 10000.0 llm_load_print_meta: freq_scale_train = 1 llm_load_print_meta: n_yarn_orig_ctx = 8192 llm_load_print_meta: rope_finetuned = unknown llm_load_print_meta: ssm_d_conv = 0 llm_load_print_meta: ssm_d_inner = 0 llm_load_print_meta: ssm_d_state = 0 llm_load_print_meta: ssm_dt_rank = 0 llm_load_print_meta: model type = 2B llm_load_print_meta: model ftype = Q4_0 llm_load_print_meta: model params = 2.51 B llm_load_print_meta: model size = 1.56 GiB (5.34 BPW) llm_load_print_meta: general.name = gemma-2b-it llm_load_print_meta: BOS token = 2 '<bos>' llm_load_print_meta: EOS token = 1 '<eos>' llm_load_print_meta: UNK token = 3 '<unk>' llm_load_print_meta: PAD token = 0 '<pad>' llm_load_print_meta: LF token = 227 '<0x0A>' llm_load_print_meta: EOT token = 107 '<end_of_turn>' ggml_cuda_init: GGML_CUDA_FORCE_MMQ: yes ggml_cuda_init: CUDA_USE_TENSOR_CORES: no ggml_cuda_init: found 1 CUDA devices: Device 0: Orin, compute capability 8.7, VMM: yes llm_load_tensors: ggml ctx size = 0.17 MiB llm_load_tensors: offloading 18 repeating layers to GPU llm_load_tensors: offloading non-repeating layers to GPU llm_load_tensors: offloaded 19/19 layers to GPU llm_load_tensors: CPU buffer size = 531.52 MiB llm_load_tensors: CUDA0 buffer size = 1594.93 MiB llama_new_context_with_model: n_ctx = 2048 llama_new_context_with_model: n_batch = 512 llama_new_context_with_model: n_ubatch = 512 llama_new_context_with_model: flash_attn = 0 llama_new_context_with_model: freq_base = 10000.0 llama_new_context_with_model: freq_scale = 1 llama_kv_cache_init: CUDA0 KV buffer size = 36.00 MiB llama_new_context_with_model: KV self size = 36.00 MiB, K (f16): 18.00 MiB, V (f16): 18.00 MiB llama_new_context_with_model: CUDA_Host output buffer size = 0.98 MiB llama_new_context_with_model: CUDA0 compute buffer size = 504.25 MiB llama_new_context_with_model: CUDA_Host compute buffer size = 8.01 MiB llama_new_context_with_model: graph nodes = 601 llama_new_context_with_model: graph splits = 2 time=2024-06-06T11:29:16.920-04:00 level=ERROR source=sched.go:344 msg="error loading llama server" error="timed out waiting for llama runner to start - progress 1.00 - " [GIN] 2024/06/06 - 11:29:16 | 500 | 7m2s | 127.0.0.1 | POST "/api/chat" time=2024-06-06T11:29:22.037-04:00 level=WARN source=sched.go:512 msg="gpu VRAM usage didn't recover within timeout" seconds=5.117139389 time=2024-06-06T11:29:22.288-04:00 level=WARN source=sched.go:512 msg="gpu VRAM usage didn't recover within timeout" seconds=5.367497274 time=2024-06-06T11:29:22.537-04:00 level=WARN source=sched.go:512 msg="gpu VRAM usage didn't recover within timeout" seconds=5.616874999 ``` If it helps, this is running on a [Jetson AGX Orin with 32GB of memory](https://www.seeedstudio.com/AGX-Orin-32GB-H01-Kit-p-5569.html) ### OS Linux ### GPU Nvidia ### CPU Other: 8-core NVIDIA Arm® Cortex A78AE v8.2 64-bit CPU 2MB L2 + 4MB L3 ### Ollama version 0.1.41
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4861/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4861/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7141
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7141/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7141/comments
https://api.github.com/repos/ollama/ollama/issues/7141/events
https://github.com/ollama/ollama/pull/7141
2,573,860,762
PR_kwDOJ0Z1Ps59-4qQ
7,141
Fix build leakages
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-10-08T18:00:51
2024-10-08T20:05:03
2024-10-08T20:05:00
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7141", "html_url": "https://github.com/ollama/ollama/pull/7141", "diff_url": "https://github.com/ollama/ollama/pull/7141.diff", "patch_url": "https://github.com/ollama/ollama/pull/7141.patch", "merged_at": "2024-10-08T20:05:00" }
The recent change to applying patches leaves the submodule dirty based on "new commits" being present. This ensures we clean up so the tree no longer reports dirty after a `go generate ./...` run. The Makefile was being a bit too aggressive in cleaning things up and would result in deleting the placeholder files which someone might accidentally commit.
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7141/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7141/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6269
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6269/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6269/comments
https://api.github.com/repos/ollama/ollama/issues/6269/events
https://github.com/ollama/ollama/issues/6269
2,456,908,603
I_kwDOJ0Z1Ps6ScXM7
6,269
Please add LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct model
{ "login": "xest", "id": 4961215, "node_id": "MDQ6VXNlcjQ5NjEyMTU=", "avatar_url": "https://avatars.githubusercontent.com/u/4961215?v=4", "gravatar_id": "", "url": "https://api.github.com/users/xest", "html_url": "https://github.com/xest", "followers_url": "https://api.github.com/users/xest/followers", "following_url": "https://api.github.com/users/xest/following{/other_user}", "gists_url": "https://api.github.com/users/xest/gists{/gist_id}", "starred_url": "https://api.github.com/users/xest/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/xest/subscriptions", "organizations_url": "https://api.github.com/users/xest/orgs", "repos_url": "https://api.github.com/users/xest/repos", "events_url": "https://api.github.com/users/xest/events{/privacy}", "received_events_url": "https://api.github.com/users/xest/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
9
2024-08-09T01:07:48
2024-12-10T08:05:04
2024-12-10T08:05:04
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
* huggingface: [LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct](https://huggingface.co/LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct)
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6269/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6269/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3931
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3931/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3931/comments
https://api.github.com/repos/ollama/ollama/issues/3931/events
https://github.com/ollama/ollama/issues/3931
2,264,958,659
I_kwDOJ0Z1Ps6HAIbD
3,931
Digest mismatch, file must be downloaded again
{ "login": "tttt-0814", "id": 39620928, "node_id": "MDQ6VXNlcjM5NjIwOTI4", "avatar_url": "https://avatars.githubusercontent.com/u/39620928?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tttt-0814", "html_url": "https://github.com/tttt-0814", "followers_url": "https://api.github.com/users/tttt-0814/followers", "following_url": "https://api.github.com/users/tttt-0814/following{/other_user}", "gists_url": "https://api.github.com/users/tttt-0814/gists{/gist_id}", "starred_url": "https://api.github.com/users/tttt-0814/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tttt-0814/subscriptions", "organizations_url": "https://api.github.com/users/tttt-0814/orgs", "repos_url": "https://api.github.com/users/tttt-0814/repos", "events_url": "https://api.github.com/users/tttt-0814/events{/privacy}", "received_events_url": "https://api.github.com/users/tttt-0814/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
9
2024-04-26T04:52:11
2025-01-30T02:39:33
2024-05-09T21:08:04
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I tried to pull nomic-embed-text, but got an error below. I also tried to pull another models, but got the same error. $ ollama pull nomic-embed-text pulling manifest pulling 970aa74c0a90... 100% ▕███████████████████████████████████████████████████████████████████████████████████████████▏ 274 MB pulling c71d239df917... 100% ▕███████████████████████████████████████████████████████████████████████████████████████████▏ 11 KB pulling ce4a164fc046... 100% ▕███████████████████████████████████████████████████████████████████████████████████████████▏ 17 B pulling 31df23ea7daa... 100% ▕███████████████████████████████████████████████████████████████████████████████████████████▏ 420 B verifying sha256 digest Error: digest mismatch, file must be downloaded again: want sha256:970aa74c0a90ef7482477cf803618e776e173c007bf957f635f1015bfcfef0e6, got sha256:bea7e365d4085c35b0cfc78f9285682c6c7df7a15ac58f1905542659871024fd ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.1.32
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3931/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3931/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4879
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4879/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4879/comments
https://api.github.com/repos/ollama/ollama/issues/4879/events
https://github.com/ollama/ollama/pull/4879
2,338,955,742
PR_kwDOJ0Z1Ps5xuCen
4,879
API app/browser access
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-06-06T18:56:08
2024-06-06T22:19:04
2024-06-06T22:19:03
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4879", "html_url": "https://github.com/ollama/ollama/pull/4879", "diff_url": "https://github.com/ollama/ollama/pull/4879.diff", "patch_url": "https://github.com/ollama/ollama/pull/4879.patch", "merged_at": "2024-06-06T22:19:03" }
Fixes #4791 Fixes #3799 Fixes #4388
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4879/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 1, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4879/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1764
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1764/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1764/comments
https://api.github.com/repos/ollama/ollama/issues/1764/events
https://github.com/ollama/ollama/pull/1764
2,063,147,841
PR_kwDOJ0Z1Ps5jFzXu
1,764
keyboard shortcut help
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-01-03T01:59:41
2024-01-03T02:04:13
2024-01-03T02:04:13
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1764", "html_url": "https://github.com/ollama/ollama/pull/1764", "diff_url": "https://github.com/ollama/ollama/pull/1764.diff", "patch_url": "https://github.com/ollama/ollama/pull/1764.patch", "merged_at": "2024-01-03T02:04:13" }
This change adds some help in the REPL for using the keyboard shortcut commands.
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1764/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1764/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4200
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4200/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4200/comments
https://api.github.com/repos/ollama/ollama/issues/4200/events
https://github.com/ollama/ollama/issues/4200
2,280,915,245
I_kwDOJ0Z1Ps6H9AEt
4,200
http://localhost:11434/api endpoint giving 404 error
{ "login": "ritesh7911", "id": 64787172, "node_id": "MDQ6VXNlcjY0Nzg3MTcy", "avatar_url": "https://avatars.githubusercontent.com/u/64787172?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ritesh7911", "html_url": "https://github.com/ritesh7911", "followers_url": "https://api.github.com/users/ritesh7911/followers", "following_url": "https://api.github.com/users/ritesh7911/following{/other_user}", "gists_url": "https://api.github.com/users/ritesh7911/gists{/gist_id}", "starred_url": "https://api.github.com/users/ritesh7911/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ritesh7911/subscriptions", "organizations_url": "https://api.github.com/users/ritesh7911/orgs", "repos_url": "https://api.github.com/users/ritesh7911/repos", "events_url": "https://api.github.com/users/ritesh7911/events{/privacy}", "received_events_url": "https://api.github.com/users/ritesh7911/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
3
2024-05-06T13:42:10
2024-05-08T20:27:28
2024-05-08T20:27:24
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I am using latest version of windows . As per readme file when I am hitting http://localhost:11434 i am getting "ollama is running" but "http://localhost:11434/api" is giving error 404 ### OS Windows ### GPU AMD ### CPU Intel ### Ollama version 0.1.33
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4200/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4200/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3554
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3554/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3554/comments
https://api.github.com/repos/ollama/ollama/issues/3554/events
https://github.com/ollama/ollama/issues/3554
2,233,075,542
I_kwDOJ0Z1Ps6FGgdW
3,554
Potential problems with the `llm/ext_server/server.cpp` not accepting `--ubatch-size ` option
{ "login": "jukofyork", "id": 69222624, "node_id": "MDQ6VXNlcjY5MjIyNjI0", "avatar_url": "https://avatars.githubusercontent.com/u/69222624?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jukofyork", "html_url": "https://github.com/jukofyork", "followers_url": "https://api.github.com/users/jukofyork/followers", "following_url": "https://api.github.com/users/jukofyork/following{/other_user}", "gists_url": "https://api.github.com/users/jukofyork/gists{/gist_id}", "starred_url": "https://api.github.com/users/jukofyork/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jukofyork/subscriptions", "organizations_url": "https://api.github.com/users/jukofyork/orgs", "repos_url": "https://api.github.com/users/jukofyork/repos", "events_url": "https://api.github.com/users/jukofyork/events{/privacy}", "received_events_url": "https://api.github.com/users/jukofyork/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
{ "login": "jessegross", "id": 6468499, "node_id": "MDQ6VXNlcjY0Njg0OTk=", "avatar_url": "https://avatars.githubusercontent.com/u/6468499?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jessegross", "html_url": "https://github.com/jessegross", "followers_url": "https://api.github.com/users/jessegross/followers", "following_url": "https://api.github.com/users/jessegross/following{/other_user}", "gists_url": "https://api.github.com/users/jessegross/gists{/gist_id}", "starred_url": "https://api.github.com/users/jessegross/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jessegross/subscriptions", "organizations_url": "https://api.github.com/users/jessegross/orgs", "repos_url": "https://api.github.com/users/jessegross/repos", "events_url": "https://api.github.com/users/jessegross/events{/privacy}", "received_events_url": "https://api.github.com/users/jessegross/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "jessegross", "id": 6468499, "node_id": "MDQ6VXNlcjY0Njg0OTk=", "avatar_url": "https://avatars.githubusercontent.com/u/6468499?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jessegross", "html_url": "https://github.com/jessegross", "followers_url": "https://api.github.com/users/jessegross/followers", "following_url": "https://api.github.com/users/jessegross/following{/other_user}", "gists_url": "https://api.github.com/users/jessegross/gists{/gist_id}", "starred_url": "https://api.github.com/users/jessegross/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jessegross/subscriptions", "organizations_url": "https://api.github.com/users/jessegross/orgs", "repos_url": "https://api.github.com/users/jessegross/repos", "events_url": "https://api.github.com/users/jessegross/events{/privacy}", "received_events_url": "https://api.github.com/users/jessegross/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
1
2024-04-09T10:01:50
2024-11-23T20:17:29
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Not sure what to list this issue under (it's a potential bug I think). Recently `llama.cpp` has added an option called `--ubatch-size ` and appears to have changed the default value (and possibly meaning of) the old `--batch-size ` option: https://github.com/ggerganov/llama.cpp/pull/6017 The PR says: > By default `n_batch` is 4096, `n_ubatch` is 512 but I think it is actually 2048 and 512 now. This better explains the difference between the options: https://github.com/ggerganov/llama.cpp/discussions/6328 There are several potential problems and implications of this change for Ollama: - It looks like `llm/ext_server/server.cpp` is using some old code that handles the `n_batch` in the old way, but this in turn seems to be importing headers, etc from a newer version of `lamma.cpp` (b2581), who's `examples/server.cpp` code does appear to have the new `n_batch` / `n_ubatch` logic in it. - It looks like `api/types.go` is setting the `n_batch` value back down to 512, when the new default is supposed to be 2048. - Since Ollama is using the `--embedding` option there are some potential problems mentioned here: https://github.com/ggerganov/llama.cpp/issues/6263 - Not being able to set the `n_ubatch` value is causing regressions for people using multiple GPUs (https://github.com/ggerganov/llama.cpp/issues/6075 & https://github.com/ggerganov/llama.cpp/issues/6476). --- https://github.com/ollama/ollama/blob/main/llm/ext_server/server.cpp: ``` printf(" --pooling {none,mean,cls}\n"); printf(" pooling type for embeddings, use model default if unspecified\n"); printf(" -b N, --batch-size N batch size for prompt processing (default: %d)\n", params.n_batch); printf(" --memory-f32 use f32 instead of f16 for memory key+value (default: disabled)\n"); ``` vs: https://github.com/ggerganov/llama.cpp/blob/37e7854c104301c5b5323ccc40e07699f3a62c3e/examples/server/server.cpp: ``` printf(" --pooling {none,mean,cls} pooling type for embeddings, use model default if unspecified\n"); printf(" -dt N, --defrag-thold N\n"); printf(" KV cache defragmentation threshold (default: %.1f, < 0 - disabled)\n", params.defrag_thold); printf(" -b N, --batch-size N logical maximum batch size (default: %d)\n", params.n_batch); printf(" -ub N, --ubatch-size N physical maximum batch size (default: %d)\n", params.n_ubatch); printf(" --memory-f32 use f32 instead of f16 for memory key+value (default: disabled)\n"); ``` --- ``` - `-b N`, `--batch-size N`: Set the batch size for prompt processing. Default: `2048` - `-ub N`, `--ubatch-size N`: Physical maximum batch size. Default: `512` ```
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3554/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3554/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/4742
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4742/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4742/comments
https://api.github.com/repos/ollama/ollama/issues/4742/events
https://github.com/ollama/ollama/issues/4742
2,327,001,333
I_kwDOJ0Z1Ps6Kszj1
4,742
VRAM allocation error when loading different models with different OLLAMA_VRAM_MAX configurations
{ "login": "hamkido", "id": 43724352, "node_id": "MDQ6VXNlcjQzNzI0MzUy", "avatar_url": "https://avatars.githubusercontent.com/u/43724352?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hamkido", "html_url": "https://github.com/hamkido", "followers_url": "https://api.github.com/users/hamkido/followers", "following_url": "https://api.github.com/users/hamkido/following{/other_user}", "gists_url": "https://api.github.com/users/hamkido/gists{/gist_id}", "starred_url": "https://api.github.com/users/hamkido/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hamkido/subscriptions", "organizations_url": "https://api.github.com/users/hamkido/orgs", "repos_url": "https://api.github.com/users/hamkido/repos", "events_url": "https://api.github.com/users/hamkido/events{/privacy}", "received_events_url": "https://api.github.com/users/hamkido/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
3
2024-05-31T05:38:13
2024-06-05T06:34:19
2024-05-31T06:31:53
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I have two amd 7900xtx 24g gpu. When using ollama, I encounter different memory allocation errors and exit errors. 1. No OLLAMA_VRAM_MAX configuration The large model deepseek-llm:67b-chat can be loaded correctly But if you call something bigger, such as qwen:72b and command-r-plus, the display memory allocation will report an error and exit. 2. Use OLLAMA_VRAM_MAX configuration Models larger than 67b-q4 load correctly, such as qwen:72b and command-r-plus. However, smaller models cannot be loaded correctly, and memory allocation errors will be reported and exited, such as deepseek-llm:67b-chat. There might be some errors in the vram configuration. ### OS Linux ### GPU AMD ### CPU AMD ### Ollama version 0.1.39
{ "login": "hamkido", "id": 43724352, "node_id": "MDQ6VXNlcjQzNzI0MzUy", "avatar_url": "https://avatars.githubusercontent.com/u/43724352?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hamkido", "html_url": "https://github.com/hamkido", "followers_url": "https://api.github.com/users/hamkido/followers", "following_url": "https://api.github.com/users/hamkido/following{/other_user}", "gists_url": "https://api.github.com/users/hamkido/gists{/gist_id}", "starred_url": "https://api.github.com/users/hamkido/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hamkido/subscriptions", "organizations_url": "https://api.github.com/users/hamkido/orgs", "repos_url": "https://api.github.com/users/hamkido/repos", "events_url": "https://api.github.com/users/hamkido/events{/privacy}", "received_events_url": "https://api.github.com/users/hamkido/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4742/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4742/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8526
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8526/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8526/comments
https://api.github.com/repos/ollama/ollama/issues/8526/events
https://github.com/ollama/ollama/issues/8526
2,803,177,332
I_kwDOJ0Z1Ps6nFRd0
8,526
how to get English output
{ "login": "jarkkop", "id": 5814285, "node_id": "MDQ6VXNlcjU4MTQyODU=", "avatar_url": "https://avatars.githubusercontent.com/u/5814285?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarkkop", "html_url": "https://github.com/jarkkop", "followers_url": "https://api.github.com/users/jarkkop/followers", "following_url": "https://api.github.com/users/jarkkop/following{/other_user}", "gists_url": "https://api.github.com/users/jarkkop/gists{/gist_id}", "starred_url": "https://api.github.com/users/jarkkop/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jarkkop/subscriptions", "organizations_url": "https://api.github.com/users/jarkkop/orgs", "repos_url": "https://api.github.com/users/jarkkop/repos", "events_url": "https://api.github.com/users/jarkkop/events{/privacy}", "received_events_url": "https://api.github.com/users/jarkkop/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
1
2025-01-22T01:51:56
2025-01-22T04:17:26
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? M:\AI\ollama>ollama run deepseek-r1:7b >>> list philosophers <think> </think> Here is a list of some of the most influential and notable philosophers throughout history, organized by era and region: ### Ancient Philosophy (c. 600–321 BCE) - **Thales of Miletus** (c. 624–548 BCE):被认为是第一个哲学家,提出“万物源于水”的学说。 - **Anaximander of Miletus** (c. 570–495 BCE): 提出“无限”概念,并认为万物起源于自然。 - **Anaximenes of Mileti** (c. 510–441 BCE): 认为万物来源于某种原始物质,如“air”(空气)。 ### OS Microsoft Windows [Version 10.0.22635.4800] ### GPU NVIDIA GeForce GTX 1660 SUPER ### CPU Processor AMD Ryzen 7 3700X 8-Core Processor, 3600 Mhz, 8 Core(s), 16 Logical Processor(s) ### Ollama version ollama version is 0.5.7
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8526/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8526/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/672
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/672/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/672/comments
https://api.github.com/repos/ollama/ollama/issues/672/events
https://github.com/ollama/ollama/pull/672
1,922,337,068
PR_kwDOJ0Z1Ps5bse6w
672
Relay default values to llama runner
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
3
2023-10-02T17:32:53
2023-10-02T18:53:17
2023-10-02T18:53:16
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/672", "html_url": "https://github.com/ollama/ollama/pull/672", "diff_url": "https://github.com/ollama/ollama/pull/672.diff", "patch_url": "https://github.com/ollama/ollama/pull/672.patch", "merged_at": "2023-10-02T18:53:16" }
Thanks to @hallh for #663. This change cherry-picks that PR, relays all our defaults, and does some re-organizing of the code to make it easier to read.
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/672/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/672/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8113
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8113/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8113/comments
https://api.github.com/repos/ollama/ollama/issues/8113/events
https://github.com/ollama/ollama/pull/8113
2,741,618,542
PR_kwDOJ0Z1Ps6FULcl
8,113
llama: add qwen2vl support
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
0
2024-12-16T07:59:10
2025-01-15T11:14:59
null
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
true
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8113", "html_url": "https://github.com/ollama/ollama/pull/8113", "diff_url": "https://github.com/ollama/ollama/pull/8113.diff", "patch_url": "https://github.com/ollama/ollama/pull/8113.patch", "merged_at": null }
Still missing: add 4 positions per embedding when creating a batch
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8113/reactions", "total_count": 5, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 1, "heart": 0, "rocket": 0, "eyes": 4 }
https://api.github.com/repos/ollama/ollama/issues/8113/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/685
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/685/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/685/comments
https://api.github.com/repos/ollama/ollama/issues/685/events
https://github.com/ollama/ollama/issues/685
1,923,133,063
I_kwDOJ0Z1Ps5yoK6H
685
Question: where are all the `Modelfile`s?
{ "login": "jamesbraza", "id": 8990777, "node_id": "MDQ6VXNlcjg5OTA3Nzc=", "avatar_url": "https://avatars.githubusercontent.com/u/8990777?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jamesbraza", "html_url": "https://github.com/jamesbraza", "followers_url": "https://api.github.com/users/jamesbraza/followers", "following_url": "https://api.github.com/users/jamesbraza/following{/other_user}", "gists_url": "https://api.github.com/users/jamesbraza/gists{/gist_id}", "starred_url": "https://api.github.com/users/jamesbraza/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jamesbraza/subscriptions", "organizations_url": "https://api.github.com/users/jamesbraza/orgs", "repos_url": "https://api.github.com/users/jamesbraza/repos", "events_url": "https://api.github.com/users/jamesbraza/events{/privacy}", "received_events_url": "https://api.github.com/users/jamesbraza/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
8
2023-10-03T01:50:20
2023-10-06T15:15:28
2023-10-04T02:40:43
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
https://ollama.ai/library has a lot of models. I would like to add a new model, and want to make sure it uses the GPU. So I am looking to refer to `Modelfile`s for models featured on https://ollama.ai/library. Where is the source `Modelfile`s for the current "built in" models?
{ "login": "technovangelist", "id": 633681, "node_id": "MDQ6VXNlcjYzMzY4MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4", "gravatar_id": "", "url": "https://api.github.com/users/technovangelist", "html_url": "https://github.com/technovangelist", "followers_url": "https://api.github.com/users/technovangelist/followers", "following_url": "https://api.github.com/users/technovangelist/following{/other_user}", "gists_url": "https://api.github.com/users/technovangelist/gists{/gist_id}", "starred_url": "https://api.github.com/users/technovangelist/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/technovangelist/subscriptions", "organizations_url": "https://api.github.com/users/technovangelist/orgs", "repos_url": "https://api.github.com/users/technovangelist/repos", "events_url": "https://api.github.com/users/technovangelist/events{/privacy}", "received_events_url": "https://api.github.com/users/technovangelist/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/685/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/685/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2246
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2246/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2246/comments
https://api.github.com/repos/ollama/ollama/issues/2246/events
https://github.com/ollama/ollama/pull/2246
2,104,401,475
PR_kwDOJ0Z1Ps5lRIEy
2,246
Don't disable GPUs on arm without AVX
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-01-28T23:38:00
2024-01-29T00:26:58
2024-01-29T00:26:55
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2246", "html_url": "https://github.com/ollama/ollama/pull/2246", "diff_url": "https://github.com/ollama/ollama/pull/2246.diff", "patch_url": "https://github.com/ollama/ollama/pull/2246.patch", "merged_at": "2024-01-29T00:26:55" }
AVX is an x86 feature, so ARM should be excluded from the check. Related to #1979
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2246/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2246/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2489
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2489/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2489/comments
https://api.github.com/repos/ollama/ollama/issues/2489/events
https://github.com/ollama/ollama/issues/2489
2,133,991,624
I_kwDOJ0Z1Ps5_MiDI
2,489
what is smallest model that know about comp system administration, network admin, etc?
{ "login": "zinwelzl", "id": 113045180, "node_id": "U_kgDOBrzuvA", "avatar_url": "https://avatars.githubusercontent.com/u/113045180?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zinwelzl", "html_url": "https://github.com/zinwelzl", "followers_url": "https://api.github.com/users/zinwelzl/followers", "following_url": "https://api.github.com/users/zinwelzl/following{/other_user}", "gists_url": "https://api.github.com/users/zinwelzl/gists{/gist_id}", "starred_url": "https://api.github.com/users/zinwelzl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zinwelzl/subscriptions", "organizations_url": "https://api.github.com/users/zinwelzl/orgs", "repos_url": "https://api.github.com/users/zinwelzl/repos", "events_url": "https://api.github.com/users/zinwelzl/events{/privacy}", "received_events_url": "https://api.github.com/users/zinwelzl/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
null
[]
null
1
2024-02-14T10:15:25
2024-03-14T00:01:41
2024-03-14T00:01:40
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I run ollama locally and need some small model for help with system administration, network administration, etc? I try few small but are really bad.
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2489/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2489/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1872
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1872/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1872/comments
https://api.github.com/repos/ollama/ollama/issues/1872/events
https://github.com/ollama/ollama/issues/1872
2,072,777,144
I_kwDOJ0Z1Ps57jBG4
1,872
Error when install on Ubuntu 22.04
{ "login": "dekogroup", "id": 126862835, "node_id": "U_kgDOB4_F8w", "avatar_url": "https://avatars.githubusercontent.com/u/126862835?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dekogroup", "html_url": "https://github.com/dekogroup", "followers_url": "https://api.github.com/users/dekogroup/followers", "following_url": "https://api.github.com/users/dekogroup/following{/other_user}", "gists_url": "https://api.github.com/users/dekogroup/gists{/gist_id}", "starred_url": "https://api.github.com/users/dekogroup/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dekogroup/subscriptions", "organizations_url": "https://api.github.com/users/dekogroup/orgs", "repos_url": "https://api.github.com/users/dekogroup/repos", "events_url": "https://api.github.com/users/dekogroup/events{/privacy}", "received_events_url": "https://api.github.com/users/dekogroup/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5755339642, "node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg", "url": "https://api.github.com/repos/ollama/ollama/labels/linux", "name": "linux", "color": "516E70", "default": false, "description": "" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
5
2024-01-09T16:55:31
2024-03-13T00:13:27
2024-03-13T00:13:27
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
curl https://ollama.ai/install.sh | sh % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 8354 0 8354 0 0 16163 0 --:--:-- --:--:-- --:--:-- 16189 >>> Downloading ollama... ######################################################################## 100.0%##O=# # Warning: Failed to open the file /tmp/tmp.AO1TPHxNpB/ollama: No such file or Warning: directory 0.0%curl: (23) Failure writing output to destination
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1872/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1872/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6461
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6461/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6461/comments
https://api.github.com/repos/ollama/ollama/issues/6461/events
https://github.com/ollama/ollama/issues/6461
2,480,507,753
I_kwDOJ0Z1Ps6T2Ytp
6,461
"/clear" command is not clearing history
{ "login": "devstefancho", "id": 61320923, "node_id": "MDQ6VXNlcjYxMzIwOTIz", "avatar_url": "https://avatars.githubusercontent.com/u/61320923?v=4", "gravatar_id": "", "url": "https://api.github.com/users/devstefancho", "html_url": "https://github.com/devstefancho", "followers_url": "https://api.github.com/users/devstefancho/followers", "following_url": "https://api.github.com/users/devstefancho/following{/other_user}", "gists_url": "https://api.github.com/users/devstefancho/gists{/gist_id}", "starred_url": "https://api.github.com/users/devstefancho/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/devstefancho/subscriptions", "organizations_url": "https://api.github.com/users/devstefancho/orgs", "repos_url": "https://api.github.com/users/devstefancho/repos", "events_url": "https://api.github.com/users/devstefancho/events{/privacy}", "received_events_url": "https://api.github.com/users/devstefancho/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-08-22T11:07:39
2024-08-22T17:00:16
2024-08-22T17:00:16
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? `Ctrl` + `l` is clearing history but `/clear` command is not clearing history https://github.com/user-attachments/assets/511ee922-9252-41d4-8b5f-ac324a75aaf1 ### OS macOS ### GPU Apple ### CPU Apple ### Ollama version 0.3.6
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6461/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6461/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3473
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3473/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3473/comments
https://api.github.com/repos/ollama/ollama/issues/3473/events
https://github.com/ollama/ollama/pull/3473
2,222,435,172
PR_kwDOJ0Z1Ps5rimQ_
3,473
Add BrainSoup to compatible clients list
{ "login": "Nurgo", "id": 11637957, "node_id": "MDQ6VXNlcjExNjM3OTU3", "avatar_url": "https://avatars.githubusercontent.com/u/11637957?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Nurgo", "html_url": "https://github.com/Nurgo", "followers_url": "https://api.github.com/users/Nurgo/followers", "following_url": "https://api.github.com/users/Nurgo/following{/other_user}", "gists_url": "https://api.github.com/users/Nurgo/gists{/gist_id}", "starred_url": "https://api.github.com/users/Nurgo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Nurgo/subscriptions", "organizations_url": "https://api.github.com/users/Nurgo/orgs", "repos_url": "https://api.github.com/users/Nurgo/repos", "events_url": "https://api.github.com/users/Nurgo/events{/privacy}", "received_events_url": "https://api.github.com/users/Nurgo/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-04-03T09:41:58
2024-05-06T20:42:16
2024-05-06T20:42:16
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3473", "html_url": "https://github.com/ollama/ollama/pull/3473", "diff_url": "https://github.com/ollama/ollama/pull/3473.diff", "patch_url": "https://github.com/ollama/ollama/pull/3473.patch", "merged_at": "2024-05-06T20:42:16" }
Hi there, BrainSoup is a native multi-LLM client for Windows with advanced features such as local document indexing, RAG, multi-modality, multi-agent automation, code interpreter, sandboxed file system and the ability for agents to interact with the local system via customizable events and tools. More information can be found [here](https://www.nurgo-software.com/products/brainsoup). The latest version is know compatible with Ollama, making BrainSoup an excellent choice for users looking to leverage Ollama's LLMs to their full potential. I kindly ask for your review of this PR. Thanks for creating such an awesome platform!
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3473/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3473/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6431
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6431/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6431/comments
https://api.github.com/repos/ollama/ollama/issues/6431/events
https://github.com/ollama/ollama/issues/6431
2,474,397,214
I_kwDOJ0Z1Ps6TfE4e
6,431
GLM4 tools support
{ "login": "napa3um", "id": 665538, "node_id": "MDQ6VXNlcjY2NTUzOA==", "avatar_url": "https://avatars.githubusercontent.com/u/665538?v=4", "gravatar_id": "", "url": "https://api.github.com/users/napa3um", "html_url": "https://github.com/napa3um", "followers_url": "https://api.github.com/users/napa3um/followers", "following_url": "https://api.github.com/users/napa3um/following{/other_user}", "gists_url": "https://api.github.com/users/napa3um/gists{/gist_id}", "starred_url": "https://api.github.com/users/napa3um/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/napa3um/subscriptions", "organizations_url": "https://api.github.com/users/napa3um/orgs", "repos_url": "https://api.github.com/users/napa3um/repos", "events_url": "https://api.github.com/users/napa3um/events{/privacy}", "received_events_url": "https://api.github.com/users/napa3um/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-08-19T22:57:39
2024-08-19T22:57:39
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
GML4 support tools - https://github.com/THUDM/GLM-4/blob/main/finetune_demo/README_en.md How to fix the template in https://ollama.com/library/glm4 to make the ollama-tools mechanism work?
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6431/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6431/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/4823
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4823/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4823/comments
https://api.github.com/repos/ollama/ollama/issues/4823/events
https://github.com/ollama/ollama/issues/4823
2,334,769,831
I_kwDOJ0Z1Ps6LKcKn
4,823
I encountered this error when converting the Tongyi-Finance-14B-Chat-Int4-AWQ model
{ "login": "wangkai111111", "id": 74865581, "node_id": "MDQ6VXNlcjc0ODY1NTgx", "avatar_url": "https://avatars.githubusercontent.com/u/74865581?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wangkai111111", "html_url": "https://github.com/wangkai111111", "followers_url": "https://api.github.com/users/wangkai111111/followers", "following_url": "https://api.github.com/users/wangkai111111/following{/other_user}", "gists_url": "https://api.github.com/users/wangkai111111/gists{/gist_id}", "starred_url": "https://api.github.com/users/wangkai111111/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wangkai111111/subscriptions", "organizations_url": "https://api.github.com/users/wangkai111111/orgs", "repos_url": "https://api.github.com/users/wangkai111111/repos", "events_url": "https://api.github.com/users/wangkai111111/events{/privacy}", "received_events_url": "https://api.github.com/users/wangkai111111/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
2
2024-06-05T02:27:12
2024-06-05T20:39:40
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? `(.venv) [root@bastion ollama]# python llm/llama.cpp/convert-hf-to-gguf.py ./model --outtype f16 --outfile converted.bin INFO:hf-to-gguf:Loading model: model INFO:gguf.gguf_writer:gguf: This GGUF file is for Little Endian only INFO:hf-to-gguf:Set model parameters INFO:hf-to-gguf:Set model tokenizer Traceback (most recent call last): File "/root/model/ollama/llm/llama.cpp/convert-hf-to-gguf.py", line 2865, in <module> main() File "/root/model/ollama/llm/llama.cpp/convert-hf-to-gguf.py", line 2850, in main model_instance.set_vocab() File "/root/model/ollama/llm/llama.cpp/convert-hf-to-gguf.py", line 1592, in set_vocab self._set_vocab_qwen() File "/root/model/ollama/llm/llama.cpp/convert-hf-to-gguf.py", line 530, in _set_vocab_qwen assert len(merged) == 2 AssertionError ` ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.1.41
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4823/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4823/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/7973
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7973/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7973/comments
https://api.github.com/repos/ollama/ollama/issues/7973/events
https://github.com/ollama/ollama/pull/7973
2,723,621,546
PR_kwDOJ0Z1Ps6EW47n
7,973
Document that `--format` now supports passing JSON Schemas
{ "login": "joliss", "id": 524783, "node_id": "MDQ6VXNlcjUyNDc4Mw==", "avatar_url": "https://avatars.githubusercontent.com/u/524783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joliss", "html_url": "https://github.com/joliss", "followers_url": "https://api.github.com/users/joliss/followers", "following_url": "https://api.github.com/users/joliss/following{/other_user}", "gists_url": "https://api.github.com/users/joliss/gists{/gist_id}", "starred_url": "https://api.github.com/users/joliss/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/joliss/subscriptions", "organizations_url": "https://api.github.com/users/joliss/orgs", "repos_url": "https://api.github.com/users/joliss/repos", "events_url": "https://api.github.com/users/joliss/events{/privacy}", "received_events_url": "https://api.github.com/users/joliss/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
0
2024-12-06T17:53:16
2024-12-12T23:50:09
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7973", "html_url": "https://github.com/ollama/ollama/pull/7973", "diff_url": "https://github.com/ollama/ollama/pull/7973.diff", "patch_url": "https://github.com/ollama/ollama/pull/7973.patch", "merged_at": null }
JSON Schema support was added in #7900. -------- I removed `e.g.` because I don't believe it supports anything else, right? Let me know if that's wrong.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7973/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7973/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5466
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5466/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5466/comments
https://api.github.com/repos/ollama/ollama/issues/5466/events
https://github.com/ollama/ollama/pull/5466
2,389,318,735
PR_kwDOJ0Z1Ps50XSIm
5,466
Fix clip model loading with unicode paths
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-07-03T19:38:27
2024-07-05T15:17:01
2024-07-05T15:16:58
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5466", "html_url": "https://github.com/ollama/ollama/pull/5466", "diff_url": "https://github.com/ollama/ollama/pull/5466.diff", "patch_url": "https://github.com/ollama/ollama/pull/5466.patch", "merged_at": "2024-07-05T15:16:58" }
On windows, if the model dir contained unicode characters clip models would fail to load. This fixes the file name handling in clip.cpp to support utf16 on windows. Fixes #5329 #4365
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5466/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5466/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3581
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3581/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3581/comments
https://api.github.com/repos/ollama/ollama/issues/3581/events
https://github.com/ollama/ollama/issues/3581
2,236,296,576
I_kwDOJ0Z1Ps6FSy2A
3,581
MacOS Ollama not binding to 0.0.0.0
{ "login": "kellerkind84", "id": 2842721, "node_id": "MDQ6VXNlcjI4NDI3MjE=", "avatar_url": "https://avatars.githubusercontent.com/u/2842721?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kellerkind84", "html_url": "https://github.com/kellerkind84", "followers_url": "https://api.github.com/users/kellerkind84/followers", "following_url": "https://api.github.com/users/kellerkind84/following{/other_user}", "gists_url": "https://api.github.com/users/kellerkind84/gists{/gist_id}", "starred_url": "https://api.github.com/users/kellerkind84/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kellerkind84/subscriptions", "organizations_url": "https://api.github.com/users/kellerkind84/orgs", "repos_url": "https://api.github.com/users/kellerkind84/repos", "events_url": "https://api.github.com/users/kellerkind84/events{/privacy}", "received_events_url": "https://api.github.com/users/kellerkind84/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
32
2024-04-10T19:37:49
2025-01-22T12:51:14
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? So when set the OLLAMA_HOST to 0.0.0.0, I cannot access Ollama via the IP, but I can still access it via localhost. ### What did you expect to see? I expect it to be available under <myIP>:11434 ### Steps to reproduce _No response_ ### Are there any recent changes that introduced the issue? _No response_ ### OS macOS ### Architecture arm64 ### Platform _No response_ ### Ollama version 0.1.31 ### GPU Apple ### GPU info _No response_ ### CPU Apple ### Other software _No response_
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3581/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3581/timeline
null
reopened
false
https://api.github.com/repos/ollama/ollama/issues/2049
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2049/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2049/comments
https://api.github.com/repos/ollama/ollama/issues/2049/events
https://github.com/ollama/ollama/issues/2049
2,088,341,295
I_kwDOJ0Z1Ps58eY8v
2,049
Embedding API could return empty embedding while using completion API from LiteLLM
{ "login": "James4Ever0", "id": 103997068, "node_id": "U_kgDOBjLejA", "avatar_url": "https://avatars.githubusercontent.com/u/103997068?v=4", "gravatar_id": "", "url": "https://api.github.com/users/James4Ever0", "html_url": "https://github.com/James4Ever0", "followers_url": "https://api.github.com/users/James4Ever0/followers", "following_url": "https://api.github.com/users/James4Ever0/following{/other_user}", "gists_url": "https://api.github.com/users/James4Ever0/gists{/gist_id}", "starred_url": "https://api.github.com/users/James4Ever0/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/James4Ever0/subscriptions", "organizations_url": "https://api.github.com/users/James4Ever0/orgs", "repos_url": "https://api.github.com/users/James4Ever0/repos", "events_url": "https://api.github.com/users/James4Ever0/events{/privacy}", "received_events_url": "https://api.github.com/users/James4Ever0/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677485533, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgJX3Q", "url": "https://api.github.com/repos/ollama/ollama/labels/embeddings", "name": "embeddings", "color": "76BF9F", "default": false, "description": "Issues around embeddings" }, { "id": 7706482389, "node_id": "LA_kwDOJ0Z1Ps8AAAABy1eW1Q", "url": "https://api.github.com/repos/ollama/ollama/labels/api", "name": "api", "color": "bfdadc", "default": false, "description": "" } ]
open
false
null
[]
null
0
2024-01-18T13:54:52
2024-11-06T19:02:39
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
To reproduce: Launch a LiteLLM service: ```bash litellm --model ollama/openhermes2.5-mistral --drop_params ``` Call the service `/completion` API continuously first, meanwhile you call embedding API via Langchain, and hopefully during the very gap (very short) between each `/completion` call you get empty embedding from Langchain. To call the `/completion` API: ```python import os os.environ['OPENAI_API_KEY'] = 'any' os.environ['OPENAI_API_BASE'] = 'http://0.0.0.0:8000' from contextlib import contextmanager from langchain.llms import OpenAI import tiktoken def print_center(banner: str): print(banner.center(50, "=")) class LLM: """ A class for running a Language Model Chain. """ def __init__(self, prompt: str, temperature=0, gpt_4=False): """ Initializes the LLM class. Args: prompt (PromptTemplate): The prompt template to use. temperature (int): The temperature to use for the model. gpt_4 (bool): Whether to use GPT-4 or Text-Davinci-003. Side Effects: Sets the class attributes. """ self.prompt = prompt self.prompt_size = self.number_of_tokens(prompt) self.temperature = temperature self.gpt_4 = gpt_4 self.model_name = "gpt-4" if self.gpt_4 else "text-davinci-003" self.max_tokens = 4097 * 2 if self.gpt_4 else 4097 self.show_init_config() def show_init_config(self): print_center("init params") print(f"Model: {self.model_name}") print(f"Max Tokens: {self.max_tokens}") print(f"Prompt Size: {self.prompt_size}") print(f"Temperature: {self.temperature}") print_center("init config") print(self.prompt) def run(self, query): """ Runs the Language Model Chain. Args: code (str): The code to use for the chain. **kwargs (dict): Additional keyword arguments. Returns: str: The generated text. """ llm = OpenAI( temperature=self.temperature, max_tokens=-1, model_name=self.model_name, disallowed_special=(), # to suppress error when special tokens within the input text (encode special tokens as normal text) ) # chain = LLMChain(llm=llm, prompt=self.prompt) chunk_list = [] print_center("query") print(query) print_center("response") _input = "\n".join([self.prompt, query]) for chunk in llm.stream(input=_input): print(chunk, end="", flush=True) chunk_list.append(chunk) print() result = "".join(chunk_list) return result def number_of_tokens(self, text): """ Counts the number of tokens in a given text. Args: text (str): The text to count tokens for. Returns: int: The number of tokens in the text. """ encoding = tiktoken.encoding_for_model("gpt-4") return len(encoding.encode(text, disallowed_special=())) @contextmanager def llm_context(prompt: str, temperature=0, gpt_4=False): model = LLM(prompt, temperature=temperature, gpt_4=gpt_4) try: yield model finally: del model if __name__ == "__main__": while True: with llm_context("You are a helpful assistant.") as model: model.run("Write me a 100 words news.") ``` To call the embedding API: ```python from langchain.embeddings import OllamaEmbeddings ollama_emb = OllamaEmbeddings( model="openhermes2.5-mistral:latest", ) while True: embed = ollama_emb.embed_query("Hello world") print("empty embedding?", embed is None) ```
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2049/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2049/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/946
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/946/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/946/comments
https://api.github.com/repos/ollama/ollama/issues/946/events
https://github.com/ollama/ollama/issues/946
1,967,119,694
I_kwDOJ0Z1Ps51P91O
946
ollama show --modelfile gives incorrect FROM when multiple tags of base model are downloaded.
{ "login": "easp", "id": 414705, "node_id": "MDQ6VXNlcjQxNDcwNQ==", "avatar_url": "https://avatars.githubusercontent.com/u/414705?v=4", "gravatar_id": "", "url": "https://api.github.com/users/easp", "html_url": "https://github.com/easp", "followers_url": "https://api.github.com/users/easp/followers", "following_url": "https://api.github.com/users/easp/following{/other_user}", "gists_url": "https://api.github.com/users/easp/gists{/gist_id}", "starred_url": "https://api.github.com/users/easp/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/easp/subscriptions", "organizations_url": "https://api.github.com/users/easp/orgs", "repos_url": "https://api.github.com/users/easp/repos", "events_url": "https://api.github.com/users/easp/events{/privacy}", "received_events_url": "https://api.github.com/users/easp/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5667396210, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2acg", "url": "https://api.github.com/repos/ollama/ollama/labels/good%20first%20issue", "name": "good first issue", "color": "7057ff", "default": true, "description": "Good for newcomers" } ]
closed
false
null
[]
null
6
2023-10-29T19:13:09
2023-12-04T18:32:40
2023-12-04T18:32:40
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I've pulled two tags for codellama and created two new models, one based on each. ``` % ollama list NAME ID SIZE MODIFIED [...] codellama:13b 9f438cb9cd58 7.4 GB 27 hours ago codellama:13b-16k e86141f13814 7.4 GB 45 hours ago codellama:34b-16k be95e5b84e7f 16 GB 41 seconds ago codellama:34b-instruct-q3_K_M f534f618ea64 16 GB 27 hours ago [...] ``` Note the file sizes; it's clear that ollama isn't confused about which data file to use for which. However, if I `ollama show --modelfile`, the FROM just lists the base model name, not the tag. ``` % ollama show codellama:13b-16k --modelfile # Modelfile generated by "ollama show" # To build a new Modelfile based on this one, replace the FROM line with: # FROM codellama:13b-16k FROM library/codellama TEMPLATE """[INST] <<SYS>>{{ .System }}<</SYS>> {{ .Prompt }} [/INST] """ PARAMETER num_ctx 16384 PARAMETER rope_frequency_base 1e+06 PARAMETER stop "[INST]" PARAMETER stop "[/INST]" PARAMETER stop "<<SYS>>" PARAMETER stop "<</SYS>>" % ollama show codellama:34b-16k --modelfile # Modelfile generated by "ollama show" # To build a new Modelfile based on this one, replace the FROM line with: # FROM codellama:34b-16k FROM library/codellama TEMPLATE """[INST] <<SYS>>{{ .System }}<</SYS>> {{ .Prompt }} [/INST] """ PARAMETER num_ctx 16384 PARAMETER num_gqa 8 PARAMETER rope_frequency_base 1e+06 PARAMETER stop "[INST]" PARAMETER stop "[/INST]" PARAMETER stop "<<SYS>>" PARAMETER stop "<</SYS>>" ```
{ "login": "technovangelist", "id": 633681, "node_id": "MDQ6VXNlcjYzMzY4MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4", "gravatar_id": "", "url": "https://api.github.com/users/technovangelist", "html_url": "https://github.com/technovangelist", "followers_url": "https://api.github.com/users/technovangelist/followers", "following_url": "https://api.github.com/users/technovangelist/following{/other_user}", "gists_url": "https://api.github.com/users/technovangelist/gists{/gist_id}", "starred_url": "https://api.github.com/users/technovangelist/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/technovangelist/subscriptions", "organizations_url": "https://api.github.com/users/technovangelist/orgs", "repos_url": "https://api.github.com/users/technovangelist/repos", "events_url": "https://api.github.com/users/technovangelist/events{/privacy}", "received_events_url": "https://api.github.com/users/technovangelist/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/946/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/946/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2743
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2743/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2743/comments
https://api.github.com/repos/ollama/ollama/issues/2743/events
https://github.com/ollama/ollama/issues/2743
2,152,745,980
I_kwDOJ0Z1Ps6AUEv8
2,743
What is the different between "gemma-instruct", "gemma-text" and "gemma". Same to other models.
{ "login": "XinyueZ", "id": 7869833, "node_id": "MDQ6VXNlcjc4Njk4MzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7869833?v=4", "gravatar_id": "", "url": "https://api.github.com/users/XinyueZ", "html_url": "https://github.com/XinyueZ", "followers_url": "https://api.github.com/users/XinyueZ/followers", "following_url": "https://api.github.com/users/XinyueZ/following{/other_user}", "gists_url": "https://api.github.com/users/XinyueZ/gists{/gist_id}", "starred_url": "https://api.github.com/users/XinyueZ/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/XinyueZ/subscriptions", "organizations_url": "https://api.github.com/users/XinyueZ/orgs", "repos_url": "https://api.github.com/users/XinyueZ/repos", "events_url": "https://api.github.com/users/XinyueZ/events{/privacy}", "received_events_url": "https://api.github.com/users/XinyueZ/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-02-25T12:39:09
2024-02-26T11:01:38
2024-02-26T11:01:38
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
null
{ "login": "XinyueZ", "id": 7869833, "node_id": "MDQ6VXNlcjc4Njk4MzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7869833?v=4", "gravatar_id": "", "url": "https://api.github.com/users/XinyueZ", "html_url": "https://github.com/XinyueZ", "followers_url": "https://api.github.com/users/XinyueZ/followers", "following_url": "https://api.github.com/users/XinyueZ/following{/other_user}", "gists_url": "https://api.github.com/users/XinyueZ/gists{/gist_id}", "starred_url": "https://api.github.com/users/XinyueZ/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/XinyueZ/subscriptions", "organizations_url": "https://api.github.com/users/XinyueZ/orgs", "repos_url": "https://api.github.com/users/XinyueZ/repos", "events_url": "https://api.github.com/users/XinyueZ/events{/privacy}", "received_events_url": "https://api.github.com/users/XinyueZ/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2743/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2743/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5166
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5166/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5166/comments
https://api.github.com/repos/ollama/ollama/issues/5166/events
https://github.com/ollama/ollama/issues/5166
2,364,002,946
I_kwDOJ0Z1Ps6M59KC
5,166
In dockerGPU containers ollama still uses the CPU
{ "login": "Zxyy-mo", "id": 48347974, "node_id": "MDQ6VXNlcjQ4MzQ3OTc0", "avatar_url": "https://avatars.githubusercontent.com/u/48347974?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Zxyy-mo", "html_url": "https://github.com/Zxyy-mo", "followers_url": "https://api.github.com/users/Zxyy-mo/followers", "following_url": "https://api.github.com/users/Zxyy-mo/following{/other_user}", "gists_url": "https://api.github.com/users/Zxyy-mo/gists{/gist_id}", "starred_url": "https://api.github.com/users/Zxyy-mo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Zxyy-mo/subscriptions", "organizations_url": "https://api.github.com/users/Zxyy-mo/orgs", "repos_url": "https://api.github.com/users/Zxyy-mo/repos", "events_url": "https://api.github.com/users/Zxyy-mo/events{/privacy}", "received_events_url": "https://api.github.com/users/Zxyy-mo/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
6
2024-06-20T09:43:38
2024-06-21T15:38:53
2024-06-21T15:38:53
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? ## desc I implemented the deployment following the official Docker GPU container tutorial. And successfully got the graphics card information using nvidia-smi in the Docker container. I'm using a nvidia discrete graphics card 3090 ```info # ollama ps NAME ID SIZE PROCESSOR UNTIL # ollama ps NAME ID SIZE PROCESSOR UNTIL qwen2:7b e0d4e1163c58 4.7 GB 100% CPU 4 minutes from now # nvidia-smi Thu Jun 20 09:36:11 2024 +-----------------------------------------------------------------------------+ | NVIDIA-SMI 525.147.05 Driver Version: 525.147.05 CUDA Version: 12.0 | |-------------------------------+----------------------+----------------------+ | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |===============================+======================+======================| | 0 NVIDIA GeForce ... Off | 00000000:02:00.0 On | N/A | | 0% 38C P8 16W / 350W | 114MiB / 24576MiB | 0% Default | | | | N/A | +-------------------------------+----------------------+----------------------+ +-----------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=============================================================================| +-----------------------------------------------------------------------------+ # ollama -v ollama version is 0.1.44 # ollama -v ollama version is 0.1.44 # ollama ps NAME ID SIZE PROCESSOR UNTIL qwen2:7b e0d4e1163c58 4.7 GB 100% CPU 37 seconds from now # ``` ### OS Linux, Docker ### GPU Nvidia ### CPU Intel ### Ollama version ollama version is 0.1.44
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5166/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5166/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8122
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8122/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8122/comments
https://api.github.com/repos/ollama/ollama/issues/8122/events
https://github.com/ollama/ollama/pull/8122
2,743,390,539
PR_kwDOJ0Z1Ps6FaUC2
8,122
build: streamline build
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
0
2024-12-16T21:00:53
2024-12-16T21:14:49
null
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8122", "html_url": "https://github.com/ollama/ollama/pull/8122", "diff_url": "https://github.com/ollama/ollama/pull/8122.diff", "patch_url": "https://github.com/ollama/ollama/pull/8122.patch", "merged_at": null }
This wiring was intended to make a faster developer build by disabling flash attention but the added complexity and friction on updates makes this less useful
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8122/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8122/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1106
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1106/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1106/comments
https://api.github.com/repos/ollama/ollama/issues/1106/events
https://github.com/ollama/ollama/pull/1106
1,989,866,478
PR_kwDOJ0Z1Ps5fQfz5
1,106
Add Dart library to README.md
{ "login": "breitburg", "id": 25728414, "node_id": "MDQ6VXNlcjI1NzI4NDE0", "avatar_url": "https://avatars.githubusercontent.com/u/25728414?v=4", "gravatar_id": "", "url": "https://api.github.com/users/breitburg", "html_url": "https://github.com/breitburg", "followers_url": "https://api.github.com/users/breitburg/followers", "following_url": "https://api.github.com/users/breitburg/following{/other_user}", "gists_url": "https://api.github.com/users/breitburg/gists{/gist_id}", "starred_url": "https://api.github.com/users/breitburg/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/breitburg/subscriptions", "organizations_url": "https://api.github.com/users/breitburg/orgs", "repos_url": "https://api.github.com/users/breitburg/repos", "events_url": "https://api.github.com/users/breitburg/events{/privacy}", "received_events_url": "https://api.github.com/users/breitburg/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-11-13T04:25:49
2023-11-14T04:08:36
2023-11-13T19:50:42
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1106", "html_url": "https://github.com/ollama/ollama/pull/1106", "diff_url": "https://github.com/ollama/ollama/pull/1106.diff", "patch_url": "https://github.com/ollama/ollama/pull/1106.patch", "merged_at": "2023-11-13T19:50:42" }
Good afternoon! I have completed the first version of the Ollama library for Dart, making it possible to integrate Ollama into Flutter applications. I thought it would be nice to mention it in the readme file. ![](https://media.giphy.com/media/3oNMQtqpnse0dbFe06/giphy.gif)
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1106/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1106/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6317
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6317/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6317/comments
https://api.github.com/repos/ollama/ollama/issues/6317/events
https://github.com/ollama/ollama/issues/6317
2,459,986,027
I_kwDOJ0Z1Ps6SoGhr
6,317
Feature request : Tools support of Qwen2
{ "login": "trinhkiet0105", "id": 76981747, "node_id": "MDQ6VXNlcjc2OTgxNzQ3", "avatar_url": "https://avatars.githubusercontent.com/u/76981747?v=4", "gravatar_id": "", "url": "https://api.github.com/users/trinhkiet0105", "html_url": "https://github.com/trinhkiet0105", "followers_url": "https://api.github.com/users/trinhkiet0105/followers", "following_url": "https://api.github.com/users/trinhkiet0105/following{/other_user}", "gists_url": "https://api.github.com/users/trinhkiet0105/gists{/gist_id}", "starred_url": "https://api.github.com/users/trinhkiet0105/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/trinhkiet0105/subscriptions", "organizations_url": "https://api.github.com/users/trinhkiet0105/orgs", "repos_url": "https://api.github.com/users/trinhkiet0105/repos", "events_url": "https://api.github.com/users/trinhkiet0105/events{/privacy}", "received_events_url": "https://api.github.com/users/trinhkiet0105/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
5
2024-08-12T03:54:30
2024-09-02T23:49:18
2024-09-02T23:49:18
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
**Why ?** So i just found out that Qwen2 have tools support. However, ollama current do not have tools support for qwen2 models. And there is a section of [Qwen2 github talking about ollama on tools use](https://github.com/QwenLM/Qwen2?tab=readme-ov-file#-run-locally ). And this seems prombles of ollama in older versions. I think this should be resolvable with version 0.3.x **Request** - I would like to request tools support for Qwen2 models - If possible I would like to know how the process of adding tools support for a model (like ollama team can add tools to Qwen2 directly or participation of Qwen2 team is required) Thanks for reading
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6317/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6317/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7517
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7517/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7517/comments
https://api.github.com/repos/ollama/ollama/issues/7517/events
https://github.com/ollama/ollama/pull/7517
2,636,321,952
PR_kwDOJ0Z1Ps6A-N6i
7,517
Doc updates for supporting Llama3.2
{ "login": "frances720", "id": 8753634, "node_id": "MDQ6VXNlcjg3NTM2MzQ=", "avatar_url": "https://avatars.githubusercontent.com/u/8753634?v=4", "gravatar_id": "", "url": "https://api.github.com/users/frances720", "html_url": "https://github.com/frances720", "followers_url": "https://api.github.com/users/frances720/followers", "following_url": "https://api.github.com/users/frances720/following{/other_user}", "gists_url": "https://api.github.com/users/frances720/gists{/gist_id}", "starred_url": "https://api.github.com/users/frances720/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/frances720/subscriptions", "organizations_url": "https://api.github.com/users/frances720/orgs", "repos_url": "https://api.github.com/users/frances720/repos", "events_url": "https://api.github.com/users/frances720/events{/privacy}", "received_events_url": "https://api.github.com/users/frances720/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-11-05T19:44:45
2024-11-15T23:41:09
2024-11-11T03:04:24
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7517", "html_url": "https://github.com/ollama/ollama/pull/7517", "diff_url": "https://github.com/ollama/ollama/pull/7517.diff", "patch_url": "https://github.com/ollama/ollama/pull/7517.patch", "merged_at": "2024-11-11T03:04:24" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7517/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7517/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/757
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/757/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/757/comments
https://api.github.com/repos/ollama/ollama/issues/757/events
https://github.com/ollama/ollama/pull/757
1,938,405,377
PR_kwDOJ0Z1Ps5cjJ1f
757
cleanup format time
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-10-11T18:06:06
2023-10-11T18:12:30
2023-10-11T18:12:29
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/757", "html_url": "https://github.com/ollama/ollama/pull/757", "diff_url": "https://github.com/ollama/ollama/pull/757.diff", "patch_url": "https://github.com/ollama/ollama/pull/757.patch", "merged_at": "2023-10-11T18:12:29" }
only `HumanTime` is actually being used
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/757/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/757/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5870
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5870/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5870/comments
https://api.github.com/repos/ollama/ollama/issues/5870/events
https://github.com/ollama/ollama/issues/5870
2,424,725,170
I_kwDOJ0Z1Ps6Qhl6y
5,870
The embeddings api interface is not working properly.
{ "login": "xldistance", "id": 29418474, "node_id": "MDQ6VXNlcjI5NDE4NDc0", "avatar_url": "https://avatars.githubusercontent.com/u/29418474?v=4", "gravatar_id": "", "url": "https://api.github.com/users/xldistance", "html_url": "https://github.com/xldistance", "followers_url": "https://api.github.com/users/xldistance/followers", "following_url": "https://api.github.com/users/xldistance/following{/other_user}", "gists_url": "https://api.github.com/users/xldistance/gists{/gist_id}", "starred_url": "https://api.github.com/users/xldistance/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/xldistance/subscriptions", "organizations_url": "https://api.github.com/users/xldistance/orgs", "repos_url": "https://api.github.com/users/xldistance/repos", "events_url": "https://api.github.com/users/xldistance/events{/privacy}", "received_events_url": "https://api.github.com/users/xldistance/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
11
2024-07-23T09:33:21
2025-01-04T10:50:19
2024-07-30T17:55:02
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I use the bge-m3 model in graphrag with the following parameters ``` embeddings: ## parallelization: override the global parallelization settings for embeddings async_mode: asyncio llm: api_key: type: openai_embedding # or azure_openai_embedding model: chatfire/bge-m3:q8_0 api_base: http://localhost:11434/api ``` The following error is returned ``` 17:11:30,126 httpx INFO HTTP Request: POST http://localhost:11434/api/embeddings "HTTP/1.1 200 OK" 17:11:30,129 datashaper.workflow.workflow ERROR Error executing verb "text_embed" in create_final_entities: 'NoneType' object is not iterable Traceback (most recent call last): File "E:\Langchain-Chatchat\glut\lib\site-packages\datashaper\workflow\workflow.py", line 415, in _execute_verb result = await result File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\index\verbs\text\embed\text_embed.py", line 105, in text_embed return await _text_embed_in_memory( File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\index\verbs\text\embed\text_embed.py", line 130, in _text_embed_in_memory result = await strategy_exec(texts, callbacks, cache, strategy_args) File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\index\verbs\text\embed\strategies\openai.py", line 61, in run embeddings = await _execute(llm, text_batches, ticker, semaphore) File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\index\verbs\text\embed\strategies\openai.py", line 105, in _execute results = await asyncio.gather(*futures) File "E:\Langchain-Chatchat\glut\lib\asyncio\tasks.py", line 304, in __wakeup future.result() File "E:\Langchain-Chatchat\glut\lib\asyncio\tasks.py", line 232, in __step result = coro.send(None) File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\index\verbs\text\embed\strategies\openai.py", line 99, in embed chunk_embeddings = await llm(chunk) File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\base\caching_llm.py", line 104, in __call__ result = await self._delegate(input, **kwargs) File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\base\rate_limiting_llm.py", line 177, in __call__ result, start = await execute_with_retry() File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\base\rate_limiting_llm.py", line 159, in execute_with_retry async for attempt in retryer: File "E:\Langchain-Chatchat\glut\lib\site-packages\tenacity\_asyncio.py", line 71, in __anext__ do = self.iter(retry_state=self._retry_state) File "E:\Langchain-Chatchat\glut\lib\site-packages\tenacity\__init__.py", line 314, in iter return fut.result() File "E:\Langchain-Chatchat\glut\lib\concurrent\futures\_base.py", line 451, in result return self.__get_result() File "E:\Langchain-Chatchat\glut\lib\concurrent\futures\_base.py", line 403, in __get_result raise self._exception File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\base\rate_limiting_llm.py", line 165, in execute_with_retry return await do_attempt(), start File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\base\rate_limiting_llm.py", line 147, in do_attempt return await self._delegate(input, **kwargs) File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\base\base_llm.py", line 49, in __call__ return await self._invoke(input, **kwargs) File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\base\base_llm.py", line 53, in _invoke output = await self._execute_llm(input, **kwargs) File "E:\Langchain-Chatchat\glut\lib\site-packages\graphrag\llm\openai\openai_embeddings_llm.py", line 36, in _execute_llm embedding = await self.client.embeddings.create( File "E:\Langchain-Chatchat\glut\lib\site-packages\openai\resources\embeddings.py", line 215, in create return await self._post( File "E:\Langchain-Chatchat\glut\lib\site-packages\openai\_base_client.py", line 1826, in post return await self.request(cast_to, opts, stream=stream, stream_cls=stream_cls) File "E:\Langchain-Chatchat\glut\lib\site-packages\openai\_base_client.py", line 1519, in request return await self._request( File "E:\Langchain-Chatchat\glut\lib\site-packages\openai\_base_client.py", line 1622, in _request return await self._process_response( File "E:\Langchain-Chatchat\glut\lib\site-packages\openai\_base_client.py", line 1714, in _process_response return await api_response.parse() File "E:\Langchain-Chatchat\glut\lib\site-packages\openai\_response.py", line 419, in parse parsed = self._options.post_parser(parsed) File "E:\Langchain-Chatchat\glut\lib\site-packages\openai\resources\embeddings.py", line 203, in parser for embedding in obj.data: TypeError: 'NoneType' object is not iterable 17:11:30,131 graphrag.index.reporting.file_workflow_callbacks INFO Error executing verb "text_embed" in create_final_entities: 'NoneType' object is not iterable details=None 17:11:30,142 graphrag.index.run ERROR error running workflow create_final_entities ``` ### OS Windows ### GPU Nvidia ### CPU AMD ### Ollama version 0.2.8
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5870/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5870/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5291
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5291/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5291/comments
https://api.github.com/repos/ollama/ollama/issues/5291/events
https://github.com/ollama/ollama/issues/5291
2,374,355,746
I_kwDOJ0Z1Ps6Nhcsi
5,291
请上架cogvlm2
{ "login": "enryteam", "id": 20081090, "node_id": "MDQ6VXNlcjIwMDgxMDkw", "avatar_url": "https://avatars.githubusercontent.com/u/20081090?v=4", "gravatar_id": "", "url": "https://api.github.com/users/enryteam", "html_url": "https://github.com/enryteam", "followers_url": "https://api.github.com/users/enryteam/followers", "following_url": "https://api.github.com/users/enryteam/following{/other_user}", "gists_url": "https://api.github.com/users/enryteam/gists{/gist_id}", "starred_url": "https://api.github.com/users/enryteam/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/enryteam/subscriptions", "organizations_url": "https://api.github.com/users/enryteam/orgs", "repos_url": "https://api.github.com/users/enryteam/repos", "events_url": "https://api.github.com/users/enryteam/events{/privacy}", "received_events_url": "https://api.github.com/users/enryteam/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
1
2024-06-26T05:57:32
2024-06-26T12:12:42
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
https://huggingface.co/THUDM/cogvlm2-llama3-chinese-chat-19B thanks 谢谢 ollama0.1.43 error format not yet support! 错误格式尚不支持! 错误格式尚不支持! 搞了多次 均报错
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5291/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5291/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/3544
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3544/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3544/comments
https://api.github.com/repos/ollama/ollama/issues/3544/events
https://github.com/ollama/ollama/issues/3544
2,232,444,606
I_kwDOJ0Z1Ps6FEGa-
3,544
ollama 0.1.31 Segmentation fault (core dumped)
{ "login": "zhqfdn", "id": 25156863, "node_id": "MDQ6VXNlcjI1MTU2ODYz", "avatar_url": "https://avatars.githubusercontent.com/u/25156863?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zhqfdn", "html_url": "https://github.com/zhqfdn", "followers_url": "https://api.github.com/users/zhqfdn/followers", "following_url": "https://api.github.com/users/zhqfdn/following{/other_user}", "gists_url": "https://api.github.com/users/zhqfdn/gists{/gist_id}", "starred_url": "https://api.github.com/users/zhqfdn/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zhqfdn/subscriptions", "organizations_url": "https://api.github.com/users/zhqfdn/orgs", "repos_url": "https://api.github.com/users/zhqfdn/repos", "events_url": "https://api.github.com/users/zhqfdn/events{/privacy}", "received_events_url": "https://api.github.com/users/zhqfdn/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
3
2024-04-09T01:36:09
2024-05-01T16:43:13
2024-05-01T16:43:13
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? [root@localhost ~]# cat /etc/redhat-release AlmaLinux release 9.3 (Shamrock Pampas Cat) [root@localhost ~]# ollama -v Warning: could not connect to a running Ollama instance Warning: client version is 0.1.30 [root@localhost ~]# ./ollama -v Segmentation fault (core dumped) [root@localhost ~]# ### What did you expect to see? [root@localhost ~]# ./ollama -v Segmentation fault (core dumped) ### Steps to reproduce [root@localhost ~]# ./ollama -v Segmentation fault (core dumped) ### Are there any recent changes that introduced the issue? [root@localhost ~]# ./ollama -v Segmentation fault (core dumped) ### OS Linux ### Architecture amd64 ### Platform _No response_ ### Ollama version 0.1.31 ### GPU Nvidia ### GPU info [root@localhost ~]# nvidia-smi Tue Apr 9 09:31:38 2024 +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 550.54.15 Driver Version: 550.54.15 CUDA Version: 12.4 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 Tesla T4 Off | 00000000:06:00.0 Off | 0 | | N/A 34C P0 27W / 70W | 0MiB / 15360MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 1 Tesla T4 Off | 00000000:86:00.0 Off | 0 | | N/A 34C P0 26W / 70W | 0MiB / 15360MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ +-----------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=========================================================================================| | No running processes found | +-----------------------------------------------------------------------------------------+ ### CPU Intel ### Other software AlmaLinux release 9.3 (Shamrock Pampas Cat)
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3544/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3544/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5712
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5712/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5712/comments
https://api.github.com/repos/ollama/ollama/issues/5712/events
https://github.com/ollama/ollama/pull/5712
2,409,800,448
PR_kwDOJ0Z1Ps51cb6s
5,712
Add Windows arm64 support to official builds
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
14
2024-07-15T23:18:15
2024-09-20T20:09:41
2024-09-20T20:09:38
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5712", "html_url": "https://github.com/ollama/ollama/pull/5712", "diff_url": "https://github.com/ollama/ollama/pull/5712.diff", "patch_url": "https://github.com/ollama/ollama/pull/5712.patch", "merged_at": "2024-09-20T20:09:38" }
Wire up CI and build rigging to generate a unified Windows installer with x64 and arm64 payloads. At install time, the correct binaries will be installed for the platform. I was unable to find a combination of hand-picked msvc redist DLLs manually that yielded a working setup on a pristine Windows 11 install, but running the vc_redist installer works reliably, so for arm64, we run the nested installer conditionally. If it is already installed, that step will be skipped. Fixes #2589 Note: I've tested most of the CI steps in the PR, but signing isn't yet verified and might require minor fixes on the first release after this merges. Resulting build artifacts: (Note: current OllamaSetup.exe with only x64 binaries is 273MB) ``` % ls -lh dist/ total 932M -rw-r--r-- 1 daniel 197609 12K Jul 17 09:24 ollama_welcome.ps1 -rwxr-xr-x 1 daniel 197609 291M Jul 17 09:27 OllamaSetup.exe* -rw-r--r-- 1 daniel 197609 649M Jul 17 09:27 ollama-windows-amd64.zip -rw-r--r-- 1 daniel 197609 20M Jul 19 15:41 ollama-windows-arm64.zip drwxr-xr-x 1 daniel 197609 0 Jul 17 09:24 windows-amd64/ -rwxr-xr-x 1 daniel 197609 5.9M Jul 17 09:24 windows-amd64-app.exe* drwxr-xr-x 1 daniel 197609 0 Jul 16 15:53 windows-arm64/ -rwxr-xr-x 1 daniel 197609 5.5M Jul 16 16:12 windows-arm64-app.exe* % du -sh dist/windows-a*64 2.1G dist/windows-amd64 37M dist/windows-arm64 ``` On a Snapdragon X 12-core laptop: ``` > ollama run --verbose llama3 why is the sky blue ... total duration: 23.6819409s load duration: 4.738127s prompt eval count: 16 token(s) prompt eval duration: 430.297ms prompt eval rate: 37.18 tokens/s eval count: 348 token(s) eval duration: 18.513796s eval rate: 18.80 tokens/s ```
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5712/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5712/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2120
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2120/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2120/comments
https://api.github.com/repos/ollama/ollama/issues/2120/events
https://github.com/ollama/ollama/issues/2120
2,092,519,133
I_kwDOJ0Z1Ps58uU7d
2,120
How to install libnvidia-ml.so?
{ "login": "silverwind63", "id": 104142549, "node_id": "U_kgDOBjUW1Q", "avatar_url": "https://avatars.githubusercontent.com/u/104142549?v=4", "gravatar_id": "", "url": "https://api.github.com/users/silverwind63", "html_url": "https://github.com/silverwind63", "followers_url": "https://api.github.com/users/silverwind63/followers", "following_url": "https://api.github.com/users/silverwind63/following{/other_user}", "gists_url": "https://api.github.com/users/silverwind63/gists{/gist_id}", "starred_url": "https://api.github.com/users/silverwind63/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/silverwind63/subscriptions", "organizations_url": "https://api.github.com/users/silverwind63/orgs", "repos_url": "https://api.github.com/users/silverwind63/repos", "events_url": "https://api.github.com/users/silverwind63/events{/privacy}", "received_events_url": "https://api.github.com/users/silverwind63/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
4
2024-01-21T10:16:16
2024-01-27T11:25:08
2024-01-26T21:06:50
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi guys! I have been using ollama with ollama webui this month.However,it output ``` WARNING: You should always run with libnvidia-ml.so that is installed with your NVIDIA Display Driver. By default it's installed in /usr/lib and /usr/lib64. libnvidia-ml.so in GDK package is a stub library that is attached only for build purposes (e.g. machine that you build your application doesn't have to have Display Driver installed). ``` And whenever I want to run any model(which is capable to load it and the speed is about 5 tokens/s) it will always run into cuda memory error. My system: RAM:16GB GPU:3060ti 8GB SYSTEM:archlinux Kernel:6.7.0-arch3-1 Nvidia GPU Driver:nvidia-dkms 545.29.06-1 I have also installed following package which is related to nvidia: ``` lib32-nvidia-utils 545.29.06-1 libnvidia-container 1.14.3-1 libnvidia-container-tools 1.14.3-1 libva-nvidia-driver-git 0.0.11.r1.gea6d862-1 nvidia-container-toolkit 1.14.3-9 nvidia-docker-compose 0.1.6-1 nvidia-utils 545.29.06-1 ```
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2120/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2120/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2639
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2639/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2639/comments
https://api.github.com/repos/ollama/ollama/issues/2639/events
https://github.com/ollama/ollama/issues/2639
2,147,069,563
I_kwDOJ0Z1Ps5_-a57
2,639
History via up arrow and down arrow not working on windows using `ollama run`
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" } ]
closed
false
null
[]
null
0
2024-02-21T15:39:35
2024-03-26T22:21:57
2024-03-26T22:21:57
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2639/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2639/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/612
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/612/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/612/comments
https://api.github.com/repos/ollama/ollama/issues/612/events
https://github.com/ollama/ollama/pull/612
1,914,484,120
PR_kwDOJ0Z1Ps5bSJCL
612
prune empty directories
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-09-27T00:40:27
2023-09-29T18:23:41
2023-09-29T18:23:40
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/612", "html_url": "https://github.com/ollama/ollama/pull/612", "diff_url": "https://github.com/ollama/ollama/pull/612.diff", "patch_url": "https://github.com/ollama/ollama/pull/612.patch", "merged_at": "2023-09-29T18:23:40" }
Resolves #270
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/612/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/612/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1033
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1033/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1033/comments
https://api.github.com/repos/ollama/ollama/issues/1033/events
https://github.com/ollama/ollama/issues/1033
1,981,678,318
I_kwDOJ0Z1Ps52HgLu
1,033
Are these system specs good enough for any models?
{ "login": "simoovara", "id": 100516318, "node_id": "U_kgDOBf3B3g", "avatar_url": "https://avatars.githubusercontent.com/u/100516318?v=4", "gravatar_id": "", "url": "https://api.github.com/users/simoovara", "html_url": "https://github.com/simoovara", "followers_url": "https://api.github.com/users/simoovara/followers", "following_url": "https://api.github.com/users/simoovara/following{/other_user}", "gists_url": "https://api.github.com/users/simoovara/gists{/gist_id}", "starred_url": "https://api.github.com/users/simoovara/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/simoovara/subscriptions", "organizations_url": "https://api.github.com/users/simoovara/orgs", "repos_url": "https://api.github.com/users/simoovara/repos", "events_url": "https://api.github.com/users/simoovara/events{/privacy}", "received_events_url": "https://api.github.com/users/simoovara/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
6
2023-11-07T15:50:36
2023-11-07T21:32:31
2023-11-07T21:32:14
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Just a question, I have an old laptop that i turned into a server with Ubuntu LTS. It has an AMD E1-6015 APU and 8gb of ram. I would like to know if that's enough to run any of these models, thank you!
{ "login": "simoovara", "id": 100516318, "node_id": "U_kgDOBf3B3g", "avatar_url": "https://avatars.githubusercontent.com/u/100516318?v=4", "gravatar_id": "", "url": "https://api.github.com/users/simoovara", "html_url": "https://github.com/simoovara", "followers_url": "https://api.github.com/users/simoovara/followers", "following_url": "https://api.github.com/users/simoovara/following{/other_user}", "gists_url": "https://api.github.com/users/simoovara/gists{/gist_id}", "starred_url": "https://api.github.com/users/simoovara/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/simoovara/subscriptions", "organizations_url": "https://api.github.com/users/simoovara/orgs", "repos_url": "https://api.github.com/users/simoovara/repos", "events_url": "https://api.github.com/users/simoovara/events{/privacy}", "received_events_url": "https://api.github.com/users/simoovara/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1033/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1033/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5907
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5907/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5907/comments
https://api.github.com/repos/ollama/ollama/issues/5907/events
https://github.com/ollama/ollama/issues/5907
2,427,317,383
I_kwDOJ0Z1Ps6QreyH
5,907
Support token embeddings for `v1/embeddings`
{ "login": "WoJiaoFuXiaoYun", "id": 30924105, "node_id": "MDQ6VXNlcjMwOTI0MTA1", "avatar_url": "https://avatars.githubusercontent.com/u/30924105?v=4", "gravatar_id": "", "url": "https://api.github.com/users/WoJiaoFuXiaoYun", "html_url": "https://github.com/WoJiaoFuXiaoYun", "followers_url": "https://api.github.com/users/WoJiaoFuXiaoYun/followers", "following_url": "https://api.github.com/users/WoJiaoFuXiaoYun/following{/other_user}", "gists_url": "https://api.github.com/users/WoJiaoFuXiaoYun/gists{/gist_id}", "starred_url": "https://api.github.com/users/WoJiaoFuXiaoYun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/WoJiaoFuXiaoYun/subscriptions", "organizations_url": "https://api.github.com/users/WoJiaoFuXiaoYun/orgs", "repos_url": "https://api.github.com/users/WoJiaoFuXiaoYun/repos", "events_url": "https://api.github.com/users/WoJiaoFuXiaoYun/events{/privacy}", "received_events_url": "https://api.github.com/users/WoJiaoFuXiaoYun/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 7706482389, "node_id": "LA_kwDOJ0Z1Ps8AAAABy1eW1Q", "url": "https://api.github.com/repos/ollama/ollama/labels/api", "name": "api", "color": "bfdadc", "default": false, "description": "" } ]
open
false
null
[]
null
3
2024-07-24T11:17:36
2024-11-06T01:00:33
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? When encoding with `tiktoken`, the interface is no longer compatible ``` tiktoken.get_encoding("cl100k_base").encode(text) ``` ```json { "input": [30624,99849,64479,51392,31809,29207,233,45829], "model": "nomic-embed-text" } ``` ``` { "error": { "message": "invalid input type", "type": "api_error", "param": null, "code": null } } ``` I hope it can be compatible, thank you ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version ollama version is 0.2.7
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5907/reactions", "total_count": 2, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/ollama/ollama/issues/5907/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/5107
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5107/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5107/comments
https://api.github.com/repos/ollama/ollama/issues/5107/events
https://github.com/ollama/ollama/issues/5107
2,358,707,183
I_kwDOJ0Z1Ps6MlwPv
5,107
ollama 模型授权
{ "login": "yawzhe", "id": 127652671, "node_id": "U_kgDOB5vTPw", "avatar_url": "https://avatars.githubusercontent.com/u/127652671?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yawzhe", "html_url": "https://github.com/yawzhe", "followers_url": "https://api.github.com/users/yawzhe/followers", "following_url": "https://api.github.com/users/yawzhe/following{/other_user}", "gists_url": "https://api.github.com/users/yawzhe/gists{/gist_id}", "starred_url": "https://api.github.com/users/yawzhe/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yawzhe/subscriptions", "organizations_url": "https://api.github.com/users/yawzhe/orgs", "repos_url": "https://api.github.com/users/yawzhe/repos", "events_url": "https://api.github.com/users/yawzhe/events{/privacy}", "received_events_url": "https://api.github.com/users/yawzhe/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-06-18T02:16:01
2024-06-18T11:28:38
2024-06-18T11:28:37
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? 我想问一下ollama怎么自定义设置KEY_传参, 每个模型定义不同的key,2.ollama是否支持模型授权,加密模型之类的, ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 最新的
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5107/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5107/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4621
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4621/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4621/comments
https://api.github.com/repos/ollama/ollama/issues/4621/events
https://github.com/ollama/ollama/issues/4621
2,316,267,571
I_kwDOJ0Z1Ps6KD3Az
4,621
phi3-medium-128k wrong number of tensors
{ "login": "EthanGraber", "id": 18070053, "node_id": "MDQ6VXNlcjE4MDcwMDUz", "avatar_url": "https://avatars.githubusercontent.com/u/18070053?v=4", "gravatar_id": "", "url": "https://api.github.com/users/EthanGraber", "html_url": "https://github.com/EthanGraber", "followers_url": "https://api.github.com/users/EthanGraber/followers", "following_url": "https://api.github.com/users/EthanGraber/following{/other_user}", "gists_url": "https://api.github.com/users/EthanGraber/gists{/gist_id}", "starred_url": "https://api.github.com/users/EthanGraber/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/EthanGraber/subscriptions", "organizations_url": "https://api.github.com/users/EthanGraber/orgs", "repos_url": "https://api.github.com/users/EthanGraber/repos", "events_url": "https://api.github.com/users/EthanGraber/events{/privacy}", "received_events_url": "https://api.github.com/users/EthanGraber/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-05-24T21:11:10
2024-05-24T22:13:06
2024-05-24T22:13:05
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I'm getting the following error when testing the new 128k versions of phi3-medium: ```sh $ ollama run phi3:14b-medium-128k-instruct-q4_0 Error: llama runner process has terminated: signal: abort trap error:done_getting_tensors: wrong number of tensors; expected 245, got 243 ``` ```sh $ ollama -v ollama version is 0.1.38 ``` ### OS macOS ### GPU Apple ### CPU Apple ### Ollama version 0.1.38
{ "login": "EthanGraber", "id": 18070053, "node_id": "MDQ6VXNlcjE4MDcwMDUz", "avatar_url": "https://avatars.githubusercontent.com/u/18070053?v=4", "gravatar_id": "", "url": "https://api.github.com/users/EthanGraber", "html_url": "https://github.com/EthanGraber", "followers_url": "https://api.github.com/users/EthanGraber/followers", "following_url": "https://api.github.com/users/EthanGraber/following{/other_user}", "gists_url": "https://api.github.com/users/EthanGraber/gists{/gist_id}", "starred_url": "https://api.github.com/users/EthanGraber/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/EthanGraber/subscriptions", "organizations_url": "https://api.github.com/users/EthanGraber/orgs", "repos_url": "https://api.github.com/users/EthanGraber/repos", "events_url": "https://api.github.com/users/EthanGraber/events{/privacy}", "received_events_url": "https://api.github.com/users/EthanGraber/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4621/reactions", "total_count": 2, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/ollama/ollama/issues/4621/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4226
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4226/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4226/comments
https://api.github.com/repos/ollama/ollama/issues/4226/events
https://github.com/ollama/ollama/issues/4226
2,282,922,640
I_kwDOJ0Z1Ps6IEqKQ
4,226
run llama3-70B-q8_0 error
{ "login": "leoHostProject", "id": 87935281, "node_id": "MDQ6VXNlcjg3OTM1Mjgx", "avatar_url": "https://avatars.githubusercontent.com/u/87935281?v=4", "gravatar_id": "", "url": "https://api.github.com/users/leoHostProject", "html_url": "https://github.com/leoHostProject", "followers_url": "https://api.github.com/users/leoHostProject/followers", "following_url": "https://api.github.com/users/leoHostProject/following{/other_user}", "gists_url": "https://api.github.com/users/leoHostProject/gists{/gist_id}", "starred_url": "https://api.github.com/users/leoHostProject/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/leoHostProject/subscriptions", "organizations_url": "https://api.github.com/users/leoHostProject/orgs", "repos_url": "https://api.github.com/users/leoHostProject/repos", "events_url": "https://api.github.com/users/leoHostProject/events{/privacy}", "received_events_url": "https://api.github.com/users/leoHostProject/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" }, { "id": 6677745918, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgZQ_g", "url": "https://api.github.com/repos/ollama/ollama/labels/gpu", "name": "gpu", "color": "76C49E", "default": false, "description": "" } ]
closed
false
null
[]
null
1
2024-05-07T10:44:08
2024-07-25T18:53:03
2024-07-25T18:53:02
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? api call error message: {"error":{"message":"timed out waiting for llama runner to start:CUDA error:uncorrectable ECC error encountered\ n current device:0,in function ggml cuda_compute_forward at /go/src/github.com/ollama/ollama/11m/1lama.cpp/ggml -cuda.cu:2300\n err\nGGML_ASSERT:/go/src/github.com/ollama/ollama/11m/1lama.cpp/ggml-cuda.cu:60:!\"CUDA error\" ""type":"api_error","param"null,"code":null}} ### OS Linux, Docker ### GPU Nvidia ### CPU Intel ### Ollama version _No response_
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4226/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4226/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5750
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5750/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5750/comments
https://api.github.com/repos/ollama/ollama/issues/5750/events
https://github.com/ollama/ollama/pull/5750
2,414,156,448
PR_kwDOJ0Z1Ps51q_iz
5,750
stub response
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-07-17T17:28:46
2024-07-17T17:39:25
2024-07-17T17:39:22
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5750", "html_url": "https://github.com/ollama/ollama/pull/5750", "diff_url": "https://github.com/ollama/ollama/pull/5750.diff", "patch_url": "https://github.com/ollama/ollama/pull/5750.patch", "merged_at": "2024-07-17T17:39:22" }
for compatibility, `{{ .Response }}` cannot be in any template control flow structures. therefore any template execution should set an empty Response if one should not be rendered otherwise the output will contain `<no value>` in place of `{{ .Response }}`
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5750/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5750/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4183
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4183/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4183/comments
https://api.github.com/repos/ollama/ollama/issues/4183/events
https://github.com/ollama/ollama/issues/4183
2,279,709,927
I_kwDOJ0Z1Ps6H4Zzn
4,183
pull orca2:7b-fp16 Error: EOF
{ "login": "MarkWard0110", "id": 90335263, "node_id": "MDQ6VXNlcjkwMzM1MjYz", "avatar_url": "https://avatars.githubusercontent.com/u/90335263?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MarkWard0110", "html_url": "https://github.com/MarkWard0110", "followers_url": "https://api.github.com/users/MarkWard0110/followers", "following_url": "https://api.github.com/users/MarkWard0110/following{/other_user}", "gists_url": "https://api.github.com/users/MarkWard0110/gists{/gist_id}", "starred_url": "https://api.github.com/users/MarkWard0110/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MarkWard0110/subscriptions", "organizations_url": "https://api.github.com/users/MarkWard0110/orgs", "repos_url": "https://api.github.com/users/MarkWard0110/repos", "events_url": "https://api.github.com/users/MarkWard0110/events{/privacy}", "received_events_url": "https://api.github.com/users/MarkWard0110/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-05-05T20:14:27
2024-05-05T20:17:17
2024-05-05T20:17:17
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? `ollama pull orca2:7b-fp16` results in `Error: EOF` ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.1.33
{ "login": "MarkWard0110", "id": 90335263, "node_id": "MDQ6VXNlcjkwMzM1MjYz", "avatar_url": "https://avatars.githubusercontent.com/u/90335263?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MarkWard0110", "html_url": "https://github.com/MarkWard0110", "followers_url": "https://api.github.com/users/MarkWard0110/followers", "following_url": "https://api.github.com/users/MarkWard0110/following{/other_user}", "gists_url": "https://api.github.com/users/MarkWard0110/gists{/gist_id}", "starred_url": "https://api.github.com/users/MarkWard0110/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MarkWard0110/subscriptions", "organizations_url": "https://api.github.com/users/MarkWard0110/orgs", "repos_url": "https://api.github.com/users/MarkWard0110/repos", "events_url": "https://api.github.com/users/MarkWard0110/events{/privacy}", "received_events_url": "https://api.github.com/users/MarkWard0110/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4183/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4183/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3397
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3397/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3397/comments
https://api.github.com/repos/ollama/ollama/issues/3397/events
https://github.com/ollama/ollama/pull/3397
2,214,232,629
PR_kwDOJ0Z1Ps5rG7Q3
3,397
Parallel requests
{ "login": "0x77dev", "id": 46429701, "node_id": "MDQ6VXNlcjQ2NDI5NzAx", "avatar_url": "https://avatars.githubusercontent.com/u/46429701?v=4", "gravatar_id": "", "url": "https://api.github.com/users/0x77dev", "html_url": "https://github.com/0x77dev", "followers_url": "https://api.github.com/users/0x77dev/followers", "following_url": "https://api.github.com/users/0x77dev/following{/other_user}", "gists_url": "https://api.github.com/users/0x77dev/gists{/gist_id}", "starred_url": "https://api.github.com/users/0x77dev/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/0x77dev/subscriptions", "organizations_url": "https://api.github.com/users/0x77dev/orgs", "repos_url": "https://api.github.com/users/0x77dev/repos", "events_url": "https://api.github.com/users/0x77dev/events{/privacy}", "received_events_url": "https://api.github.com/users/0x77dev/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-03-28T21:54:46
2024-03-30T22:41:18
2024-03-30T22:41:18
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
true
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3397", "html_url": "https://github.com/ollama/ollama/pull/3397", "diff_url": "https://github.com/ollama/ollama/pull/3397.diff", "patch_url": "https://github.com/ollama/ollama/pull/3397.patch", "merged_at": null }
Stage: PoC Related issue: #358 - loaded.mu.{Lock,Unlock}() is not implemented correctly in this change - sparams.n_parallel is hardcoded to 4
{ "login": "0x77dev", "id": 46429701, "node_id": "MDQ6VXNlcjQ2NDI5NzAx", "avatar_url": "https://avatars.githubusercontent.com/u/46429701?v=4", "gravatar_id": "", "url": "https://api.github.com/users/0x77dev", "html_url": "https://github.com/0x77dev", "followers_url": "https://api.github.com/users/0x77dev/followers", "following_url": "https://api.github.com/users/0x77dev/following{/other_user}", "gists_url": "https://api.github.com/users/0x77dev/gists{/gist_id}", "starred_url": "https://api.github.com/users/0x77dev/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/0x77dev/subscriptions", "organizations_url": "https://api.github.com/users/0x77dev/orgs", "repos_url": "https://api.github.com/users/0x77dev/repos", "events_url": "https://api.github.com/users/0x77dev/events{/privacy}", "received_events_url": "https://api.github.com/users/0x77dev/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3397/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3397/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1158
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1158/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1158/comments
https://api.github.com/repos/ollama/ollama/issues/1158/events
https://github.com/ollama/ollama/issues/1158
1,997,987,411
I_kwDOJ0Z1Ps53Ft5T
1,158
max retries exceeded: unexpected EOF
{ "login": "priamai", "id": 57333254, "node_id": "MDQ6VXNlcjU3MzMzMjU0", "avatar_url": "https://avatars.githubusercontent.com/u/57333254?v=4", "gravatar_id": "", "url": "https://api.github.com/users/priamai", "html_url": "https://github.com/priamai", "followers_url": "https://api.github.com/users/priamai/followers", "following_url": "https://api.github.com/users/priamai/following{/other_user}", "gists_url": "https://api.github.com/users/priamai/gists{/gist_id}", "starred_url": "https://api.github.com/users/priamai/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/priamai/subscriptions", "organizations_url": "https://api.github.com/users/priamai/orgs", "repos_url": "https://api.github.com/users/priamai/repos", "events_url": "https://api.github.com/users/priamai/events{/privacy}", "received_events_url": "https://api.github.com/users/priamai/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
15
2023-11-16T23:47:19
2025-01-28T16:11:44
2024-03-11T18:25:00
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi there, I am not sure if this is related to your file service, but I am getting this connection drops out very often. ![Screenshot from 2023-11-16 23-45-54](https://github.com/jmorganca/ollama/assets/57333254/d530f24e-af82-49d8-9435-0653922d1eec) Maybe there is a way to throttle requests?
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1158/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1158/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/216
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/216/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/216/comments
https://api.github.com/repos/ollama/ollama/issues/216/events
https://github.com/ollama/ollama/issues/216
1,822,180,868
I_kwDOJ0Z1Ps5snEYE
216
Something might still be wrong with K-Quant
{ "login": "nkoehring", "id": 246402, "node_id": "MDQ6VXNlcjI0NjQwMg==", "avatar_url": "https://avatars.githubusercontent.com/u/246402?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nkoehring", "html_url": "https://github.com/nkoehring", "followers_url": "https://api.github.com/users/nkoehring/followers", "following_url": "https://api.github.com/users/nkoehring/following{/other_user}", "gists_url": "https://api.github.com/users/nkoehring/gists{/gist_id}", "starred_url": "https://api.github.com/users/nkoehring/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nkoehring/subscriptions", "organizations_url": "https://api.github.com/users/nkoehring/orgs", "repos_url": "https://api.github.com/users/nkoehring/repos", "events_url": "https://api.github.com/users/nkoehring/events{/privacy}", "received_events_url": "https://api.github.com/users/nkoehring/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
3
2023-07-26T11:17:46
2023-08-02T19:03:27
2023-08-02T19:03:27
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
When I run a 30B model (in this case upstage-llama-30b-instruct-2048.ggmlv3.q5_K_M.bin) the debug output in ollama talks about a 13B model size: ![Screenshot from 2023-07-26 13-07-51](https://github.com/jmorganca/ollama/assets/246402/36bb44f1-a534-44ae-94bb-3e87d7ce5a74) when running the same model with llama.cpp it outputs the correct size: ![Screenshot from 2023-07-26 13-11-07](https://github.com/jmorganca/ollama/assets/246402/2eb0621f-683b-4ea7-82a0-5aedf8292a03) I tested with a 13B model and the output seems correct. Both models seem to work (as in generating output).
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/216/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/216/timeline
null
not_planned
false
https://api.github.com/repos/ollama/ollama/issues/4880
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4880/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4880/comments
https://api.github.com/repos/ollama/ollama/issues/4880/events
https://github.com/ollama/ollama/issues/4880
2,339,156,469
I_kwDOJ0Z1Ps6LbLH1
4,880
Extend ollama show command
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
0
2024-06-06T21:07:21
2024-06-26T17:31:00
2024-06-26T17:31:00
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
In reference to #3570
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4880/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4880/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7355
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7355/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7355/comments
https://api.github.com/repos/ollama/ollama/issues/7355/events
https://github.com/ollama/ollama/issues/7355
2,613,814,322
I_kwDOJ0Z1Ps6by6Qy
7,355
Released binaries have High severity CVEs due to Go version 1.22.5
{ "login": "pivotal-marcela-campo", "id": 20945140, "node_id": "MDQ6VXNlcjIwOTQ1MTQw", "avatar_url": "https://avatars.githubusercontent.com/u/20945140?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pivotal-marcela-campo", "html_url": "https://github.com/pivotal-marcela-campo", "followers_url": "https://api.github.com/users/pivotal-marcela-campo/followers", "following_url": "https://api.github.com/users/pivotal-marcela-campo/following{/other_user}", "gists_url": "https://api.github.com/users/pivotal-marcela-campo/gists{/gist_id}", "starred_url": "https://api.github.com/users/pivotal-marcela-campo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pivotal-marcela-campo/subscriptions", "organizations_url": "https://api.github.com/users/pivotal-marcela-campo/orgs", "repos_url": "https://api.github.com/users/pivotal-marcela-campo/repos", "events_url": "https://api.github.com/users/pivotal-marcela-campo/events{/privacy}", "received_events_url": "https://api.github.com/users/pivotal-marcela-campo/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
1
2024-10-25T11:17:12
2024-10-27T00:03:38
2024-10-27T00:03:38
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Scanning linux binary with `grype` yields the following report ![Screenshot 2024-10-25 at 11 49 39](https://github.com/user-attachments/assets/7c4fe7af-13d4-4ddc-9339-2bef323691a8) Upgrading to 1.22.7+ for building would fix this issue: https://github.com/ollama/ollama/blob/3085c47bea508e638c0b9f15a2f4c00afd83b66d/Dockerfile#L1 https://github.com/ollama/ollama/blob/3085c47bea508e638c0b9f15a2f4c00afd83b66d/llama/Dockerfile#L2 https://github.com/ollama/ollama/blob/3085c47bea508e638c0b9f15a2f4c00afd83b66d/go.mod#L3 ### OS Linux, macOS, Docker ### GPU Nvidia ### CPU Intel ### Ollama version 0.3.14
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7355/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7355/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5004
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5004/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5004/comments
https://api.github.com/repos/ollama/ollama/issues/5004/events
https://github.com/ollama/ollama/pull/5004
2,349,490,461
PR_kwDOJ0Z1Ps5yRrGr
5,004
fix: multiple templates when creating from model
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-06-12T19:00:13
2024-06-12T21:39:29
2024-06-12T21:39:29
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5004", "html_url": "https://github.com/ollama/ollama/pull/5004", "diff_url": "https://github.com/ollama/ollama/pull/5004.diff", "patch_url": "https://github.com/ollama/ollama/pull/5004.patch", "merged_at": "2024-06-12T21:39:29" }
multiple templates may appear in a model if a model is created from another model that 1) has an autodetected template and 2) defines a custom template this fixes the bug by not detecting chat template when inheriting from another model
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5004/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5004/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6676
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6676/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6676/comments
https://api.github.com/repos/ollama/ollama/issues/6676/events
https://github.com/ollama/ollama/issues/6676
2,510,163,293
I_kwDOJ0Z1Ps6Vng1d
6,676
on ollama.com , the centrate new profile picture page , looked on andro chrome canary , out of bound
{ "login": "fxmbsw7", "id": 39368685, "node_id": "MDQ6VXNlcjM5MzY4Njg1", "avatar_url": "https://avatars.githubusercontent.com/u/39368685?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fxmbsw7", "html_url": "https://github.com/fxmbsw7", "followers_url": "https://api.github.com/users/fxmbsw7/followers", "following_url": "https://api.github.com/users/fxmbsw7/following{/other_user}", "gists_url": "https://api.github.com/users/fxmbsw7/gists{/gist_id}", "starred_url": "https://api.github.com/users/fxmbsw7/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/fxmbsw7/subscriptions", "organizations_url": "https://api.github.com/users/fxmbsw7/orgs", "repos_url": "https://api.github.com/users/fxmbsw7/repos", "events_url": "https://api.github.com/users/fxmbsw7/events{/privacy}", "received_events_url": "https://api.github.com/users/fxmbsw7/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6573197867, "node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw", "url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com", "name": "ollama.com", "color": "ffffff", "default": false, "description": "" } ]
open
false
{ "login": "hoyyeva", "id": 63033505, "node_id": "MDQ6VXNlcjYzMDMzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoyyeva", "html_url": "https://github.com/hoyyeva", "followers_url": "https://api.github.com/users/hoyyeva/followers", "following_url": "https://api.github.com/users/hoyyeva/following{/other_user}", "gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}", "starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions", "organizations_url": "https://api.github.com/users/hoyyeva/orgs", "repos_url": "https://api.github.com/users/hoyyeva/repos", "events_url": "https://api.github.com/users/hoyyeva/events{/privacy}", "received_events_url": "https://api.github.com/users/hoyyeva/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "hoyyeva", "id": 63033505, "node_id": "MDQ6VXNlcjYzMDMzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoyyeva", "html_url": "https://github.com/hoyyeva", "followers_url": "https://api.github.com/users/hoyyeva/followers", "following_url": "https://api.github.com/users/hoyyeva/following{/other_user}", "gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}", "starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions", "organizations_url": "https://api.github.com/users/hoyyeva/orgs", "repos_url": "https://api.github.com/users/hoyyeva/repos", "events_url": "https://api.github.com/users/hoyyeva/events{/privacy}", "received_events_url": "https://api.github.com/users/hoyyeva/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
0
2024-09-06T10:55:18
2024-09-10T21:07:41
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? when upload profile pic , a page to move the pic into o round place .. the canvas doesnt fit into the display , of andro chrome canary ![IMG_20240906_125448_752](https://github.com/user-attachments/assets/56ce834d-575a-4570-9052-3cf683bb2b19) ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6676/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6676/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/4902
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4902/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4902/comments
https://api.github.com/repos/ollama/ollama/issues/4902/events
https://github.com/ollama/ollama/issues/4902
2,339,983,408
I_kwDOJ0Z1Ps6LeVAw
4,902
Performance issue with CPU only inference start 0.1.39 - to latest version of todate.
{ "login": "raymond-infinitecode", "id": 4714784, "node_id": "MDQ6VXNlcjQ3MTQ3ODQ=", "avatar_url": "https://avatars.githubusercontent.com/u/4714784?v=4", "gravatar_id": "", "url": "https://api.github.com/users/raymond-infinitecode", "html_url": "https://github.com/raymond-infinitecode", "followers_url": "https://api.github.com/users/raymond-infinitecode/followers", "following_url": "https://api.github.com/users/raymond-infinitecode/following{/other_user}", "gists_url": "https://api.github.com/users/raymond-infinitecode/gists{/gist_id}", "starred_url": "https://api.github.com/users/raymond-infinitecode/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/raymond-infinitecode/subscriptions", "organizations_url": "https://api.github.com/users/raymond-infinitecode/orgs", "repos_url": "https://api.github.com/users/raymond-infinitecode/repos", "events_url": "https://api.github.com/users/raymond-infinitecode/events{/privacy}", "received_events_url": "https://api.github.com/users/raymond-infinitecode/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5808482718, "node_id": "LA_kwDOJ0Z1Ps8AAAABWjZpng", "url": "https://api.github.com/repos/ollama/ollama/labels/performance", "name": "performance", "color": "A5B5C6", "default": false, "description": "" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
7
2024-06-07T09:17:48
2024-07-03T23:34:02
2024-07-03T23:34:02
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I am running the ollama on intel xeon 32 processors (CPU only) previously which high token generation count using version 0.1.38 However, once I migrate to the latest ollama version 0.1.41, I found that the inference speed for even a model like phi3 on pure CPU slow to a halt. I retest the version and reproducing the slowness start with 0.1.39. Unable to provide log details as there is no error just pure slowness. Didn't change any model nor configuration. Revert back to ver 0.1.38, the performance turn high speed again. Using centos 8 linux Xeon gold processor 32 core ### OS Linux ### GPU Other ### CPU Intel ### Ollama version 0.1.39
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4902/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4902/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6115
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6115/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6115/comments
https://api.github.com/repos/ollama/ollama/issues/6115/events
https://github.com/ollama/ollama/pull/6115
2,441,859,247
PR_kwDOJ0Z1Ps53Fqs2
6,115
Fix context in /api/generate grows too much (#5980).
{ "login": "slouffka", "id": 8129, "node_id": "MDQ6VXNlcjgxMjk=", "avatar_url": "https://avatars.githubusercontent.com/u/8129?v=4", "gravatar_id": "", "url": "https://api.github.com/users/slouffka", "html_url": "https://github.com/slouffka", "followers_url": "https://api.github.com/users/slouffka/followers", "following_url": "https://api.github.com/users/slouffka/following{/other_user}", "gists_url": "https://api.github.com/users/slouffka/gists{/gist_id}", "starred_url": "https://api.github.com/users/slouffka/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/slouffka/subscriptions", "organizations_url": "https://api.github.com/users/slouffka/orgs", "repos_url": "https://api.github.com/users/slouffka/repos", "events_url": "https://api.github.com/users/slouffka/events{/privacy}", "received_events_url": "https://api.github.com/users/slouffka/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
6
2024-08-01T08:47:37
2024-08-01T22:14:00
2024-08-01T22:13:59
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6115", "html_url": "https://github.com/ollama/ollama/pull/6115", "diff_url": "https://github.com/ollama/ollama/pull/6115.diff", "patch_url": "https://github.com/ollama/ollama/pull/6115.patch", "merged_at": "2024-08-01T22:13:59" }
This PR fixes [Context in /api/generate response grows too big. #5980 ](https://github.com/ollama/ollama/issues/5980)
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6115/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6115/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6094
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6094/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6094/comments
https://api.github.com/repos/ollama/ollama/issues/6094/events
https://github.com/ollama/ollama/issues/6094
2,439,598,593
I_kwDOJ0Z1Ps6RaVIB
6,094
"embedding generation failed: do embedding request: Post \"http://127.0.0.1:33967/embedding\": EOF"
{ "login": "yeexiangzhen1001", "id": 70881071, "node_id": "MDQ6VXNlcjcwODgxMDcx", "avatar_url": "https://avatars.githubusercontent.com/u/70881071?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yeexiangzhen1001", "html_url": "https://github.com/yeexiangzhen1001", "followers_url": "https://api.github.com/users/yeexiangzhen1001/followers", "following_url": "https://api.github.com/users/yeexiangzhen1001/following{/other_user}", "gists_url": "https://api.github.com/users/yeexiangzhen1001/gists{/gist_id}", "starred_url": "https://api.github.com/users/yeexiangzhen1001/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yeexiangzhen1001/subscriptions", "organizations_url": "https://api.github.com/users/yeexiangzhen1001/orgs", "repos_url": "https://api.github.com/users/yeexiangzhen1001/repos", "events_url": "https://api.github.com/users/yeexiangzhen1001/events{/privacy}", "received_events_url": "https://api.github.com/users/yeexiangzhen1001/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
19
2024-07-31T09:39:08
2025-01-10T08:14:20
2024-09-02T23:36:51
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? 2024/07/31 09:18:15 routes.go:1099: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:2562047h47m16.854775807s OLLAMA_LLM_LIBRARY: OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://*] OLLAMA_RUNNERS_DIR: OLLAMA_SCHED_SPREAD:false OLLAMA_TMPDIR: ROCR_VISIBLE_DEVICES:]" time=2024-07-31T09:18:16.095Z level=INFO source=images.go:786 msg="total blobs: 2" time=2024-07-31T09:18:16.095Z level=INFO source=images.go:793 msg="total unused blobs removed: 0" time=2024-07-31T09:18:16.095Z level=INFO source=routes.go:1146 msg="Listening on [::]:11434 (version 0.3.1)" time=2024-07-31T09:18:16.095Z level=INFO source=payload.go:30 msg="extracting embedded files" dir=/tmp/ollama37639419/runners time=2024-07-31T09:18:18.739Z level=INFO source=payload.go:44 msg="Dynamic LLM libraries [rocm_v60102 cpu cpu_avx cpu_avx2 cuda_v11]" time=2024-07-31T09:18:18.739Z level=INFO source=gpu.go:205 msg="looking for compatible GPUs" time=2024-07-31T09:18:18.808Z level=INFO source=types.go:105 msg="inference compute" id=GPU-31fa3c8c-f42e-bade-72ec-f936eb48ac45 library=cuda compute=8.6 driver=12.2 name="NVIDIA GeForce RTX 3090 Ti" total="23.7 GiB" available="17.2 GiB" time=2024-07-31T09:20:14.214Z level=INFO source=sched.go:701 msg="new model will fit in available VRAM in single GPU, loading" model=/root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 gpu=GPU-31fa3c8c-f42e-bade-72ec-f936eb48ac45 parallel=4 available=18469158912 required="737.9 MiB" time=2024-07-31T09:20:14.214Z level=INFO source=memory.go:309 msg="offload to cuda" layers.requested=-1 layers.model=13 layers.offload=13 layers.split="" memory.available="[17.2 GiB]" memory.required.full="737.9 MiB" memory.required.partial="737.9 MiB" memory.required.kv="24.0 MiB" memory.required.allocations="[737.9 MiB]" memory.weights.total="186.5 MiB" memory.weights.repeating="155.5 MiB" memory.weights.nonrepeating="30.9 MiB" memory.graph.full="48.0 MiB" memory.graph.partial="48.0 MiB" time=2024-07-31T09:20:14.214Z level=INFO source=server.go:384 msg="starting llama server" cmd="/tmp/ollama37639419/runners/cuda_v11/ollama_llama_server --model /root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 --ctx-size 8192 --batch-size 512 --embedding --log-disable --n-gpu-layers 13 --parallel 4 --port 44985" time=2024-07-31T09:20:14.214Z level=INFO source=sched.go:437 msg="loaded runners" count=1 time=2024-07-31T09:20:14.214Z level=INFO source=server.go:584 msg="waiting for llama runner to start responding" time=2024-07-31T09:20:14.214Z level=INFO source=server.go:618 msg="waiting for server to become available" status="llm server error" INFO [main] build info | build=1 commit="6eeaeba" tid="127422522179584" timestamp=1722417614 INFO [main] system info | n_threads=8 n_threads_batch=-1 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 0 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="127422522179584" timestamp=1722417614 total_threads=16 INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="15" port="44985" tid="127422522179584" timestamp=1722417614 llama_model_loader: loaded meta data with 22 key-value pairs and 197 tensors from /root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = bert llama_model_loader: - kv 1: general.name str = Dmeta-embedding-zh llama_model_loader: - kv 2: bert.block_count u32 = 12 llama_model_loader: - kv 3: bert.context_length u32 = 1024 llama_model_loader: - kv 4: bert.embedding_length u32 = 768 llama_model_loader: - kv 5: bert.feed_forward_length u32 = 3072 llama_model_loader: - kv 6: bert.attention.head_count u32 = 12 llama_model_loader: - kv 7: bert.attention.layer_norm_epsilon f32 = 0.000000 llama_model_loader: - kv 8: general.file_type u32 = 1 llama_model_loader: - kv 9: bert.attention.causal bool = false llama_model_loader: - kv 10: bert.pooling_type u32 = 2 llama_model_loader: - kv 11: tokenizer.ggml.token_type_count u32 = 2 llama_model_loader: - kv 12: tokenizer.ggml.model str = bert llama_model_loader: - kv 13: tokenizer.ggml.tokens arr[str,21128] = ["[PAD]", "[unused1]", "[unused2]", "... llama_model_loader: - kv 14: tokenizer.ggml.token_type arr[i32,21128] = [3, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 15: tokenizer.ggml.unknown_token_id u32 = 100 llama_model_loader: - kv 16: tokenizer.ggml.seperator_token_id u32 = 102 llama_model_loader: - kv 17: tokenizer.ggml.padding_token_id u32 = 0 llama_model_loader: - kv 18: tokenizer.ggml.cls_token_id u32 = 101 llama_model_loader: - kv 19: tokenizer.ggml.mask_token_id u32 = 103 llama_model_loader: - kv 20: tokenizer.ggml.bos_token_id u32 = 0 llama_model_loader: - kv 21: tokenizer.ggml.eos_token_id u32 = 2 llama_model_loader: - type f32: 123 tensors llama_model_loader: - type f16: 74 tensors llm_load_vocab: special tokens cache size = 5 llm_load_vocab: token to piece cache size = 0.0769 MB llm_load_print_meta: format = GGUF V3 (latest) llm_load_print_meta: arch = bert llm_load_print_meta: vocab type = WPM llm_load_print_meta: n_vocab = 21128 llm_load_print_meta: n_merges = 0 llm_load_print_meta: vocab_only = 0 llm_load_print_meta: n_ctx_train = 1024 llm_load_print_meta: n_embd = 768 llm_load_print_meta: n_layer = 12 llm_load_print_meta: n_head = 12 llm_load_print_meta: n_head_kv = 12 llm_load_print_meta: n_rot = 64 llm_load_print_meta: n_swa = 0 llm_load_print_meta: n_embd_head_k = 64 llm_load_print_meta: n_embd_head_v = 64 llm_load_print_meta: n_gqa = 1 llm_load_print_meta: n_embd_k_gqa = 768 llm_load_print_meta: n_embd_v_gqa = 768 llm_load_print_meta: f_norm_eps = 1.0e-12 llm_load_print_meta: f_norm_rms_eps = 0.0e+00 llm_load_print_meta: f_clamp_kqv = 0.0e+00 llm_load_print_meta: f_max_alibi_bias = 0.0e+00 llm_load_print_meta: f_logit_scale = 0.0e+00 llm_load_print_meta: n_ff = 3072 llm_load_print_meta: n_expert = 0 llm_load_print_meta: n_expert_used = 0 llm_load_print_meta: causal attn = 0 llm_load_print_meta: pooling type = 2 llm_load_print_meta: rope type = 2 llm_load_print_meta: rope scaling = linear llm_load_print_meta: freq_base_train = 10000.0 llm_load_print_meta: freq_scale_train = 1 llm_load_print_meta: n_ctx_orig_yarn = 1024 llm_load_print_meta: rope_finetuned = unknown llm_load_print_meta: ssm_d_conv = 0 llm_load_print_meta: ssm_d_inner = 0 llm_load_print_meta: ssm_d_state = 0 llm_load_print_meta: ssm_dt_rank = 0 llm_load_print_meta: model type = 109M llm_load_print_meta: model ftype = F16 llm_load_print_meta: model params = 102.07 M llm_load_print_meta: model size = 194.92 MiB (16.02 BPW) llm_load_print_meta: general.name = Dmeta-embedding-zh llm_load_print_meta: BOS token = 0 '[PAD]' llm_load_print_meta: EOS token = 2 '[unused2]' llm_load_print_meta: UNK token = 100 '[UNK]' llm_load_print_meta: SEP token = 102 '[SEP]' llm_load_print_meta: PAD token = 0 '[PAD]' llm_load_print_meta: CLS token = 101 '[CLS]' llm_load_print_meta: MASK token = 103 '[MASK]' llm_load_print_meta: LF token = 0 '[PAD]' llm_load_print_meta: max token length = 48 ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 1 CUDA devices: Device 0: NVIDIA GeForce RTX 3090 Ti, compute capability 8.6, VMM: yes llm_load_tensors: ggml ctx size = 0.16 MiB llm_load_tensors: offloading 12 repeating layers to GPU llm_load_tensors: offloading non-repeating layers to GPU llm_load_tensors: offloaded 13/13 layers to GPU llm_load_tensors: CPU buffer size = 32.46 MiB llm_load_tensors: CUDA0 buffer size = 162.46 MiB llama_new_context_with_model: n_ctx = 8192 llama_new_context_with_model: n_batch = 512 llama_new_context_with_model: n_ubatch = 512 llama_new_context_with_model: flash_attn = 0 llama_new_context_with_model: freq_base = 10000.0 llama_new_context_with_model: freq_scale = 1 llama_kv_cache_init: CUDA0 KV buffer size = 288.00 MiB llama_new_context_with_model: KV self size = 288.00 MiB, K (f16): 144.00 MiB, V (f16): 144.00 MiB llama_new_context_with_model: CPU output buffer size = 0.00 MiB llama_new_context_with_model: CUDA0 compute buffer size = 19.00 MiB llama_new_context_with_model: CUDA_Host compute buffer size = 4.00 MiB llama_new_context_with_model: graph nodes = 429 llama_new_context_with_model: graph splits = 2 time=2024-07-31T09:20:14.465Z level=INFO source=server.go:618 msg="waiting for server to become available" status="llm server loading model" INFO [main] model loaded | tid="127422522179584" timestamp=1722417614 time=2024-07-31T09:20:14.966Z level=INFO source=server.go:623 msg="llama runner started in 0.75 seconds" [GIN] 2024/07/31 - 09:20:15 | 200 | 862.184786ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:20:15 | 200 | 91.260258ms | 10.234.218.0 | POST "/api/embeddings" time=2024-07-31T09:20:15.383Z level=INFO source=routes.go:426 msg="embedding generation failed: do embedding request: Post \"http://127.0.0.1:44985/embedding\": EOF" [GIN] 2024/07/31 - 09:20:15 | 500 | 140.114654ms | 10.234.218.0 | POST "/api/embeddings" time=2024-07-31T09:23:45.923Z level=WARN source=server.go:503 msg="llama runner process no longer running" sys=139 string="signal: segmentation fault (core dumped)" time=2024-07-31T09:23:50.993Z level=WARN source=sched.go:634 msg="gpu VRAM usage didn't recover within timeout" seconds=5.069197565 model=/root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 time=2024-07-31T09:23:51.075Z level=INFO source=sched.go:701 msg="new model will fit in available VRAM in single GPU, loading" model=/root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 gpu=GPU-31fa3c8c-f42e-bade-72ec-f936eb48ac45 parallel=4 available=18469158912 required="737.9 MiB" time=2024-07-31T09:23:51.075Z level=INFO source=memory.go:309 msg="offload to cuda" layers.requested=-1 layers.model=13 layers.offload=13 layers.split="" memory.available="[17.2 GiB]" memory.required.full="737.9 MiB" memory.required.partial="737.9 MiB" memory.required.kv="24.0 MiB" memory.required.allocations="[737.9 MiB]" memory.weights.total="186.5 MiB" memory.weights.repeating="155.5 MiB" memory.weights.nonrepeating="30.9 MiB" memory.graph.full="48.0 MiB" memory.graph.partial="48.0 MiB" time=2024-07-31T09:23:51.075Z level=INFO source=server.go:384 msg="starting llama server" cmd="/tmp/ollama37639419/runners/cuda_v11/ollama_llama_server --model /root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 --ctx-size 8192 --batch-size 512 --embedding --log-disable --n-gpu-layers 13 --parallel 4 --port 42155" time=2024-07-31T09:23:51.075Z level=INFO source=sched.go:437 msg="loaded runners" count=1 time=2024-07-31T09:23:51.075Z level=INFO source=server.go:584 msg="waiting for llama runner to start responding" time=2024-07-31T09:23:51.076Z level=INFO source=server.go:618 msg="waiting for server to become available" status="llm server error" INFO [main] build info | build=1 commit="6eeaeba" tid="131709034942464" timestamp=1722417831 INFO [main] system info | n_threads=8 n_threads_batch=-1 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 0 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="131709034942464" timestamp=1722417831 total_threads=16 INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="15" port="42155" tid="131709034942464" timestamp=1722417831 llama_model_loader: loaded meta data with 22 key-value pairs and 197 tensors from /root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 (version GGUF V3 (latest)) llama_model_loader: D umping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = bert llama_model_loader: - kv 1: general.name str = Dmeta-embedding-zh llama_model_loader: - kv 2: bert.block_count u32 = 12 llama_model_loader: - kv 3: bert.context_length u32 = 1024 llama_model_loader: - kv 4: bert.embedding_length u32 = 768 llama_model_loader: - kv 5: bert.feed_forward_length u32 = 3072 llama_model_loader: - kv 6: bert.attention.head_count u32 = 12 llama_model_loader: - kv 7: bert.attention.layer_norm_epsilon f32 = 0.000000 llama_model_loader: - kv 8: general.file_type u32 = 1 llama_model_loader: - kv 9: bert.attention.causal bool = false llama_model_loader: - kv 10: bert.pooling_type u32 = 2 llama_model_loader: - kv 11: tokenizer.ggml.token_type_count u32 = 2 llama_model_loader: - kv 12: tokenizer.ggml.model str = bert llama_model_loader: - kv 13: tokenizer.ggml.tokens arr[str,21128] = ["[PAD]", "[unused1]", "[unused2]", "... llama_model_loader: - kv 14: tokenizer.ggml.token_type arr[i32,21128] = [3, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 15: tokenizer.ggml.unknown_token_id u32 = 100 llama_model_loader: - kv 16: tokenizer.ggml.seperator_token_id u32 = 102 llama_model_loader: - kv 17: tokenizer.ggml.padding_token_id u32 = 0 llama_model_loader: - kv 18: tokenizer.ggml.cls_token_id u32 = 101 llama_model_loader: - kv 19: tokenizer.ggml.mask_token_id u32 = 103 llama_model_loader: - kv 20: tokenizer.ggml.bos_token_id u32 = 0 llama_model_loader: - kv 21: tokenizer.ggml.eos_token_id u32 = 2 llama_model_loader: - type f32: 123 tensors llama_model_loader: - type f16: 74 tensors llm_load_vocab: special tokens cache size = 5 llm_load_vocab: token to piece cache size = 0.0769 MB llm_load_print_meta: format = GGUF V3 (latest) llm_load_print_meta: arch = bert llm_load_print_meta: vocab type = WPM llm_load_print_meta: n_vocab = 21128 llm_load_print_meta: n_merges = 0 llm_load_print_meta: vocab_only = 0 llm_load_print_meta: n_ctx_train = 1024 llm_load_print_meta: n_embd = 768 llm_load_print_meta: n_layer = 12 llm_load_print_meta: n_head = 12 llm_load_print_meta: n_head_kv = 12 llm_load_print_meta: n_rot = 64 llm_load_print_meta: n_swa = 0 llm_load_print_meta: n_embd_head_k = 64 llm_load_print_meta: n_embd_head_v = 64 llm_load_print_meta: n_gqa = 1 llm_load_print_meta: n_embd_k_gqa = 768 llm_load_print_meta: n_embd_v_gqa = 768 llm_load_print_meta: f_norm_eps = 1.0e-12 llm_load_print_meta: f_norm_rms_eps = 0.0e+00 llm_load_print_meta: f_clamp_kqv = 0.0e+00 llm_load_print_meta: f_max_alibi_bias = 0.0e+00 llm_load_print_meta: f_logit_scale = 0.0e+00 llm_load_print_meta: n_ff = 3072 llm_load_print_meta: n_expert = 0 llm_load_print_meta: n_expert_used = 0 llm_load_print_meta: causal attn = 0 llm_load_print_meta: pooling type = 2 llm_load_print_meta: rope type = 2 llm_load_print_meta: rope scaling = linear llm_load_print_meta: freq_base_train = 10000.0 llm_load_print_meta: freq_scale_train = 1 llm_load_print_meta: n_ctx_orig_yarn = 1024 llm_load_print_meta: rope_finetuned = unknown llm_load_print_meta: ssm_d_conv = 0 llm_load_print_meta: ssm_d_inner = 0 llm_load_print_meta: ssm_d_state = 0 llm_load_print_meta: ssm_dt_rank = 0 llm_load_print_meta: model type = 109M llm_load_print_meta: model ftype = F16 llm_load_print_meta: model params = 102.07 M llm_load_print_meta: model size = 194.92 MiB (16.02 BPW) llm_load_print_meta: general.name = Dmeta-embedding-zh llm_load_print_meta: BOS token = 0 '[PAD]' llm_load_print_meta: EOS token = 2 '[unused2]' llm_load_print_meta: UNK token = 100 '[UNK]' llm_load_print_meta: SEP token = 102 '[SEP]' llm_load_print_meta: PAD token = 0 '[PAD]' llm_load_print_meta: CLS token = 101 '[CLS]' llm_load_print_meta: MASK token = 103 '[MASK]' llm_load_print_meta: LF token = 0 '[PAD]' llm_load_print_meta: max token length = 48 ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 1 CUDA devices: Device 0: NVIDIA GeForce RTX 3090 Ti, compute capability 8.6, VMM: yes llm_load_tensors: ggml ctx size = 0.16 MiB llm_load_tensors: offloading 12 repeating layers to GPU llm_load_tensors: offloading non-repeating layers to GPU llm_load_tensors: offloaded 13/13 layers to GPU llm_load_tensors: CPU buffer size = 32.46 MiB llm_load_tensors: CUDA0 buffer size = 162.46 MiB llama_new_context_with_model: n_ctx = 8192 llama_new_context_with_model: n_batch = 512 llama_new_context_with_model: n_ubatch = 512 llama_new_context_with_model: flash_attn = 0 llama_new_context_with_model: freq_base = 10000.0 llama_new_context_with_model: freq_scale = 1 llama_kv_cache_init: CUDA0 KV buffer size = 288.00 MiB llama_new_context_with_model: KV self size = 288.00 MiB, K (f16): 144.00 MiB, V (f16): 144.00 MiB llama_new_context_with_model: CPU output buffer size = 0.00 MiB llama_new_context_with_model: CUDA0 compute buffer size = 19.00 MiB llama_new_context_with_model: CUDA_Host compute buffer size = 4.00 MiB llama_new_context_with_model: graph nodes = 429 llama_new_context_with_model: graph splits = 2 time=2024-07-31T09:23:51.243Z level=WARN source=sched.go:634 msg="gpu VRAM usage didn't recover within timeout" seconds=5.319657234 model=/root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 time=2024-07-31T09:23:51.327Z level=INFO source=server.go:618 msg="waiting for server to become available" status="llm server loading model" INFO [main] model loaded | tid="131709034942464" timestamp=1722417831 time=2024-07-31T09:23:51.829Z level=INFO source=server.go:623 msg="llama runner started in 0.75 seconds" [GIN] 2024/07/31 - 09:23:51 | 200 | 5.954027368s | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:51 | 200 | 5.997875851s | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:51 | 200 | 6.001301156s | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:51 | 200 | 6.05401596s | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 6.093406397s | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 6.093515843s | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 141.106871ms | 10.234.218.0 | POST "/api/embeddings" INFO [update_slots] input truncated | n_ctx=2048 n_erase=1989 n_keep=0 n_left=2048 n_shift=1024 tid="131709034942464" timestamp=1722417832 [GIN] 2024/07/31 - 09:23:52 | 200 | 156.396038ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 159.160468ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 155.371305ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 150.237024ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 161.78585ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 158.374292ms | 10.234.218.0 | POST "/api/embeddings" INFO [update_slots] input truncated | n_ctx=2048 n_erase=1517 n_keep=0 n_left=2048 n_shift=1024 tid="131709034942464" timestamp=1722417832 [GIN] 2024/07/31 - 09:23:52 | 200 | 144.427285ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 192.549717ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 131.371235ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 185.844931ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 151.950066ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 141.888776ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 171.173954ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 130.251712ms | 10.234.218.0 | POST "/api/embeddings" INFO [update_slots] input truncated | n_ctx=2048 n_erase=1709 n_keep=0 n_left=2048 n_shift=1024 tid="131709034942464" timestamp=1722417832 [GIN] 2024/07/31 - 09:23:52 | 200 | 140.112505ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 171.12123ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 227.184409ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 264.346952ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 189.302007ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 183.643992ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 165.703255ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 200 | 229.741451ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:52 | 500 | 303.282026ms | 10.234.218.0 | POST "/api/embeddings" time=2024-07-31T09:23:52.825Z level=INFO source=routes.go:426 msg="embedding generation failed: do embedding request: Post \"http://127.0.0.1:42155/embedding\": EOF" time=2024-07-31T09:23:57.889Z level=WARN source=sched.go:634 msg="gpu VRAM usage didn't recover within timeout" seconds=5.063724982 model=/root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 time=2024-07-31T09:23:57.975Z level=INFO source=sched.go:701 msg="new model will fit in available VRAM in single GPU, loading" model=/root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 gpu=GPU-31fa3c8c-f42e-bade-72ec-f936eb48ac45 parallel=4 available=18469158912 required="737.9 MiB" time=2024-07-31T09:23:57.975Z level=INFO source=memory.go:309 msg="offload to cuda" layers.requested=-1 layers.model=13 layers.offload=13 layers.split="" memory.available="[17.2 GiB]" memory.required.full="737.9 MiB" memory.required.partial="737.9 MiB" memory.required.kv="24.0 MiB" memory.required.allocations="[737.9 MiB]" memory.weights.total="186.5 MiB" memory.weights.repeating="155.5 MiB" memory.weights.nonrepeating="30.9 MiB" memory.graph.full="48.0 MiB" memory.graph.partial="48.0 MiB" time=2024-07-31T09:23:57.975Z level=INFO source=server.go:384 msg="starting llama server" cmd="/tmp/ollama37639419/runners/cuda_v11/ollama_llama_server --model /root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 --ctx-size 8192 --batch-size 512 --embedding --log-disable --n-gpu-layers 1 3 --parallel 4 --port 33967" time=2024-07-31T09:23:57.976Z level=INFO source=sched.go:437 msg="loaded runners" count=1 time=2024-07-31T09:23:57.976Z level=INFO source=server.go:584 msg="waiting for llama runner to start responding" time=2024-07-31T09:23:57.976Z level=INFO source=server.go:618 msg="waiting for server to become available" status="llm server error" INFO [main] build info | build=1 commit="6eeaeba" tid="125558191894528" timestamp=1722417837 INFO [main] system info | n_threads=8 n_threads_batch=-1 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 0 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="125558191894528" timestamp=1722417837 total_threads=16 INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="15" port="33967" tid="125558191894528" timestamp=1722417837 llama_model_loader: loaded meta data with 22 key-value pairs and 197 tensors from /root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = bert llama_model_loader: - kv 1: general.name str = Dmeta-embedding-zh llama_model_loader: - kv 2: bert.block_count u32 = 12 llama_model_loader: - kv 3: bert.context_length u32 = 1024 llama_model_loader: - kv 4: bert.embedding_length u32 = 768 llama_model_loader: - kv 5: bert.feed_forward_length u32 = 3072 llama_model_loader: - kv 6: bert.attention.head_count u32 = 12 llama_model_loader: - kv 7: bert.attention.layer_norm _epsilon f32 = 0.000000 llama_model_loader: - kv 8: general.file_type u32 = 1 llama_model_loader: - kv 9: bert.attention.causal bool = false llama_model_loader: - kv 10: bert.pooling_type u32 = 2 llama_model_loader: - kv 11: tokenizer.ggml.token_type_count u32 = 2 llama_model_loader: - kv 12: tokenizer.ggml.model str = bert llama_model_loader: - kv 13: tokenizer.ggml.tokens arr[str,21128] = ["[PAD]", "[unused1]", "[unused2]", "... llama_model_loader: - kv 14: tokenizer.ggml.token_type arr[i32,21128] = [3, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 15: tokenizer.ggml.unknown_token_id u32 = 100 llama_model_loader: - kv 16: tokenizer.ggml.seperator_token_id u32 = 102 llama_model_loader: - kv 17: tokenizer.ggml.padding_token_id u32 = 0 llama_model_loader: - kv 18: tokenizer.ggml.cls_token_id u32 = 101 llama_model_loader: - kv 19: tokenizer.ggml.mask_token_id u32 = 103 llama_model_loader: - kv 20: tokenizer.ggml.bos_token_id u32 = 0 llama_model_loader: - kv 21: tokenizer.ggml.eos_token_id u32 = 2 llama_model_loader: - type f32: 123 tensors llama_model_loader: - type f16: 74 tensors llm_load_vocab: special tokens cache size = 5 llm_load_vocab: token to piece cache size = 0.0769 MB llm_load_print_meta: format = GGUF V3 (latest) llm_load_print_meta: arch = bert llm_load_print_meta: vocab type = WPM llm_load_print_meta: n_vocab = 21128 llm_load_print_meta: n_merges = 0 llm_load_print_meta: vocab_only = 0 llm_load_print_meta: n_ctx_train = 1024 llm_load_print_meta: n_embd = 768 llm_load_print_meta: n_layer = 12 llm_load_print_meta: n_head = 12 llm_load_print_meta: n_head_kv = 12 llm_load_print_meta: n_rot = 64 llm_load_print_meta: n_swa = 0 llm_load_print_meta: n_embd_head_k = 64 llm_load_print_meta: n_embd_head_v = 64 llm_load_print_meta: n_gqa = 1 llm_load_print_meta: n_embd_k_gqa = 768 llm_load_print_meta: n_embd_v_gqa = 768 llm_load_print_meta: f_norm_eps = 1.0e-12 llm_load_print_meta: f_norm_rms_eps = 0.0e+00 llm_load_print_meta: f_clamp_kqv = 0.0e+00 llm_load_print_meta: f_max_alibi_bias = 0.0e+00 llm_load_print_meta: f_logit_scale = 0.0e+00 llm_load_print_meta: n_ff = 3072 llm_load_print_meta: n_expert = 0 llm_load_print_meta: n_expert_used = 0 llm_load_print_meta: causal attn = 0 llm_load_print_meta: pooling type = 2 llm_load_print_meta: rope type = 2 llm_load_print_meta: rope scaling = linear llm_load_print_meta: freq_base_train = 10000.0 llm_load_print_meta: freq_scale_train = 1 llm_load_print_meta: n_ctx_orig_yarn = 1024 llm_load_print_meta: rope_finetuned = unknown llm_load_print_meta: ssm_d_conv = 0 llm_load_print_meta: ssm_d_inner = 0 llm_load_print_meta: ssm_d_state = 0 llm_load_print_meta: ssm_dt_rank = 0 llm_load_print_meta: model type = 109M llm_load_print_meta: model ftype = F16 llm_load_print_meta: model params = 102.07 M llm_load_print_meta: model size = 194.92 MiB (16.02 BPW) llm_load_print_meta: general.name = Dmeta-embedding-zh llm_load_print_meta: BOS token = 0 '[PAD]' llm_load_print_meta: EOS token = 2 '[unused2]' llm_load_print_meta: UNK token = 100 '[UNK]' llm_load_print_meta: SEP token = 102 '[SEP]' llm_load_print_meta: PAD token = 0 '[PAD]' llm_load_print_meta: CLS token = 101 '[CLS]' llm_load_print_meta: MASK token = 103 '[MASK]' llm_load_print_meta: LF token = 0 '[PAD]' llm_load_print_meta: max token length = 48 ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 1 CUDA devices: Device 0: NVIDIA GeForce RTX 3090 Ti, compute capability 8.6, VMM: yes llm_load_tensors: ggml ctx size = 0.16 MiB llm_load_tensors: offloading 12 repeating layers to GPU llm_load_tensors: offloading non-repeating layers to GPU llm_load_tensors: offloaded 13/13 layers to GPU llm_load_tensors: CPU buffer size = 32.46 MiB llm_load_tensors: CUDA0 buffer size = 162.46 MiB llama_new_context_with_model: n_ctx = 8192 llama_new_context_with_model: n_batch = 512 llama_new_context_with_model: n_ubatch = 512 llama_new_context_with_model: flash_attn = 0 llama_new_context_with_model: freq_base = 10000.0 llama_new_context_with_model: freq_scale = 1 llama_kv_cache_init: CUDA0 KV buffer size = 288.00 MiB llama_new_context_with_model: KV self size = 288.00 MiB, K (f16): 144.00 MiB, V (f16): 144.00 MiB llama_new_context_with_model: CPU output buffer size = 0.00 MiB llama_new_context_with_model: CUDA0 compute buffer size = 19.00 MiB llama_new_context_with_model: CUDA_Host compute buffer size = 4.00 MiB llama_new_context_with_model: graph nodes = 429 llama_new_context_with_model: graph splits = 2 time=2024-07-31T09:23:58.139Z level=WARN source=sched.go:634 msg="gpu VRAM usage didn't recover within timeout" seconds=5.312995606 model=/root/.ollama/models/blobs/sha256-9b18b416fe232d5a834e15ce0d6cc353d7f6366423b8a7ef236db9ecee320527 time=2024-07-31T09:23:58.226Z level=INFO source=server.go:618 msg="waiting for server to become available" status="llm server loading model" INFO [main] model loaded | tid="125558191894528" timestamp=1722417838 time=2024-07-31T09:23:58.729Z level=INFO source=server.go:623 msg="llama runner started in 0.75 seconds" [GIN] 2024/07/31 - 09:23:58 | 200 | 6.175518609s | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:58 | 200 | 6.173129645s | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/ 31 - 09:23:58 | 200 | 6.181901759s | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:58 | 200 | 6.217999442s | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:58 | 200 | 6.128390115s | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:58 | 200 | 139.275881ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:58 | 200 | 141.805964ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:58 | 200 | 147.553231ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:59 | 200 | 147.626781ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:59 | 200 | 90.649859ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:59 | 200 | 134.183906ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:59 | 200 | 100.703301ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:59 | 200 | 76.093064ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:59 | 200 | 139.579148ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:59 | 200 | 195.963998ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:59 | 200 | 184.951077ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:59 | 200 | 204.863879ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:59 | 200 | 93.607337ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:59 | 200 | 92.691741ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:59 | 200 | 122.460956ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:23:59 | 200 | 164.876363ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:26:50 | 200 | 93.430143ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:26:50 | 200 | 51.56662ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:26: 50 | 200 | 139.845262ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:26:50 | 200 | 48.229681ms | 10.234.218.0 | POST "/api/embeddings" INFO [update_slots] input truncated | n_ctx=2048 n_erase=1522 n_keep=0 n_left=2048 n_shift=1024 tid="125558191894528" timestamp=1722418010 [GIN] 2024/07/31 - 09:26:50 | 200 | 103.527766ms | 10.234.218.0 | POST "/api/embeddings" [GIN] 2024/07/31 - 09:26:50 | 500 | 138.709641ms | 10.234.218.0 | POST "/api/embeddings" time=2024-07-31T09:26:50.849Z level=INFO source=routes.go:426 msg="embedding generation failed: do embedding request: Post \"http://127.0.0.1:33967/embedding\": EOF" [GIN] 2024/07/31 - 09:37:35 | 200 | 19.4µs | 127.0.0.1 | GET "/api/version" ### OS Docker ### GPU Nvidia ### CPU Intel ### Ollama version 0.3.1
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6094/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 1, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6094/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6743
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6743/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6743/comments
https://api.github.com/repos/ollama/ollama/issues/6743/events
https://github.com/ollama/ollama/pull/6743
2,518,666,601
PR_kwDOJ0Z1Ps57GpZP
6,743
Fixed no redirect URL scenario when downloading blobs
{ "login": "JingWoo", "id": 21989093, "node_id": "MDQ6VXNlcjIxOTg5MDkz", "avatar_url": "https://avatars.githubusercontent.com/u/21989093?v=4", "gravatar_id": "", "url": "https://api.github.com/users/JingWoo", "html_url": "https://github.com/JingWoo", "followers_url": "https://api.github.com/users/JingWoo/followers", "following_url": "https://api.github.com/users/JingWoo/following{/other_user}", "gists_url": "https://api.github.com/users/JingWoo/gists{/gist_id}", "starred_url": "https://api.github.com/users/JingWoo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JingWoo/subscriptions", "organizations_url": "https://api.github.com/users/JingWoo/orgs", "repos_url": "https://api.github.com/users/JingWoo/repos", "events_url": "https://api.github.com/users/JingWoo/events{/privacy}", "received_events_url": "https://api.github.com/users/JingWoo/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
1
2024-09-11T06:34:01
2024-09-30T09:08:37
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6743", "html_url": "https://github.com/ollama/ollama/pull/6743", "diff_url": "https://github.com/ollama/ollama/pull/6743.diff", "patch_url": "https://github.com/ollama/ollama/pull/6743.patch", "merged_at": null }
null
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6743/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6743/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1312
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1312/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1312/comments
https://api.github.com/repos/ollama/ollama/issues/1312/events
https://github.com/ollama/ollama/issues/1312
2,016,449,283
I_kwDOJ0Z1Ps54MJMD
1,312
trouble with deepseek-coder
{ "login": "niknoproblems", "id": 3484515, "node_id": "MDQ6VXNlcjM0ODQ1MTU=", "avatar_url": "https://avatars.githubusercontent.com/u/3484515?v=4", "gravatar_id": "", "url": "https://api.github.com/users/niknoproblems", "html_url": "https://github.com/niknoproblems", "followers_url": "https://api.github.com/users/niknoproblems/followers", "following_url": "https://api.github.com/users/niknoproblems/following{/other_user}", "gists_url": "https://api.github.com/users/niknoproblems/gists{/gist_id}", "starred_url": "https://api.github.com/users/niknoproblems/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/niknoproblems/subscriptions", "organizations_url": "https://api.github.com/users/niknoproblems/orgs", "repos_url": "https://api.github.com/users/niknoproblems/repos", "events_url": "https://api.github.com/users/niknoproblems/events{/privacy}", "received_events_url": "https://api.github.com/users/niknoproblems/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
4
2023-11-29T12:10:21
2024-03-12T01:15:06
2024-03-12T01:15:05
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I'm having trouble getting this model to run on mac m1 16gb ram: ollama run deepseek-coder:6.7b-base-q8_0 but this model work without any troubles: ollama run neural-chat:7b-v3.1-q8_0 it has more weights and bigger file size.
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1312/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1312/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2529
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2529/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2529/comments
https://api.github.com/repos/ollama/ollama/issues/2529/events
https://github.com/ollama/ollama/issues/2529
2,137,653,184
I_kwDOJ0Z1Ps5_af_A
2,529
Ollama Windows is much slower at inference than Ollama on WSL2
{ "login": "devinprater", "id": 15256014, "node_id": "MDQ6VXNlcjE1MjU2MDE0", "avatar_url": "https://avatars.githubusercontent.com/u/15256014?v=4", "gravatar_id": "", "url": "https://api.github.com/users/devinprater", "html_url": "https://github.com/devinprater", "followers_url": "https://api.github.com/users/devinprater/followers", "following_url": "https://api.github.com/users/devinprater/following{/other_user}", "gists_url": "https://api.github.com/users/devinprater/gists{/gist_id}", "starred_url": "https://api.github.com/users/devinprater/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/devinprater/subscriptions", "organizations_url": "https://api.github.com/users/devinprater/orgs", "repos_url": "https://api.github.com/users/devinprater/repos", "events_url": "https://api.github.com/users/devinprater/events{/privacy}", "received_events_url": "https://api.github.com/users/devinprater/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" } ]
closed
false
null
[]
null
6
2024-02-16T00:18:17
2024-02-21T09:25:21
2024-02-19T21:23:33
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
CPU: AMD 5500U with Radion internal GPU. Ollama runs on CPU mode on both WSL2 and Windows. Attached are the logs from Windows, and Linux. [server.log](https://github.com/ollama/ollama/files/14303692/server.log) [ollama-log-linux.log](https://github.com/ollama/ollama/files/14303696/ollama-log-linux.log)
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2529/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2529/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3068
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3068/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3068/comments
https://api.github.com/repos/ollama/ollama/issues/3068/events
https://github.com/ollama/ollama/pull/3068
2,180,371,586
PR_kwDOJ0Z1Ps5pT6Cw
3,068
Use stdin for term discovery on windows
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-03-11T22:28:55
2024-03-14T18:55:22
2024-03-14T18:55:19
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3068", "html_url": "https://github.com/ollama/ollama/pull/3068", "diff_url": "https://github.com/ollama/ollama/pull/3068.diff", "patch_url": "https://github.com/ollama/ollama/pull/3068.patch", "merged_at": "2024-03-14T18:55:19" }
When you feed input to the cmd via a pipe it no longer reports a warning Before: ``` > echo "what is the captial of australia" | .\ollama.exe run phi failed to get console mode for stdin: The handle is invalid. The capital of Australia is Canberra. It's located in the Australian Capital Territory, about 120 kilometers northwest of Sydney and 130 kilometers southwest of Melbourne. ``` After fix: ``` > echo "what is the captial of australia" | .\ollama.exe run phi The capital city of Australia is Canberra. It is located in the Australian Capital Territory (ACT) and is home to important government buildings, such as Parliament House and the National Museum of Australia ``` Fixes #2698
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3068/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3068/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5146
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5146/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5146/comments
https://api.github.com/repos/ollama/ollama/issues/5146/events
https://github.com/ollama/ollama/pull/5146
2,362,711,736
PR_kwDOJ0Z1Ps5y-1ig
5,146
Put back temporary intel GPU env var
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-06-19T15:58:50
2024-06-19T16:12:48
2024-06-19T16:12:45
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5146", "html_url": "https://github.com/ollama/ollama/pull/5146", "diff_url": "https://github.com/ollama/ollama/pull/5146.diff", "patch_url": "https://github.com/ollama/ollama/pull/5146.patch", "merged_at": "2024-06-19T16:12:45" }
Until we merge #4876 lets keep the opt-in env var to avoid confusion in the binary releases if we discover an Intel GPU but don't actually have the runner built in. This reverts commit 755b4e4fc291366595ed7bfb37c2a91ff5834df8.
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5146/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5146/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4219
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4219/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4219/comments
https://api.github.com/repos/ollama/ollama/issues/4219/events
https://github.com/ollama/ollama/issues/4219
2,282,184,189
I_kwDOJ0Z1Ps6IB139
4,219
模型整理 - Categorize models on ollama.com
{ "login": "syssbs", "id": 129733386, "node_id": "U_kgDOB7uTCg", "avatar_url": "https://avatars.githubusercontent.com/u/129733386?v=4", "gravatar_id": "", "url": "https://api.github.com/users/syssbs", "html_url": "https://github.com/syssbs", "followers_url": "https://api.github.com/users/syssbs/followers", "following_url": "https://api.github.com/users/syssbs/following{/other_user}", "gists_url": "https://api.github.com/users/syssbs/gists{/gist_id}", "starred_url": "https://api.github.com/users/syssbs/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/syssbs/subscriptions", "organizations_url": "https://api.github.com/users/syssbs/orgs", "repos_url": "https://api.github.com/users/syssbs/repos", "events_url": "https://api.github.com/users/syssbs/events{/privacy}", "received_events_url": "https://api.github.com/users/syssbs/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 6573197867, "node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw", "url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com", "name": "ollama.com", "color": "ffffff", "default": false, "description": "" } ]
open
false
null
[]
null
3
2024-05-07T03:16:01
2024-07-25T18:15:27
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
麻烦把官网的模型进行分类整理下吧,模型太多了以后也会越来越多,官网的模型列表感觉会很乱
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4219/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4219/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/8307
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8307/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8307/comments
https://api.github.com/repos/ollama/ollama/issues/8307/events
https://github.com/ollama/ollama/pull/8307
2,769,004,347
PR_kwDOJ0Z1Ps6GvOdm
8,307
fix: correct endpoint URL to avoid 404 error
{ "login": "ubaldus", "id": 660076, "node_id": "MDQ6VXNlcjY2MDA3Ng==", "avatar_url": "https://avatars.githubusercontent.com/u/660076?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ubaldus", "html_url": "https://github.com/ubaldus", "followers_url": "https://api.github.com/users/ubaldus/followers", "following_url": "https://api.github.com/users/ubaldus/following{/other_user}", "gists_url": "https://api.github.com/users/ubaldus/gists{/gist_id}", "starred_url": "https://api.github.com/users/ubaldus/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ubaldus/subscriptions", "organizations_url": "https://api.github.com/users/ubaldus/orgs", "repos_url": "https://api.github.com/users/ubaldus/repos", "events_url": "https://api.github.com/users/ubaldus/events{/privacy}", "received_events_url": "https://api.github.com/users/ubaldus/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2025-01-04T21:14:22
2025-01-04T23:45:16
2025-01-04T23:45:16
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8307", "html_url": "https://github.com/ollama/ollama/pull/8307", "diff_url": "https://github.com/ollama/ollama/pull/8307.diff", "patch_url": "https://github.com/ollama/ollama/pull/8307.patch", "merged_at": "2025-01-04T23:45:16" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8307/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8307/timeline
null
null
true