omg german and AI ... sry that i must say that and iam german
#11
by
kalle07
- opened
first
only 4096 token lenght at max and you talk about RAG?
that are only 4 pages, so useless !?!? that was 3 years ago with lama3 ...
second you have no template prepared (in tokenizer_config.json) for all open software that use that gpt4all, JAN, LLM-Studido !!!
example
https://huggingface.co/ibm-granite/granite-3.3-2b-instruct/blob/main/tokenizer_config.json
last view lines
so the basics are not promissing . . .
for me this is more talkelative template:
{%- for message in messages %}
{%- if message['role'] == 'system' and message['content'] %}
System: {{ message['content']|trim }}
{%- elif message['role'] == 'user_context' and message['content'] %}
### Input:
{{ message['content']|trim }}
{%- elif message['role'] == 'user' and message['content'] %}
### Instruction:
{{ message['content']|trim }}
{%- elif message['role'] == 'assistant' and message['content'] %}
### Response:
{{ message['content']|trim }}{{ eos_token }}
{%- endif %}
{%- if not loop.last %}
{%- endif %}
{%- endfor %}
{%- if add_generation_prompt and messages[-1]['role'] != 'assistant' %}
### Response:
{%- endif %}