xldistance
xldistance
AI & ML interests
None yet
Recent Activity
new activity
1 day ago
Qwen/QwQ-32B:When will you fix the model replies missing</think>\n start tags
new activity
4 days ago
Qwen/QwQ-32B:missing opening <think>
new activity
6 days ago
Qwen/Qwen2.5-VL-72B-Instruct:Anyone pls let me know what hardware can run 72B ?
Organizations
None yet
xldistance's activity
When will you fix the model replies missing</think>\n start tags
13
#19 opened 4 days ago
by
xldistance
missing opening <think>
17
#4 opened 4 days ago
by
getfit

Anyone pls let me know what hardware can run 72B ?
2
#15 opened 13 days ago
by
haoyiharrison

Can you train this model for 4.5bpw quantization?
2
#1 opened 8 days ago
by
xldistance
Can you train this model for 4.5bpw quantization?
#1 opened 8 days ago
by
xldistance
Can you produce a 4.5bpw quantized model of this model?
2
#1 opened 10 days ago
by
xldistance
Repeated Thinking Tags in Output Generation
10
#2 opened 12 days ago
by
xldistance
Can you distill qwen-2.5-72b?
1
#30 opened about 1 month ago
by
xldistance
This model removes the limitations but the ability to write code decreases a lot.
#3 opened about 1 month ago
by
xldistance
Can you quantify the 4.0bpw weight of this model
1
#2 opened about 1 month ago
by
xldistance
This model is poorly coded and suspected of being a list swipe
1
#3 opened about 1 month ago
by
xldistance
Can you provide a 5.5bpw quantization of this model?
#2 opened about 2 months ago
by
xldistance
Model loading failure
2
#1 opened about 2 months ago
by
xldistance
Can you produce a quantized 2.4bpw model of this model?
3
#1 opened 3 months ago
by
xldistance
Phi-4 = gpt-4o-mini
6
#4 opened 3 months ago
by
maxbn
Can you produce a 2.4bpw quantization of this model?
2
#1 opened 3 months ago
by
xldistance
How to reduce the problem of 2.25bpw quantitative models often responding haphazardly
1
#2 opened 4 months ago
by
xldistance
Can you make a 2.25bpw quantization for this model?
#4 opened 3 months ago
by
xldistance
Can you use the same method to train the qwen2.5 32b model?
8
#24 opened 3 months ago
by
xldistance
The model can go off on tangents
1
#4 opened 3 months ago
by
spanspek