JJ
J22
AI & ML interests
None yet
Organizations
None yet
J22's activity
Upload tokenizer.json
1
#1 opened 8 days ago
by
J22
a horrible function in `modeling_mobilellm.py`
1
#5 opened 9 days ago
by
J22
Run this on CPU
#6 opened about 2 months ago
by
J22
Run on CPU
1
#13 opened about 2 months ago
by
J22
need gguf
18
#4 opened 3 months ago
by
windkkk
Best practice for tool calling with meta-llama/Meta-Llama-3.1-8B-Instruct
1
#33 opened 4 months ago
by
zzclynn
Run this on CPU and use tool calling
1
#38 opened 4 months ago
by
J22
My alternative quantizations.
5
#5 opened 4 months ago
by
ZeroWw
Tool calling is supported by ChatLLM.cpp
#36 opened 5 months ago
by
J22
can't say hello
1
#9 opened 6 months ago
by
J22
no system message?
8
#14 opened 6 months ago
by
mclassHF2023
"small" is so different from "mini" and "medium"
1
#8 opened 6 months ago
by
J22
how to set context in multi-turn QA?
6
#14 opened 6 months ago
by
J22
clarification on the usage of `short_factor` and `long_factor`?
1
#49 opened 6 months ago
by
J22
Continue the discussion: `long_factor` and `short_factor`
2
#32 opened 7 months ago
by
J22
is the '\n' after `'<|end|>'`?
1
#43 opened 7 months ago
by
J22
Is sliding window used or not?
1
#25 opened 7 months ago
by
J22
`long_factor` is never used?
2
#22 opened 7 months ago
by
J22
generate +6 min, +20GB V-ram
2
#17 opened 7 months ago
by
NickyNicky
`sliding_window` is larger than `max_position_embeddings`
1
#21 opened 7 months ago
by
J22