cicdatopea
cicdatopea
AI & ML interests
None yet
Recent Activity
updated
a model
about 9 hours ago
cicdatopea/small_llam4_moverick_test
published
a model
about 10 hours ago
cicdatopea/small_llam4_moverick_test
updated
a model
about 16 hours ago
cicdatopea/Llama-4-Maverick-17B-128E-Instruct_int8
Organizations
cicdatopea's activity
Qwen2.5-VL-72B-Instruct quant?
1
#1 opened 11 days ago
by
samgreen

VLLM 0.7.2 can start the model normally, but there is no output when simulating a request using Curl, it blocks!
3
#2 opened about 2 months ago
by
JZMALi
Could we get a int3 version of gptq please?
1
#1 opened 15 days ago
by
davidsyoung
Qwen-32B overflow issue
8
#1 opened 25 days ago
by
cicdatopea

how to run this model
4
#1 opened 18 days ago
by
cicdatopea

without licence
1
#2 opened 22 days ago
by
Futureli
how to inference this model?
1
#1 opened 25 days ago
by
xiximayou
so consider build a model for GPU?
1
#1 opened 26 days ago
by
kq

Your quants are not listed in the base model
2
#2 opened about 2 months ago
by
dazipe
sglang inference issue
7
#1 opened about 2 months ago
by
su400
Start on cpu with vllm.
1
#1 opened 2 months ago
by
kuliev-vitaly
“a larger accuracy drop in Chinese tasks"? how much exectaly?
1
#1 opened 3 months ago
by
chuangzhidian
A bug when running the demo inference on GPU
1
#5 opened 3 months ago
by
HuggingLianWang
vllm
23
#4 opened 3 months ago
by
NikolaSigmoid
Base model please!
2
#2 opened 3 months ago
by
deltanym

alternative serving framework
2
#1 opened 4 months ago
by
erichartford

Update README.md
#1 opened 4 months ago
by
n1ck-guo