Alexandre Marques
alexmarques
AI & ML interests
None yet
Recent Activity
updated
a model
1 day ago
RedHatAI/Qwen3-30B-A3B-quantized.w4a16
published
a model
1 day ago
RedHatAI/Qwen3-30B-A3B-quantized.w4a16
Organizations
alexmarques's activity
What is the difference between Qwen/Qwen3-32B-FP8 and this quatinized model?
4
#1 opened 5 days ago
by
traphix
How can I repeat the eval results?
3
#2 opened 2 days ago
by
bash99
Where are the safetensors?
1
1
#1 opened 2 days ago
by
traphix
Is this a QAT model?
1
#2 opened 9 days ago
by
Downtown-Case
Having trouble running this model with vLLM not sure why
2
#1 opened 19 days ago
by
zacksiri

Add chat_template to tokenizer_config
1
2
#58 opened about 1 month ago
by
alexmarques
KV Cache Quantization - what is the default precision
1
#2 opened 5 months ago
by
deepmage121
Usage of --apply_chat_template in lm_eval benchmarks
1
#1 opened 9 months ago
by
VlSav
how many resources were used for quantizing this model?
1
#4 opened 8 months ago
by
fengyang1995
4k or 128k?
1
#1 opened 8 months ago
by
pavidu
Update README.md
#1 opened 9 months ago
by
nm-research
Llama-3.1 8B quantization?
1
#1 opened 10 months ago
by
ashbo
Language Support
1
#2 opened 10 months ago
by
ashbo
possible problem in description
2
#3 opened 9 months ago
by
kuliev-vitaly
Weird variable name mistakes in code generation
1
#1 opened 10 months ago
by
krana