Tom
TomLucidor
AI & ML interests
None yet
Recent Activity
new activity 1 day ago
z-lab/gemma-4-26B-A4B-it-DFlash:This works with a quantsized version? new activity 1 day ago
z-lab/Qwen3.6-35B-A3B-DFlash:Acceptance rate drop in non-thinking scenariosOrganizations
This works with a quantsized version?
2
#4 opened 3 days ago
by
khronnuz
Acceptance rate drop in non-thinking scenarios
👍 1
1
#7 opened 17 days ago
by
lillilyh
DFlash compatibility with AutoRound (W4A16) quantized Qwen3.5-9B in vLLM?
2
#1 opened about 2 months ago
by
Vishva007
Is have Qwen3.6 35B A3B FP8 version?
➕ 1
2
#4 opened 18 days ago
by
whaozl
Qwen3.5 2B/0.5B
1
#1 opened 28 days ago
by
packetsent
30+ Billion Models?
🚀 3
1
#2 opened 22 days ago
by
bobbytaylor
Will there be Mini/Nano models as well?
#2 opened 11 days ago
by
TomLucidor
Open-weights only toggle for Web UI
👍 1
#4 opened about 2 months ago
by
TomLucidor
Sort by Pass@5 % on web UI
👍 1
#3 opened about 2 months ago
by
TomLucidor
Can you add in Qwen3.5 and other series of models for testing?
3
#2 opened about 2 months ago
by
TomLucidor
Some questions on BitNet PTQ
2
#1 opened 3 months ago
by
TomLucidor
Will there be a base model?
2
#4 opened 3 months ago
by
zianglih
Has REAM been checked for their resilience towards quantization?
4
#1 opened 3 months ago
by
TomLucidor
How is this different from the other quants?
8
#1 opened 2 months ago
by
TomLucidor
Will Hybrid Attention RP models get some love?
➕ 1
#7 opened 2 months ago
by
TomLucidor
ValueError: Model type lfm2_moe not supported.
5
#1 opened 7 months ago
by
kadirnar
Runs Amazing on M4 Macbook air!
1
#1 opened 6 months ago
by
leo253
Could more benchmarks be done on Instruction Following / Function Calling?
4
#2 opened 3 months ago
by
TomLucidor
The comparison with the original MTP
👍 1
1
#2 opened 3 months ago
by
Michalea
Using ZwZ and better VLM along side DeepGen
4
#6 opened 3 months ago
by
TomLucidor