Josh Warner
JDWarner
AI & ML interests
None yet
Recent Activity
new activity 4 days ago
AEON-7/Qwen3.6-27B-AEON-Ultimate-Uncensored-BF16:Is MTP possible? liked a model 7 days ago
rdtand/Qwen3.6-27B-PrismaQuant-5.5bit-vllm liked a model 7 days ago
GestaltLabs/Ornstein-3.6-27BOrganizations
None yet
Is MTP possible?
2
#2 opened 4 days ago
by
JDWarner
Scaling with concurrency?
2
#1 opened 13 days ago
by
JDWarner
dflash with quantize model
1
#5 opened 17 days ago
by
Shimon324
FP8 work for base model or is 16-bit of 27B required?
14
#2 opened 29 days ago
by
unoid
pruned version
๐๐ฅ 1
2
#16 opened about 1 month ago
by
pirola
There's got to be a better way.
23
#6 opened about 1 month ago
by
phil111
Recall from embed documents not as good as the original
5
#4 opened about 1 month ago
by
o0Linny0o
A wild idea / suggestion...
๐ฅ 3
2
#4 opened about 2 months ago
by
MrDevolver
Consider releasing full BF16 weights
2
#1 opened about 2 months ago
by
JDWarner
good model
5
#1 opened about 2 months ago
by
Roman1111111
Work great on 3090 except for weird (...) generation
โค๏ธ 1
6
#1 opened about 2 months ago
by
ortegaalfredo
Qwopus with visual capabilities?
2
#19 opened about 2 months ago
by
AQLabs
Security/Compliance Audit: EU AI Act & NIST Exposure
๐ฅ 1
3
#8 opened about 2 months ago
by
tradeapollo
FP8 models
3
#1 opened about 2 months ago
by
ecopoiesis
IQ5_K 136.891 GiB
๐ฅ 2
30
#9 opened 3 months ago
by
Hunterx
Request: GGUF / quantized weights for Intern-S1-Pro
1
#7 opened 3 months ago
by
gileneo
INT8 quantization for KVCache on DGX Spark/GB10
4
#6 opened 3 months ago
by
JDWarner
This just trades general performance for domain specific gains.
๐ฅ๐ 16
11
#3 opened 8 months ago
by
phil111
Disable thinking mode in Jan-v1-4B model
2
#9 opened 8 months ago
by
vuhaix95