Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up
poe's picture
6 1 193

poe

wop
andyolivers's profile picture
·

AI & ML interests

https://huggingface.co/CreativityCompound

Recent Activity

reacted to eaddario's post with ❤️ about 22 hours ago
Experimental global target bits‑per‑weight quantization of ServiceNow-AI/Apriel-1.6-15b-Thinker and zai-org/GLM-4.6V-Flash Unlike standard llama.cpp quantizations that rely on fixed type heuristics (e.g., Q4_K_M), the Target BPW approach optimizes per-tensor precision where it matters the most, and produces high quality models that meet a precise global file size target. Key Advantages: - VRAM Maximization: Can generate high quality models sized exactly to fit hardware constraints (e.g., fitting the model into exactly 24GB VRAM). - Data-Driven Precision: Quantization mix is determined by actual weight error sensitivity rather than hardcoded rules, often yielding better PPL/KLD size trade-offs. Full benchmarks (PPL, KLD, ARC, MMLU, etc.) and methodology in the models' cards https://huggingface.co/eaddario/Apriel-1.6-15b-Thinker-GGUF https://huggingface.co/eaddario/GLM-4.6V-Flash-GGUF
updated a model 9 days ago
wop/arptron-beta-5
liked a Space 9 days ago
sitammeur/Gemma-3-WebGPU
View all activity

Organizations

CreativityCompound's profile picture

wop 's datasets 2

wop/Romanian_Language

Viewer • Updated Sep 23, 2024 • 1 • 21

wop/discord_chats_1709

Viewer • Updated Dec 8, 2023 • 1.71k • 26
Company
TOS Privacy About Careers
Website
Models Datasets Spaces Pricing Docs