Moonshine Tiny -- GGUF

GGUF conversions and quantisations of UsefulSensors/moonshine-tiny for use with CrispStrobe/CrispASR.

Available variants

File Quant Size Notes
moonshine-tiny.gguf F32 104 MB Full precision
moonshine-tiny-q8_0.gguf Q8_0 33 MB High quality
moonshine-tiny-q4_k.gguf Q4_K 21 MB Best size/quality tradeoff

All variants produce correct transcription on test audio.

Model details

  • Architecture: Conv1d stem + 6L transformer encoder + 6L transformer decoder (288d, 8 heads, partial RoPE, SiLU/GELU)
  • Parameters: 27M
  • Languages: English only
  • WER: 4.55% (LibriSpeech clean), 11.68% (Other)
  • Performance: 11.2x realtime on CPU (F32)
  • License: MIT
  • Source: moonshine.cpp (MIT)

Usage with CrispASR

./build/bin/crispasr -m moonshine-tiny-q4_k.gguf -f audio.wav
./build/bin/crispasr --backend moonshine -m moonshine-tiny-q4_k.gguf -f audio.wav -osrt
Downloads last month
159
GGUF
Model size
27.1M params
Architecture
moonshine
Hardware compatibility
Log In to add your hardware

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for cstr/moonshine-tiny-GGUF

Quantized
(3)
this model