ggml : fix fallback to CPU for ununsupported ops (llama/15118) 2b7ae5e Diego Devesa commited on Aug 6
ggml-cpu : split arch-specific implementations (llama/13892) 8c833e9 xctan ggerganov commited on Jun 9
cpu: fix cpu backend's supports-op for GET_ROWS_BACK. fixes a fatal when running test-backend-ops with only the CPU backend (ggml/1190) ee7706c cmdr2 commited on Apr 11
ggml-cpu: Faster IQ1 mul_mat_vec on AVX2 using BMI2 instructions (llama/12154) 05466a9 Rémy O commited on Mar 6
ggml-cpu: Support s390x SIMD Instruction Set (llama/12019) 4aa54ec Aaron Teo Jinyang He junchao-zhao commited on Feb 22
ggml-cpu: Add CPU backend support for KleidiAI library (llama/11390) 9de6d81 Charles Xu commited on Feb 20
Fix #11802: Compile bug - RegQueryValueExA changed to RegQueryValueEx (llama/11803) 86969ac Sheldon Robinson commited on Feb 11
CPU/CUDA: fix (GQA) mul mat back, add CUDA support (llama/11380) 855a9fe JohannesGaessler commited on Jan 24
CUDA: backwards pass for misc. ops, add tests (llama/11257) 2fbcec1 JohannesGaessler commited on Jan 16
RoPE: fix back, CUDA support for back + noncont. (llama/11240) 131a21e JohannesGaessler commited on Jan 15
ggml : update ggml_backend_cpu_device_supports_op (llama/10867) 2f11d1e ggerganov commited on Dec 17, 2024
ggml : add predefined list of CPU backend variants to build (llama/10626) 1794b43 Diego Devesa commited on Dec 4, 2024
ggml-cpu: support IQ4_NL_4_4 by runtime repack (llama/10541) bf73242 shupeif commited on Nov 28, 2024
ggml : add support for dynamic loading of backends (llama/10469) b73266f Diego Devesa ggerganov commited on Nov 25, 2024
backend cpu: add online flow for aarch64 Q4_0 GEMV/GEMM kernels (llama/9921) 3541ee8 Charles Xu Diego Devesa commited on Nov 15, 2024
ggml : build backends as libraries (llama/10256) 3dc93f3 Diego Devesa ggerganov R0CKSTAR commited on Nov 14, 2024