Spaces:
Running
Running
Commit History
common : handle new quant types (ggml/0) 53bb541
ggml : add ggml-aarch64 (ggml/0) 0062819
Dibakar Gope commited on
ggml : reduce hash table reset cost (llama/8698) 9808fbf
slaren commited on
ggml: handle ggml_init failure to fix NULL pointer deref (llama/8692) dc51517
DavidKorczynski commited on
fix multi-gpu issue on sycl (llama/8554) 94a6436
ggml : add and use ggml_cpu_has_llamafile() (llama/8664) efcca56
Re-add erroneously removed -fsycl from GGML_EXTRA_LIBS (llama/8667) 6e12dfd
Joe Todd commited on
sycl : Add support for non-release DPC++ & oneMKL (llama/8644) 2a5814c
Joe Todd commited on
Vulkan IQ4_NL Support (llama/8613) 899145d
Allow all RDNA2 archs to use sdot4 intrinsic (llama/8629) 1d65fea
Jeroen Mostert commited on
fix scratch size of softmax (llama/8642) 6519fd2
ggml: fix compile error for RISC-V (llama/8623) 4eec44b
Mark Zhuang commited on
CUDA: MMQ code deduplication + iquant support (llama/8495) 6d14124
gguf : handle null name during init (llama/8587) 2f95156
ggml : fix quant dot product with odd number of blocks (llama/8549) 0083f96
ggml : add friendlier error message to fopen errors (llama/8575) ab5b4e0
CUDA: fix partial offloading for ne0 % 256 != 0 (llama/8572) afc137c
cmake : install all ggml public headers (llama/8480) 73a16f3
65a 65a commited on
Add Ascend NPU backend (llama/6035) 3175a17
make/cmake: add missing force MMQ/cuBLAS for HIP (llama/8515) 5096c91
Refactor lora adapter support (llama/8332) 76bcfc6
add concat through dim 1/2 (llama/8483) acf23d9
Vulkan MMQ Fix (llama/8479) e2989d0
vulkan : cmake integration (llama/8119) a094e22
bandoti commited on
metal : template-ify some of the kernels (llama/8447) 3c3094f
ggml : minor naming changes (llama/8433) e0c6dff
ggml : add NVPL BLAS support (ggml/8329) (llama/8425) 4816a87
cuda : suppress 'noreturn' warn in no_device_code (llama/8414) 13c1163
CUDA: optimize and refactor MMQ (llama/8416) a3fe534
Use multi_ptr to clean up deprecated warnings (llama/8256) 6dbe297
AidanBeltonS commited on
ggml : move sgemm sources to llamafile subfolder (llama/8394) 1554348
ggml : add AArch64 optimized GEMV and GEMM Q4 kernels (llama/5780) 9509586
Dibakar Gope commited on
sycl : Reenabled mmvq path for the SYCL Nvidia Backend (llama/8372) b969571
Alberto Cabrera Pérez commited on
sycl : fix powf call in device code (llama/8368) 011fbfd
Alberto Cabrera Pérez commited on
ggml : loop tiling optimizations for scalar path (ggml/898) 1c4b0ca
Mahesh Madhav commited on
ggml: add support for float16 input tensors in pooling operations (ggml/895) 8248d8e
Ivan Filipov vanaka11 commited on
vulkan : initialize vk_buffer_struct members to VK_NULL_HANDLE (ggml/893) 8c409e3
Tony Wasserka Tony Wasserka commited on
cmake : only enable GGML_NATIVE and x86 flags if not crosscompiling (ggml/885) 0456299
scripts : sync new files (#0) 49609b6
cmake : fix compile in xcode (#2311) c0c60f1 unverified
Daven Sanassy commited on
whisper : handle empty mel (#2324) 1434bfe unverified
whisper : use vulkan as gpu backend when available (#2302) 0755fa0 unverified
Matt Stephenson commited on
whisper : fix DTW assert (#2299) 1d1d815 unverified
arizhih commited on