Spaces:
Sleeping
Sleeping
Commit History
ggml : Enable MMA for BF16 in llamafile_sgemm (llama/13148)
7da5bcc
rpc : avoid uninitialized memory in serialize_tensor (llama/13210)
31cad24
Justin Santa Barbara
commited on
ggml: Don't assert fail when tensor data changes (llama/13222)
af16d74
Jesse Gross
commited on
build : fix build info on windows (llama/13239)
415b9fc
Diego Devesa
commited on
vulkan: Add bfloat16 support (llama/12554)
b21f8a1
vulkan: Handle src1 batch dimension in non-contiguous mat-vec-mul shader (llama/13191)
710fdcf
vulkan : kernels for depthwise 2D convolution (CONV_2D_DW) (ggml/1204)
43d9f3e
ci : zip windows artifacts for release uploading (#3124)
3dbef6c
unverified
ci : add zip extension to xcframework artifact name (#3120)
a8a2519
unverified
whisper: remove MSVC warnings pragmas (#3090)
e0d130c
unverified
server: update abort mechanism to handle HTTP connection closure (#3112)
02b25fa
unverified
cli : support "-" for stdout like stdin (#3050)
7e3c27c
unverified
Daniel Tang
commited on
docs : Update cli documentation (#3102)
8566207
unverified
cmake : removed stdc++fs (#3097)
e715962
unverified
server : update httplib.h to version 0.20.0 (#3101)
238f652
unverified
ruby : refine HTTP cache feature (#3109)
f1d4a23
unverified
talk-llama : sync llama.cpp
05fda4a
sync : ggml
6d29e32
CUDA: batched+noncont MMQ, refactor bs>1 MoE code (llama/13199)
a867083
vulkan: use uint array index to avoid glslang bug (llama/13193)
fd2d86d
ggml : fix ppc64le build (llama/13176)
07ec79f
feat(ggml-cpu): enable z17 compile (llama/13182)
10f7d18
Aaron Teo
commited on
CUDA: fix non-cont. inputs for batched mat mul (llama/13155)
d13b876
fix(rpc): Improve input validation and error handling (llama/13069)
9e9f2fe
Ville Vesilehto
commited on
SYCL: Add all missing unary kernels (llama/13074)
d2ce872
Akarshan Biswas
commited on
musa: fix typo in cc control (llama/13144)
5fb7320
R0CKSTAR
commited on
CUDA: fix q_nope_absorbed prec for DS 2 Lite f16 (llama/13137)
e9c9d4b
musa: fix build warning (llama/13129)
3436ba4
R0CKSTAR
commited on
ggml: move fp16/bf16 conversion optimizations to CPU backend + export conversion APIs (llama/13107)
c47823e
change the reorder tensor from init to execute OP (llama/13003)
8614863
Neo Zhang Jianyu
commited on
rpc : do not wait for response when sending RPC_CMD_SET_TENSOR (llama/12943)
691c071
ggml : fix ggml_gallocr_ptr type (ggml/1205)
cf46d5c
Diego Devesa
commited on
whisper : add check that target name exists (#3103)
60ff3ed
unverified
server : add --no-gpu option to print usage output (#3098)
1eb0f64
unverified
ruby : ignore "Downloading" output in test_log_suppress (#3106)
fdb6c7e
unverified
make : fix samples glob pattern (#3100)
0a9e5b1
unverified
ggml : suppress Windows compiler warnings (#3075)
887f7a2
unverified
whisper : fix grammar advance stack warning (#3087)
e4a0565
unverified
examples : expose language detection probabilities to server example (#3044)
6b8d348
unverified
whisper : remove empty .gitmodules file [no ci] (#3085)
aa54166
unverified
talk-llama : sync llama.cpp (#3084)
511930c
unverified
ci : disable publishing of java binding [no ci] (#3086)
4b6e041
unverified
build : Add Moore Threads GPU support and update GitHub workflow for MUSA build (#3069)
8ede9a1
unverified
R0CKSTAR
commited on