Releases: CodeLinaro/llama.cpp
Releases · CodeLinaro/llama.cpp
b2986
readme : remove trailing space (#7469)
b2957
tests : test-tokenizer-0.sh print more info (#7402)
b2953
Tokenizer SPM fixes for phi-3 and llama-spm (#7375) * Update brute force test: special tokens * Fix added tokens - Try to read 'added_tokens.json'. - Try to read 'tokenizer_config.json'. - Try to read 'tokenizer.json'. * Fix special tokens rtrim Co-authored-by: Georgi Gerganov <[email protected]> * server : fix test regexes
fix-win-artifacts-b2897-e910cea
ci: fix bin/Release path for windows-arm64 builds Switch to Ninja Multi-Config CMake generator to resurect bin/Release path that broke artifact packaging in CI.
b2894
ggml : use dynamic thread scheduling for matrix multiplication (#6915) * Just reordering some structs. * Adding in the calls to mm_pause * Passing around the state * Renaming and moving a bunch of variables around. * Extracting the logic to it's own function. * Moving some variable definitions into the chunk function. * Moving some variables around * moving src1_cont inside * Moving row_size * adding the current_chunk * Reorg the code. * Formatting to match the orig patch * starting to setup the chunking variables * Starting the buildup of the loop * The yield shouldn't be necessary. * adding the looping structure based on the chunk configuration. * Add in the re-chunking code. * Making it much more likely to rechunk. * disable resizing if numa is enabled. * Updating comments with what we've learned. * Fix formatting * Couple more formatting fixes. * More style fixes. * Fix Warnings * Going with unused because there's conditional logic that needs it. * Update ggml.c * Update ggml.c ---------
b2886
script : sync ggml-rpc
b2866
convert.py: Outfile default name change and additional metadata suppo…