Skip to content

Releases: CodeLinaro/llama.cpp

b3145

13 Jun 22:06
172c825
Compare
Choose a tag to compare
rpc : fix ggml_backend_rpc_supports_buft() (#7918)

b3087

04 Jun 20:36
1442677
Compare
Choose a tag to compare
common : refactor cli arg parsing (#7675)

* common : gpt_params_parse do not print usage

* common : rework usage print (wip)

* common : valign

* common : rework print_usage

* infill : remove cfg support

* common : reorder args

* server : deduplicate parameters

ggml-ci

* common : add missing header

ggml-ci

* common : remote --random-prompt usages

ggml-ci

* examples : migrate to gpt_params

ggml-ci

* batched-bench : migrate to gpt_params

* retrieval : migrate to gpt_params

* common : change defaults for escape and n_ctx

* common : remove chatml and instruct params

ggml-ci

* common : passkey use gpt_params

b3060

31 May 17:25
0515ad9
Compare
Choose a tag to compare
convert-hf : Handle NotImplementedError in convert-hf-to-gguf (#7660)

b3029

28 May 23:49
b864b50
Compare
Choose a tag to compare
[SYCL] Align GEMM dispatch (#7566)

* align GEMM dispatch

b3014

28 May 00:47
852aafb
Compare
Choose a tag to compare
update HIP_UMA #7399 (#7414)

* update HIP_UMA #7399

add use of hipMemAdviseSetCoarseGrain when LLAMA_HIP_UMA is enable.
- get x2 on prompte eval and x1.5 on token gen with rocm6.0 on ryzen 7940HX iGPU (780M/gfx1103)

* simplify code, more consistent style

---------

Co-authored-by: slaren <[email protected]>

b2993

25 May 05:35
902184d
Compare
Choose a tag to compare
fix missing slash in `fs_get_cache_directory()` (#7503)

* fix missing slash in fs_get_cache_directory()

* use LOCALAPPDATA for fs_get_cache_directory()

* better code style

b2990

24 May 22:05
d041d2c
Compare
Choose a tag to compare
flake.lock: Update (#7232)

Flake lock file updates:

• Updated input 'flake-parts':
    'github:hercules-ci/flake-parts/e5d10a24b66c3ea8f150e47dfdb0416ab7c3390e?narHash=sha256-yzcRNDoyVP7%2BSCNX0wmuDju1NUCt8Dz9%2BlyUXEI0dbI%3D' (2024-05-02)
  → 'github:hercules-ci/flake-parts/8dc45382d5206bd292f9c2768b8058a8fd8311d9?narHash=sha256-/GJvTdTpuDjNn84j82cU6bXztE0MSkdnTWClUCRub78%3D' (2024-05-16)
• Updated input 'nixpkgs':
    'github:NixOS/nixpkgs/63c3a29ca82437c87573e4c6919b09a24ea61b0f?narHash=sha256-4cPymbty65RvF1DWQfc%2BBc8B233A1BWxJnNULJKQ1EY%3D' (2024-05-02)
  → 'github:NixOS/nixpkgs/4a6b83b05df1a8bd7d99095ec4b4d271f2956b64?narHash=sha256-%2BNpbZRCRisUHKQJZF3CT%2Bxn14ZZQO%2BKjxIIanH3Pvn4%3D' (2024-05-17)

Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>

b2986

24 May 00:41
74f33ad
Compare
Choose a tag to compare
readme : remove trailing space (#7469)

b2957

21 May 18:03
c3f8d58
Compare
Choose a tag to compare
tests : test-tokenizer-0.sh print more info (#7402)

b2953

20 May 21:37
917dc8c
Compare
Choose a tag to compare
Tokenizer SPM fixes for phi-3 and llama-spm (#7375)

* Update brute force test: special tokens
* Fix added tokens
  - Try to read 'added_tokens.json'.
  - Try to read 'tokenizer_config.json'.
  - Try to read 'tokenizer.json'.
* Fix special tokens rtrim

Co-authored-by: Georgi Gerganov <[email protected]>
* server : fix test regexes