Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

b3267 #200

Merged
merged 23 commits into from
Jun 30, 2024
Merged

b3267 #200

Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
23 commits
Select commit Hold shift + click to select a range
9b31a40
clip : suppress unused variable warnings (#8105)
danbev Jun 26, 2024
ac14662
Fix llama-android.cpp for error - "common/common.h not found" (#8145)
criminact Jun 27, 2024
911e35b
llama : fix CodeLlama FIM token checks (#8144)
CISC Jun 27, 2024
f675b20
Added support for Viking pre-tokenizer (#8135)
kustaaya Jun 27, 2024
85a267d
CUDA: fix MMQ stream-k for --split-mode row (#8167)
JohannesGaessler Jun 27, 2024
6030c61
Add Qwen2MoE 57B-A14B model identifier (#8158)
CISC Jun 27, 2024
3879526
Delete examples/llama.android/llama/CMakeLists.txt (#8165)
criminact Jun 27, 2024
97877eb
Control vector loading fixes (#8137)
jukofyork Jun 27, 2024
ab36791
flake.lock: Update (#8071)
ggerganov Jun 27, 2024
16791b8
Add chatml fallback for cpp `llama_chat_apply_template` (#8160)
ngxson Jun 27, 2024
8172ee9
cmake : fix deprecated option names not working (#8171)
slaren Jun 27, 2024
558f44b
CI: fix release build (Ubuntu+Mac) (#8170)
loonerin Jun 27, 2024
cb0b06a
`json`: update grammars/README w/ examples & note about additionalPro…
ochafik Jun 27, 2024
a27aa50
Add missing items in makefile (#8177)
ngxson Jun 28, 2024
e57dc62
llama: Add support for Gemma2ForCausalLM (#8156)
pculliton Jun 28, 2024
139cc62
`json`: restore default additionalProperties to false, fix some patte…
ochafik Jun 28, 2024
b851b3f
cmake : allow user to override default options (#8178)
slaren Jun 28, 2024
38373cf
Add SPM infill support (#8016)
CISC Jun 28, 2024
26a39bb
Add MiniCPM, Deepseek V2 chat template + clean up `llama_chat_apply_t…
ngxson Jun 28, 2024
8748d8a
json: attempt to skip slow tests when running under emulator (#8189)
ochafik Jun 28, 2024
72272b8
fix code typo in llama-cli (#8198)
ngxson Jun 28, 2024
1c5eba6
llama: Add attention and final logit soft-capping, update scaling fac…
abetlen Jun 30, 2024
9ef0780
Fix new line issue with chat template, disable template when in-prefi…
ngxson Jun 30, 2024
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 4 additions & 3 deletions .github/workflows/build.yml
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,7 @@ jobs:
sysctl -a
mkdir build
cd build
cmake -DLLAMA_FATAL_WARNINGS=ON -DGGML_METAL_EMBED_LIBRARY=ON -DLLAMA_CURL=ON ..
cmake -DLLAMA_FATAL_WARNINGS=ON -DGGML_METAL_EMBED_LIBRARY=ON -DLLAMA_CURL=ON -DBUILD_SHARED_LIBS=OFF ..
cmake --build . --config Release -j $(sysctl -n hw.logicalcpu)

- name: Test
Expand Down Expand Up @@ -105,7 +105,7 @@ jobs:
sysctl -a
# Metal is disabled due to intermittent failures with Github runners not having a GPU:
# https://github.com/ggerganov/llama.cpp/actions/runs/8635935781/job/23674807267#step:5:2313
cmake -B build -DLLAMA_FATAL_WARNINGS=ON -DGGML_METAL=OFF -DLLAMA_CURL=ON
cmake -B build -DLLAMA_FATAL_WARNINGS=ON -DGGML_METAL=OFF -DLLAMA_CURL=ON -DBUILD_SHARED_LIBS=OFF
cmake --build build --config Release -j $(sysctl -n hw.logicalcpu)

- name: Test
Expand Down Expand Up @@ -222,7 +222,7 @@ jobs:
run: |
mkdir build
cd build
cmake .. -DLLAMA_FATAL_WARNINGS=ON -DLLAMA_CURL=ON
cmake .. -DLLAMA_FATAL_WARNINGS=ON -DLLAMA_CURL=ON -DBUILD_SHARED_LIBS=OFF
cmake --build . --config Release -j $(nproc)

- name: Test
Expand Down Expand Up @@ -799,6 +799,7 @@ jobs:
7z x "-o${env:RUNNER_TEMP}" $env:RUNNER_TEMP/sde.tar
$sde = $(join-path $env:RUNNER_TEMP sde-external-${env:SDE_VERSION}-win/sde.exe)
cd build
$env:LLAMA_SKIP_TESTS_SLOW_ON_EMULATOR = 1
& $sde -future -- ctest -L main -C Release --verbose --timeout 900

- name: Determine tag name
Expand Down
14 changes: 10 additions & 4 deletions CMakeLists.txt
Original file line number Diff line number Diff line change
Expand Up @@ -79,14 +79,21 @@ set(GGML_SANITIZE_ADDRESS ${LLAMA_SANITIZE_ADDRESS})
set(GGML_SANITIZE_UNDEFINED ${LLAMA_SANITIZE_UNDEFINED})
set(GGML_ALL_WARNINGS ${LLAMA_ALL_WARNINGS})
set(GGML_FATAL_WARNINGS ${LLAMA_FATAL_WARNINGS})
set(GGML_LLAMAFILE ON)
set(GGML_CUDA_USE_GRAPHS ON)

# change the default for these ggml options
if (NOT DEFINED GGML_LLAMAFILE)
set(GGML_LLAMAFILE ON)
endif()

if (NOT DEFINED GGML_CUDA_USE_GRAPHS)
set(GGML_CUDA_USE_GRAPHS ON)
endif()

# transition helpers
function (llama_option_depr TYPE OLD NEW)
if (${OLD})
message(${TYPE} "${OLD} is deprecated and will be removed in the future.\nUse ${NEW} instead\n")
set(${NEW} ON)
set(${NEW} ON PARENT_SCOPE)
endif()
endfunction()

Expand All @@ -96,7 +103,6 @@ llama_option_depr(WARNING LLAMA_KOMPUTE GGML_KOMPUTE)
llama_option_depr(WARNING LLAMA_METAL GGML_METAL)
llama_option_depr(WARNING LLAMA_METAL_EMBED_LIBRARY GGML_METAL_EMBED_LIBRARY)
llama_option_depr(WARNING LLAMA_NATIVE GGML_NATIVE)
llama_option_depr(WARNING LLAMA_OPENMP GGML_OPENMP)
llama_option_depr(WARNING LLAMA_RPC GGML_RPC)
llama_option_depr(WARNING LLAMA_SYCL GGML_SYCL)
llama_option_depr(WARNING LLAMA_SYCL_F16 GGML_SYCL_F16)
Expand Down
2 changes: 2 additions & 0 deletions Makefile
Original file line number Diff line number Diff line change
Expand Up @@ -45,6 +45,7 @@ BUILD_TARGETS = \
TEST_TARGETS = \
tests/test-autorelease \
tests/test-backend-ops \
tests/test-chat-template \
tests/test-double-float \
tests/test-grad0 \
tests/test-grammar-integration \
Expand Down Expand Up @@ -1070,6 +1071,7 @@ clean:
rm -rvf src/*.o
rm -rvf tests/*.o
rm -rvf examples/*.o
rm -rvf common/*.o
rm -rvf *.a
rm -rvf *.dll
rm -rvf *.so
Expand Down
Loading
Loading