From 1cc58cd4b18b80e04ecb078403bb6854eda8d01f Mon Sep 17 00:00:00 2001 From: Olivier DEBAUCHE Date: Mon, 6 May 2024 15:37:25 +0200 Subject: [PATCH] Bump llama-cpp-python to v0.2.69 (#1266) * Update serge.env * Update dev.sh * Update deploy.sh * Update dev.sh * Update dev.sh * Update deploy.sh * Bump LLaMA CPP Python to 0.2.68 * Update dev.sh * Update deploy.sh * Update deploy.sh * Update dev.sh * Update dev.sh * Update deploy.sh * Update deploy.sh * Update dev.sh * Update deploy.sh * Update dev.sh * Update serge.env * Update serge.env * Update scripts/deploy.sh Co-authored-by: coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com> * Update scripts/dev.sh Co-authored-by: coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com> * Update dev.sh * Update deploy.sh * Only use official wheels for now * Fix lint issues --------- Co-authored-by: Juan Calderon-Perez <835733+gaby@users.noreply.github.com> Co-authored-by: coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com> --- scripts/deploy.sh | 9 +++++---- scripts/dev.sh | 9 +++++---- scripts/serge.env | 2 +- sweep.yaml | 27 --------------------------- 4 files changed, 11 insertions(+), 36 deletions(-) delete mode 100644 sweep.yaml diff --git a/scripts/deploy.sh b/scripts/deploy.sh index 109a90f..7dc8293 100755 --- a/scripts/deploy.sh +++ b/scripts/deploy.sh @@ -21,12 +21,13 @@ detect_cpu_features() { } # Check if the CPU architecture is aarch64/arm64 -if [ "$cpu_arch" = "aarch64" ]; then - pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://gaby.github.io/arm64-wheels/" +if [ "$cpu_arch" = "aarch64" ] || [ "$cpu_arch" = "arm64" ]; then + pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://abetlen.github.io/llama-cpp-python/whl/cpu/" else # Use @smartappli provided wheels - cpu_feature=$(detect_cpu_features) - pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://smartappli.github.io/serge-wheels/$cpu_feature/cpu" + #cpu_feature=$(detect_cpu_features) + #pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://abetlen.github.io/llama-cpp-python/whl/cpu-$cpu_feature/" + pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://abetlen.github.io/llama-cpp-python/whl/cpu/" fi echo "Recommended install command for llama-cpp-python: $pip_command" diff --git a/scripts/dev.sh b/scripts/dev.sh index 3ed8459..4e3d668 100755 --- a/scripts/dev.sh +++ b/scripts/dev.sh @@ -21,12 +21,13 @@ detect_cpu_features() { } # Check if the CPU architecture is aarch64/arm64 -if [ "$cpu_arch" = "aarch64" ]; then - pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://gaby.github.io/arm64-wheels/" +if [ "$cpu_arch" = "aarch64" ] || [ "$cpu_arch" = "arm64" ]; then + pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://abetlen.github.io/llama-cpp-python/whl/cpu/" else # Use @smartappli provided wheels - cpu_feature=$(detect_cpu_features) - pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://smartappli.github.io/serge-wheels/$cpu_feature/cpu" + #cpu_feature=$(detect_cpu_features) + #pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://abetlen.github.io/llama-cpp-python/whl/cpu-$cpu_feature/" + pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://abetlen.github.io/llama-cpp-python/whl/cpu/" fi echo "Recommended install command for llama-cpp-python: $pip_command" diff --git a/scripts/serge.env b/scripts/serge.env index 74230a2..b7dac63 100644 --- a/scripts/serge.env +++ b/scripts/serge.env @@ -1,3 +1,3 @@ -LLAMA_PYTHON_VERSION=0.2.64 +LLAMA_PYTHON_VERSION=0.2.69 SERGE_ENABLE_IPV4=true SERGE_ENABLE_IPV6=false diff --git a/sweep.yaml b/sweep.yaml deleted file mode 100644 index 89e1d02..0000000 --- a/sweep.yaml +++ /dev/null @@ -1,27 +0,0 @@ -# Sweep AI turns bugs & feature requests into code changes (https://sweep.dev) -# For details on our config file, check out our docs at https://docs.sweep.dev/usage/config - -# This setting contains a list of rules that Sweep will check for. If any of these rules are broken in a new commit, Sweep will create an pull request to fix the broken rule. -rules: - - "All new business logic should have corresponding unit tests." - - "Refactor large functions to be more modular." - - "Add docstrings to all functions and file headers." - -# This is the branch that Sweep will develop from and make pull requests to. Most people use 'main' or 'master' but some users also use 'dev' or 'staging'. -branch: 'main' - -# By default Sweep will read the logs and outputs from your existing Github Actions. To disable this, set this to false. -gha_enabled: True - -# This is the description of your project. It will be used by sweep when creating PRs. You can tell Sweep what's unique about your project, what frameworks you use, or anything else you want. -# -# Example: -# -# description: sweepai/sweep is a python project. The main api endpoints are in sweepai/api.py. Write code that adheres to PEP8. -description: '' - -# This sets whether to create pull requests as drafts. If this is set to True, then all pull requests will be created as drafts and GitHub Actions will not be triggered. -draft: False - -# This is a list of directories that Sweep will not be able to edit. -blocked_dirs: []