Bump llama-cpp-python to v0.2.69 (#1266)
* Update serge.env * Update dev.sh * Update deploy.sh * Update dev.sh * Update dev.sh * Update deploy.sh * Bump LLaMA CPP Python to 0.2.68 * Update dev.sh * Update deploy.sh * Update deploy.sh * Update dev.sh * Update dev.sh * Update deploy.sh * Update deploy.sh * Update dev.sh * Update deploy.sh * Update dev.sh * Update serge.env * Update serge.env * Update scripts/deploy.sh Co-authored-by: coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com> * Update scripts/dev.sh Co-authored-by: coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com> * Update dev.sh * Update deploy.sh * Only use official wheels for now * Fix lint issues --------- Co-authored-by: Juan Calderon-Perez <835733+gaby@users.noreply.github.com> Co-authored-by: coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com>
This commit is contained in:
parent
19d55bc61d
commit
1cc58cd4b1
@ -21,12 +21,13 @@ detect_cpu_features() {
|
||||
}
|
||||
|
||||
# Check if the CPU architecture is aarch64/arm64
|
||||
if [ "$cpu_arch" = "aarch64" ]; then
|
||||
pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://gaby.github.io/arm64-wheels/"
|
||||
if [ "$cpu_arch" = "aarch64" ] || [ "$cpu_arch" = "arm64" ]; then
|
||||
pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://abetlen.github.io/llama-cpp-python/whl/cpu/"
|
||||
else
|
||||
# Use @smartappli provided wheels
|
||||
cpu_feature=$(detect_cpu_features)
|
||||
pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://smartappli.github.io/serge-wheels/$cpu_feature/cpu"
|
||||
#cpu_feature=$(detect_cpu_features)
|
||||
#pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://abetlen.github.io/llama-cpp-python/whl/cpu-$cpu_feature/"
|
||||
pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://abetlen.github.io/llama-cpp-python/whl/cpu/"
|
||||
fi
|
||||
|
||||
echo "Recommended install command for llama-cpp-python: $pip_command"
|
||||
|
||||
@ -21,12 +21,13 @@ detect_cpu_features() {
|
||||
}
|
||||
|
||||
# Check if the CPU architecture is aarch64/arm64
|
||||
if [ "$cpu_arch" = "aarch64" ]; then
|
||||
pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://gaby.github.io/arm64-wheels/"
|
||||
if [ "$cpu_arch" = "aarch64" ] || [ "$cpu_arch" = "arm64" ]; then
|
||||
pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://abetlen.github.io/llama-cpp-python/whl/cpu/"
|
||||
else
|
||||
# Use @smartappli provided wheels
|
||||
cpu_feature=$(detect_cpu_features)
|
||||
pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://smartappli.github.io/serge-wheels/$cpu_feature/cpu"
|
||||
#cpu_feature=$(detect_cpu_features)
|
||||
#pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://abetlen.github.io/llama-cpp-python/whl/cpu-$cpu_feature/"
|
||||
pip_command="python -m pip install -v llama-cpp-python==$LLAMA_PYTHON_VERSION --only-binary=:all: --extra-index-url=https://abetlen.github.io/llama-cpp-python/whl/cpu/"
|
||||
fi
|
||||
|
||||
echo "Recommended install command for llama-cpp-python: $pip_command"
|
||||
|
||||
@ -1,3 +1,3 @@
|
||||
LLAMA_PYTHON_VERSION=0.2.64
|
||||
LLAMA_PYTHON_VERSION=0.2.69
|
||||
SERGE_ENABLE_IPV4=true
|
||||
SERGE_ENABLE_IPV6=false
|
||||
|
||||
27
sweep.yaml
27
sweep.yaml
@ -1,27 +0,0 @@
|
||||
# Sweep AI turns bugs & feature requests into code changes (https://sweep.dev)
|
||||
# For details on our config file, check out our docs at https://docs.sweep.dev/usage/config
|
||||
|
||||
# This setting contains a list of rules that Sweep will check for. If any of these rules are broken in a new commit, Sweep will create an pull request to fix the broken rule.
|
||||
rules:
|
||||
- "All new business logic should have corresponding unit tests."
|
||||
- "Refactor large functions to be more modular."
|
||||
- "Add docstrings to all functions and file headers."
|
||||
|
||||
# This is the branch that Sweep will develop from and make pull requests to. Most people use 'main' or 'master' but some users also use 'dev' or 'staging'.
|
||||
branch: 'main'
|
||||
|
||||
# By default Sweep will read the logs and outputs from your existing Github Actions. To disable this, set this to false.
|
||||
gha_enabled: True
|
||||
|
||||
# This is the description of your project. It will be used by sweep when creating PRs. You can tell Sweep what's unique about your project, what frameworks you use, or anything else you want.
|
||||
#
|
||||
# Example:
|
||||
#
|
||||
# description: sweepai/sweep is a python project. The main api endpoints are in sweepai/api.py. Write code that adheres to PEP8.
|
||||
description: ''
|
||||
|
||||
# This sets whether to create pull requests as drafts. If this is set to True, then all pull requests will be created as drafts and GitHub Actions will not be triggered.
|
||||
draft: False
|
||||
|
||||
# This is a list of directories that Sweep will not be able to edit.
|
||||
blocked_dirs: []
|
||||
Loading…
x
Reference in New Issue
Block a user