diff --git a/.github/workflows/ci_tests.yml b/.github/workflows/ci_tests.yml index 838cbde99..de28e840c 100644 --- a/.github/workflows/ci_tests.yml +++ b/.github/workflows/ci_tests.yml @@ -23,6 +23,11 @@ jobs: steps: - uses: actions/checkout@v4 + - name: Install Rust + shell: bash + run: | + curl https://sh.rustup.rs -sSf | sh -s -- -y --default-toolchain 1.75.0 + echo "$HOME/.cargo/bin" >> $GITHUB_PATH - name: Set up Python ${{ matrix.python-version }} uses: actions/setup-python@v5 with: diff --git a/notebooks/art_of_prompt_design/react.ipynb b/notebooks/art_of_prompt_design/react.ipynb index ac7406988..58271f0b9 100644 --- a/notebooks/art_of_prompt_design/react.ipynb +++ b/notebooks/art_of_prompt_design/react.ipynb @@ -48,7 +48,7 @@ "\n", "repo_id = \"TheBloke/Mistral-7B-Instruct-v0.2-GGUF\"\n", "filename = \"mistral-7b-instruct-v0.2.Q8_0.gguf\"\n", - "model_kwargs = {\"verbose\": True, \"n_gpu_layers\": -1}\n", + "model_kwargs = {\"verbose\": True, \"n_gpu_layers\": -1, \"n_ctx\": 4096}\n", "\n", "downloaded_file = hf_hub_download(repo_id=repo_id, filename=filename)\n", "mistral7 = guidance.models.LlamaCpp(downloaded_file, **model_kwargs)" diff --git a/notebooks/art_of_prompt_design/use_clear_syntax.ipynb b/notebooks/art_of_prompt_design/use_clear_syntax.ipynb index 4be18c5b2..00362bdd5 100644 --- a/notebooks/art_of_prompt_design/use_clear_syntax.ipynb +++ b/notebooks/art_of_prompt_design/use_clear_syntax.ipynb @@ -153,7 +153,7 @@ "\n", "repo_id = \"TheBloke/Mistral-7B-Instruct-v0.2-GGUF\"\n", "filename = \"mistral-7b-instruct-v0.2.Q8_0.gguf\"\n", - "model_kwargs = {\"verbose\": True, \"n_gpu_layers\": -1}\n", + "model_kwargs = {\"verbose\": True, \"n_gpu_layers\": -1, \"n_ctx\": 1024}\n", "\n", "downloaded_file = hf_hub_download(repo_id=repo_id, filename=filename)\n", "lm = guidance.models.LlamaCpp(downloaded_file, **model_kwargs)" diff --git a/tests/conftest.py b/tests/conftest.py index 9e7bc593a..30a5c60ef 100644 --- a/tests/conftest.py +++ b/tests/conftest.py @@ -14,7 +14,7 @@ ), "hfllama7b": dict( name="huggingface_hubllama:TheBloke/Llama-2-7B-GGUF:llama-2-7b.Q5_K_M.gguf", - kwargs={"verbose": True}, + kwargs={"verbose": True, "n_ctx": 4096}, ), "transformers_mistral_7b": dict( name="transformers:mistralai/Mistral-7B-v0.1", kwargs=dict() @@ -30,7 +30,7 @@ ), "hfllama_7b_gpu": dict( name="huggingface_hubllama:TheBloke/Llama-2-7B-GGUF:llama-2-7b.Q5_K_M.gguf", - kwargs={"verbose": True, "n_gpu_layers": -1}, + kwargs={"verbose": True, "n_gpu_layers": -1, "n_ctx": 4096}, ), }