This Project welcomes contributions, suggestions, and feedback. All contributions, suggestions, and feedback you submitted are accepted under the Project's license. You represent that if you do not own copyright in the code that you have the authority to submit it under the Project's license. All feedback, suggestions, or contributions are not confidential.
The Project abides by the Organization's code of conduct and trademark policy.
We welcome contributions to guidance
, and this document exists to provide useful information contributors.
The quickest way to get started is to run (in a fresh environment):
pip install -e .[all,test,bench]
which should bring in all of the basic required dependencies.
Note that if you want to use GPU acceleration, then you will need to do whatever is required to allow torch
and llama-cpp
to access your GPU too.
There are sometimes difficulties configuring Rust during the pip installs. If you encounter such issues, then one work around (if you use Anaconda) is to create your environment along the lines of
conda create -n guidance-312 python=3.12 rust
In our experience, this has been a little more reliable. Similarly, to get GPU support, we have found that (after activating the environment) running
conda install pytorch pytorch-cuda=12.1 -c pytorch -c nvidia
works best. However, if you have your own means of installing Rust and CUDA, you should be able to continue using those.
To run a basic test suite locally:
python -m pytest ./tests/
which runs our basic test suite. Where an LLM is required, this will default to using GPT2 on the CPU.
To change that default, run
python -m pytest --selected_model <MODELNAME> ./tests/
where <MODELNAME>
is taken from the AVAILABLE_MODELS
dictionary defined in _llms_for_testing.py
.
Alternatively, the default value for --selected_model
can be set via the GUIDANCE_SELECTED_MODEL
environment variable.
This may be useful when trying to use a debugger when running pytest
, and setting the extra command line argument in the debugger configuration is tricky.
Just remember that the environment variable needs to be set before starting PyCharm/VSCode etc.
Our tests run on a variety of LLMs. These fall into three categories: CPU-based, GPU-based and endpoint-based (which need credentials).
Due to the limited resources of the regular GitHub runner machines, the LLM under test is a dimension of our test matrix (otherwise the GitHub runners will tend to run out of RAM and/or hard drive space).
New models should be configured in the AVAILABLE_MODELS
dictionary in conftest.py
, and then that key added to the model
list in unit_tests.yml
or unit_tests_gpu.yml
as appropriate.
The model will then be available via the selected_model
fixture for all tests.
If you have a test which should only run for particular models, you can use the selected_model_name
fixture to check, and call pytest.skip()
if necessary.
An example of this is given in test_llama_cpp.py
.
If your model requires credentials, then those will need to be added to our GitHub repository as secrets.
The endpoint itself (and any other required information) should be configured as environment variables too.
When the test runs, the environment variables will be set, and can then be used to configure the model as required.
See test_azureai_openai.py
for examples of this being done.
The environment variables and secrets will also need to be configured in the ci_tests.yml
file.
We run black
on our codebase, and plan to turn on enforcement of this in the GitHub workflows soon.
Part of MVG-0.1-beta. Made with love by GitHub. Licensed under the CC-BY 4.0 License.