Skip to content

Commit

Permalink
docs: removed llm_gym mentionings
Browse files Browse the repository at this point in the history
  • Loading branch information
lllAlexanderlll committed Jan 15, 2024
1 parent b7169c4 commit bbb8830
Show file tree
Hide file tree
Showing 8 changed files with 22 additions and 17 deletions.
11 changes: 7 additions & 4 deletions .github/workflows/build_documentation.yml
Original file line number Diff line number Diff line change
Expand Up @@ -10,9 +10,12 @@ jobs:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v4
- name: Check path
- name: Install dependencies
run: |
pip install myst-parser
- uses: ammaraskar/sphinx-action@master
with:
docs-folder: "docs/"
pip install -e .
pip install sphinx-rtd-theme
- name: "Parse into HTML"
run: |
sphinx-build -M html docs/source/ docs/build/
1 change: 1 addition & 0 deletions .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -149,3 +149,4 @@ tags

checkpoints
data
docs/source/generated
2 changes: 1 addition & 1 deletion docs/source/api.rst
Original file line number Diff line number Diff line change
Expand Up @@ -4,4 +4,4 @@ API
.. autosummary::
:toctree: generated

lumache
modalities
5 changes: 3 additions & 2 deletions docs/source/conf.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,9 +2,9 @@

# -- Project information

project = "LLMGym"
project = "Modalities"
copyright = "2023, Fraunhofer"
author = "Max Lübbering"
author = "Many"

release = "0.1"
version = "0.1.0"
Expand All @@ -17,6 +17,7 @@
"sphinx.ext.autodoc",
"sphinx.ext.autosummary",
"sphinx.ext.intersphinx",
"myst_parser",
]

intersphinx_mapping = {
Expand Down
2 changes: 1 addition & 1 deletion docs/source/index.rst
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
Welcome to LLMgym's documentation!
Welcome to Modalities's documentation!
===================================

**EDIT "docs/source/index.rst" IN ORDER TO MAKE CHANGES HERE**
Expand Down
4 changes: 2 additions & 2 deletions docs/source/memmap.rst
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ The :python:`MemMapDataset` requires an index file providing the necessary point

.. code-block:: bash
llm_gym create_memmap_index <path/to/jsonl/file>
modalities create_memmap_index <path/to/jsonl/file>
The index will be created in the same directory as the raw data file. For further options you may look into the usage documentation via :bash:`llm_gym create_memmap_index --help`.

Expand All @@ -27,7 +27,7 @@ The :python:`PackedMemMapDatasetContinuous` and :python:`PackedMemMapDatasetMega

.. code-block:: bash
llm_gym create_packed_data <path/to/jsonl/file>
modalities create_packed_data <path/to/jsonl/file>
The packed data file will be created in the same directory as the raw data file. For further options you may look into the usage documentation via :bash:`llm_gym create_packed_data --help`.

Expand Down
12 changes: 6 additions & 6 deletions docs/source/quickstart.rst
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ Quickstart

Installation
-----------
Setup a conda environment `conda create -n llm_gym python=3.10 & conda activate llm_gym` and install the requirements `pip install -e .`.
Setup a conda environment `conda create -n modalities python=3.10 & conda activate modalities` and install the requirements `pip install -e .`.

Setup Dataset
------------
Expand All @@ -14,28 +14,28 @@ To start a training you need to create memmap dataset out of a jsonl file first,
.. code-block:: bash
# Create memmap dataset from jsonl file.
llm_gym create_memmap_index <path/to/jsonl/file>
modalities create_memmap_index <path/to/jsonl/file>
# Create packed dataset.
llm_gym create_packed_data <path/to/jsonl/file>
modalities create_packed_data <path/to/jsonl/file>
For example, using the lorem ipsum example:

.. code-block:: bash
# Create memmap dataset from jsonl file.
llm_gym create_memmap_index data/lorem_ipsum.jsonl
modalities create_memmap_index data/lorem_ipsum.jsonl
# Create packed dataset.
llm_gym create_packed_data data/lorem_ipsum.jsonl
modalities create_packed_data data/lorem_ipsum.jsonl
Training
--------
To run a training environment variables in a multi-gpu setting are required.

.. code-block:: bash
CUDA_VISIBLE_DEVICES=0,1 torchrun --nnodes 1 --nproc_per_node 2 --rdzv-endpoint=0.0.0.0:29502 src/llm_gym/__main__.py run --config_file_path config_files/config_lorem_ipsum.yaml
CUDA_VISIBLE_DEVICES=0,1 torchrun --nnodes 1 --nproc_per_node 2 --rdzv-endpoint=0.0.0.0:29502 src/modalities/__main__.py run --config_file_path config_files/config_lorem_ipsum.yaml
4. **Evaluation:**
WIP add contents
2 changes: 1 addition & 1 deletion docs/source/vs_code_setup.rst
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ In VSCode, add this to your :file:`launch.json`:
"--nproc_per_node",
"2",
"--rdzv-endpoint=0.0.0.0:29503",
"src/llm_gym/__main__.py",
"src/modalities/__main__.py",
"run",
"--config_file_path",
"config_files/config.yaml",
Expand Down

0 comments on commit bbb8830

Please sign in to comment.