From 9f33385ba4bfd21a00aa0a266d8ab6236aa6fad0 Mon Sep 17 00:00:00 2001 From: Maziyar Panahi Date: Wed, 16 Nov 2022 21:37:27 +0100 Subject: [PATCH] Adding more information regarding load_model Adding information regarding parameters accepted by `load_model()`: - `name: str` that accepts `mini`, `base`, `standard`, `large`, and `huge` - `dtype: str=None` that accepts `float16` (or when the name is set to `huge`) and `float32` - `num_gpus: int=None` that accepts the total number of GPUs to be used (by default this is `8`) --- README.md | 5 +++++ 1 file changed, 5 insertions(+) diff --git a/README.md b/README.md index 58032c3..e44a50e 100644 --- a/README.md +++ b/README.md @@ -45,6 +45,11 @@ model.generate("Scaled dot product attention:\n\n\\[") # Scaled dot product attention:\n\n\\[ \\displaystyle\\text{Attention}(Q,K,V)=\\text{softmax}(\\frac{QK^{T}}{\\sqrt{d_{k}}}%\n)V \\] ``` +The `load_model()` function accepts the following parameters: +- `name: str` that accepts `mini`, `base`, `standard`, `large`, and `huge` +- `dtype: str=None` that accepts `float16` (or when the name is set to `huge`) and `float32` +- `num_gpus: int=None` that accepts the total number of GPUs to be used (by default this is `8`) + ## Capabilities We demonstrate some examples using the standard (6.7B) model below.