Skip to content

Commit

Permalink
Update default client-side batches
Browse files Browse the repository at this point in the history
  • Loading branch information
Maxusmusti committed Feb 12, 2024
1 parent 258c9c6 commit c5157a9
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 2 deletions.
2 changes: 1 addition & 1 deletion language/llama2-70b/SUT.py
Original file line number Diff line number Diff line change
Expand Up @@ -112,7 +112,7 @@ def __init__(self,

if not batch_size:
if device == "cpu":
batch_size = 512
batch_size = 2000
else:
batch_size = 32 # Reduce to 8 if using 4 GPUs, 16 for 8.
self.batch_size = batch_size
Expand Down
2 changes: 1 addition & 1 deletion language/llama2-70b/api-endpoint-artifacts/benchmark.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ spec:
restartPolicy: Never
containers:
- name: mlperf-env
image: quay.io/meyceoz/mlperf-inference:v6
image: quay.io/meyceoz/mlperf-inference:v7
resources:
requests:
memory: 20000Mi
Expand Down

0 comments on commit c5157a9

Please sign in to comment.