-
Notifications
You must be signed in to change notification settings - Fork 2.6k
Issues: openai/evals
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[_common.py:105] Backing off create_retrying(...) for 1.3s (False)
bug
Something isn't working
#1571
opened Dec 4, 2024 by
Khyathianeesha
Request for Global Memory Across Different Chats (Across Sessions)
#1570
opened Nov 14, 2024 by
rodrigoreis
Vector maping, origins, 2-3-4D definitions.
bug
Something isn't working
#1569
opened Nov 11, 2024 by
Really-69
Project installation fails: Something isn't working
tensorflow
conflicting dependencies
bug
#1567
opened Nov 7, 2024 by
djbb7
ERROR: Failed building wheel for numpy. clang error compiler does not support 'faltivec
bug
Something isn't working
#1566
opened Nov 6, 2024 by
jinchi2013
AttributeError: module 'openai' has no attribute 'error'
bug
Something isn't working
#1564
opened Oct 17, 2024 by
sahilrajput03
Is Evals repo being replaced by the Evaluations feature in the Playground?
#1562
opened Oct 1, 2024 by
sakher
o1 release breaks token usage stats
bug
Something isn't working
#1556
opened Sep 13, 2024 by
lucapericlp
Information exposure alert through an exception
bug
Something isn't working
#1543
opened Aug 8, 2024 by
arpitjain099
Multiple Unit Test Failures Across OpenAI Assistants, Anthropic, and Google Gemini Libraries
bug
Something isn't working
#1536
opened Jun 24, 2024 by
sakher
Setting completion function args via CLI does not work
bug
Something isn't working
#1504
opened Mar 27, 2024 by
LoryPack
OpenAIChatCompletionFn
should __init__
should accept **kwargs
bug
#1493
opened Mar 15, 2024 by
ezraporter
Proposal for Adding a New Evaluation Metric: Sentiment Analysis Accuracy
#1419
opened Nov 23, 2023 by
Sarfaraz021
Using different models in evaluating mode-graded eval and in generating the completion
#1393
opened Nov 3, 2023 by
LoryPack
In the task "balance_chemical_equation", many instances have incorrect labels.
bug
Something isn't working
#1386
opened Oct 19, 2023 by
dongZheX
Eval-running often hangs on last sample
bug
Something isn't working
#1384
opened Oct 17, 2023 by
sjadler2004
Previous Next
ProTip!
no:milestone will show everything without a milestone.