Skip to content

Conversation

@eamag
Copy link

@eamag eamag commented Apr 15, 2025

Tried running it end-to-end using gemini, ran into so many errors and hidden default models, tried to fix them here. Can split/remove things like example project, ptal!

@eamag
Copy link
Author

eamag commented Apr 15, 2025

2025-04-15_23-01-47_automated_concept_sae_eval_attempt_0_reflection1.pdf
I managed to generate this in the end, figures are still missing

@conglu1997
Copy link
Contributor

Thank you, would it be possible to make this a more minimal addition focused on adding Gemini support but changing nothing else?

@SakanaAI SakanaAI deleted a comment from fangyu05 Apr 18, 2025
@RichardScottOZ
Copy link

Adding gemini support means you have to generalise it more - because in the backend search it checks for either openai or claude models and that is all - so it will fail for llama, deepseek or other examples given in the llm list here.

@eamag
Copy link
Author

eamag commented Apr 19, 2025

@conglu1997 I assume you mean to remove examples? I deleted everything but gemini part and some bugs I found while running e2e

@RichardScottOZ Yes, there are some hardcoded models/clients in the code. I think the best way would be to add something like https://github.com/BerriAI/litellm but that was out of scope, I just wanted to test one run

@RichardScottOZ
Copy link

Yes, or the llama ideation uses OpenRouter already.. so another option for something like litellm

@RichardScottOZ
Copy link

@RichardScottOZ
Copy link

RichardScottOZ commented Apr 20, 2025

Anyway, minor conveniences aside, AI Scientist v2 is very impressive.

Well done.

@eamag
Copy link
Author

eamag commented Apr 23, 2025

ptal @conglu1997

@conglu1997
Copy link
Contributor

See #37

@eamag
Copy link
Author

eamag commented Apr 26, 2025

@conglu1997 that PR only changes models that were defined in llm.py, in your codebase you're also using just random clients defined in several places lol

@conglu1997
Copy link
Contributor

Would it be possible to make a similar style PR? I'm concerned loads of new env variables are being used, new imports, etc. The ideal would be the minimal change that makes this work.

@BradKML
Copy link

BradKML commented Jul 3, 2025

Can this be generalized to cheaper models as well? #29

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants