Really quick-and-dirty example of AI recursive learning (video in action:https://x.com/jconorgrogan/status/1853943122046144829) ; The TLDR is that we can make AI "Learn" via smart, recursive workflows that rejigger the context window. Goes as follows:
- Digest user query. Send towards two paths: 2a. Answer query 2b. Have one LLM create test cases for that query
- We then combine 2a and 2b, passing the llm best answer through those test cases
- Have a judge LLM review those responses for logic errors. Ideally the testing step is broken down to the most basic components and the judge just reviews pass/fail based on simple criteria
- Another LLM with a new context window collects those learnings (if there are any fails)
- This LLM then creates a new prompt which amends the original user prompt with the learnings of the LLMs iterations, and sends that through to the model as part of a new context window
- This process loops back to 2a until all logic tests are passed, breaking the loop
- The answer is cleaned up by another LLM returned to the user
This image is the result of two recursive iterations. The initial starting prompt was, "Construct a sentence where none of the words in that sentence are in the Bible", which o1preview constantly struggles with. As you can see, by the end the "user query" had changed a lot, with learnings
Insert your OpenAI key to the "Proofofconcept.html" run and try it.
Future exploration ideas
Example #1: [Not yet in production]
- Input prompt, have llm break down goal to the most basic components
- test those components, individually, each with a context wiped new pull
- have an llm summarize the results; if all pass then pass full response to user. if fail, then synthesize learnings
- another llm takes these learnings and then constructs a new prompt with new context, amending the original user prompt
- this happens recursively again and again until the test are all passed
- Have an LLM then synthesize the best possible answer given original goal
- Return to user
Example #2; introspective/entropix-esque [Not yet in production]
- User query
- LLM responds
- Overseer LLM #1 Check branches of that response to assess alternate answers possibilities
- Overseer LLM #2 triggers completion for those "alternative universe" tokens
- context compressed and sent to an LLM evaluates all COT output from unique branches
- LLM responds given best answer
- you can even loop this if you want to, eg if there are branches in best answer from the entropy/varentropy profile
- Return that answer to user, via an LLM that is focused on response formatting/user objective
Example #3 Introspective pruning- [See 9.11or9.9Solver html file for proof-of-concept which has an LLM replace any numbers like 9.11 (which fire calandar-related activations) with short math equations (9+0.11)]
- User querys
- LLM responds
- LLM reviews neuron activations, if any fire in domains that are counter to the user goal, seek to clamp those down
- Alter context/prompt with tokens that still compress the user goal, but in a way that uses tokens that minimize neuron activation for activity that could negatively impact the output