You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
| Jan 27th |**Inference-Time Techniques for LLM Reasoning** <br> Xinyun Chen, Google DeepMind <br> <ahref="https://rdi.berkeley.edu/adv-llm-agents/slides/llm-agents-berkeley-intro-sp25.pdf">Intro</a> <ahref="https://rdi.berkeley.edu/adv-llm-agents/slides/inference_time_techniques_lecture_sp25.pdf">Slides</a> | - [Large Language Models as Optimizers](https://arxiv.org/abs/2309.03409) <br> - [Large Language Models Cannot Self-Correct Reasoning Yet](https://arxiv.org/abs/2310.01798) <br> - [Teaching Large Language Models to Self-Debug](https://arxiv.org/abs/2304.05128) <br> *All readings are optional this week.*|
65
-
| Feb 3rd‡ |**Learning to reason with LLMs** <br> Jason Weston, Meta | - [Direct Preference Optimization: Your Language Model is Secretly a Reward Model](https://arxiv.org/abs/2305.18290) <br> - [Iterative Reasoning Preference Optimization](https://arxiv.org/abs/2404.19733) <br> - [Chain-of-Verification Reduces Hallucination in Large Language Models](https://arxiv.org/abs/2309.11495)|
65
+
| Feb 3rd‡ |**Learning to reason with LLMs** <br> Jason Weston, Meta <br> <ahref="https://rdi.berkeley.edu/adv-llm-agents/slides/Jason-Weston-Reasoning-Alignment-Berkeley-Talk.pdf">Slides</a> | - [Direct Preference Optimization: Your Language Model is Secretly a Reward Model](https://arxiv.org/abs/2305.18290) <br> - [Iterative Reasoning Preference Optimization](https://arxiv.org/abs/2404.19733) <br> - [Chain-of-Verification Reduces Hallucination in Large Language Models](https://arxiv.org/abs/2309.11495)|
66
66
| Feb 10th‡ |**Reasoning & Planning** <br> Yu Su, Ohio State University ||
67
67
| Feb 17th |*No Class - Presidents' Day*||
68
68
| Feb 24th‡ |**Reasoning and Planning in Large Language Models** <br> Hanna Hajishirzi, University of Washington ||
0 commit comments