An LLM Framework Using a Gradient Descent-Like Architecture: Self-refines its prompts with each iteration until it arrives at the correct answer. The LLM computes a "verbal loss" based on the feedback from each response, which it uses to improve the next iteration. The loss is calculated by another LLM, inspired by the concept of using an LLM as a judge.
Implementing Textgrad without using Textgrad.
-
Notifications
You must be signed in to change notification settings - Fork 0
Rev-x/GradientLLM
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
About
No description, website, or topics provided.
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published