# textgrad **Repository Path**: xiaoheaicoding/textgrad ## Basic Information - **Project Name**: textgrad - **Description**: TextGrad: Automatic ''Differentiation'' via Text -- using large language models to backpropagate textual gradients. - **Primary Language**: Unknown - **License**: MIT - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 1 - **Created**: 2026-04-24 - **Last Updated**: 2026-04-24 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README  [](https://colab.research.google.com/github/zou-group/TextGrad/blob/main/examples/notebooks/Prompt-Optimization.ipynb) [][#license-gh-package] [](https://www.nature.com/articles/s41586-025-08661-4) [][#arxiv-paper-package] [][#docs-package] [][#pypi-package] [][#pypi-package] [][#conda-forge-package] [][#conda-forge-package] [#license-gh-package]: https://lbesson.mit-license.org/ [#arxiv-paper-package]: https://arxiv.org/abs/2406.07496 [#docs-package]: https://textgrad.readthedocs.io/en/latest/?badge=latest [#pypi-package]: https://pypi.org/project/textgrad/ [#conda-forge-package]: https://anaconda.org/conda-forge/textgrad [#nature-link]: https://www.nature.com/articles/s41586-025-08661-4 ## TextGrad: Automatic ''Differentiation'' via Text An autograd engine -- for textual gradients! TextGrad is a powerful framework building automatic ``differentiation'' via text. TextGrad implements backpropagation through text feedback provided by LLMs, strongly building on the gradient metaphor We provide a simple and intuitive API that allows you to define your own loss functions and optimize them using text feedback. This API is similar to the Pytorch API, making it simple to adapt to your usecases.  ### Updates: **19th March 2025** TextGrad published in [Nature](https://www.nature.com/articles/s41586-025-08661-4)! **Past Updates**: We are introducing a new engine based on [litellm](https://github.com/BerriAI/litellm). This should allow you to use any model you like, as long as it is supported by litellm. This means that now **Bedrock, Together, Gemini and even more** are all supported by TextGrad! This should be seen as experimental but we plan to depreciate the old engines in the future. In addition to this, with the new engines it should be easy to enable and disable caching. We are in the process of testing these new engines and deprecating the old engines. If you have any issues, please let us know! The new litellm engines can be loaded with the following code: An example of loading a litellm engine: ```python engine = get_engine("experimental:gpt-4o", cache=False) # this also works with set_backward_engine("experimental:gpt-4o", cache=False) ``` Be sure to set the relevant environment variables for the new engines! An example of forward pass: ```python import httpx from textgrad.engine_experimental.litellm import LiteLLMEngine LiteLLMEngine("gpt-4o", cache=True).generate(content="hello, what's 3+4", system_prompt="you are an assistant") image_url = "https://upload.wikimedia.org/wikipedia/commons/a/a7/Camponotus_flavomarginatus_ant.jpg" image_data = httpx.get(image_url).content LiteLLMEngine("gpt-4o", cache=True).generate(content=[image_data, "what is this my boy"], system_prompt="you are an assistant") ``` In the examples folder you will find two new notebooks that show how to use the new engines. ## QuickStart If you know PyTorch, you know 80% of TextGrad. Let's walk through the key components with a simple example. Say we want to use GPT-4o to solve a simple reasoning problem. The question is *If it takes 1 hour to dry 25 shirts under the sun, how long will it take to dry 30 shirts under the sun? Reason step by step.* (Thanks, [Reddit User](https://www.reddit.com/r/OpenAI/comments/18q479x/comment/kf444es/)) ```python import textgrad as tg tg.set_backward_engine("gpt-4o", override=True) # Step 1: Get an initial response from an LLM. model = tg.BlackboxLLM("gpt-4o") question_string = ("If it takes 1 hour to dry 25 shirts under the sun, " "how long will it take to dry 30 shirts under the sun? " "Reason step by step") question = tg.Variable(question_string, role_description="question to the LLM", requires_grad=False) answer = model(question) ``` > :warning: **answer: To determine how long it will take to dry 30 shirts under the sun,** > **we can use a proportional relationship based on the given information.** > **Here’s the step-by-step reasoning: [.....]** > **So, it will take 1.2 hours (or 1 hour and 12 minutes) to dry 30 shirts under the sun.** As you can see, **the model's answer is incorrect.** We can optimize the answer using TextGrad to get the correct answer. ```python answer.set_role_description("concise and accurate answer to the question") # Step 2: Define the loss function and the optimizer, just like in PyTorch! # Here, we don't have SGD, but we have TGD (Textual Gradient Descent) # that works with "textual gradients". optimizer = tg.TGD(parameters=[answer]) evaluation_instruction = (f"Here's a question: {question_string}. " "Evaluate any given answer to this question, " "be smart, logical, and very critical. " "Just provide concise feedback.") # TextLoss is a natural-language specified loss function that describes # how we want to evaluate the reasoning. loss_fn = tg.TextLoss(evaluation_instruction) ``` > :brain: **loss: [...] Your step-by-step reasoning is clear and logical,** > **but it contains a critical flaw in the assumption that drying time is** > **directly proportional to the number of shirts. [...]** ```python # Step 3: Do the loss computation, backward pass, and update the punchline. # Exact same syntax as PyTorch! loss = loss_fn(answer) loss.backward() optimizer.step() answer ``` > :white_check_mark: **answer: It will still take 1 hour to dry 30 shirts under the sun,** > **assuming they are all laid out properly to receive equal sunlight.** We have many more examples around how TextGrad can optimize all kinds of variables -- code, solutions to problems, molecules, prompts, and all that! ### Tutorials We have prepared a couple of tutorials to get you started with TextGrad. The order of this tutorial is what we would recommend to follow for a beginner. You can run them directly in Google Colab by clicking on the links below (but you need an OpenAI/Anthropic key to run the LLMs).
|
Federico Bianchi |
Mert Yuksekgonul |
Nihal Nayak |
Sugato Ray |
Pan Lu |
David Ruan |
|
San |
Zhi Huang |
Albert |
tboen1 |
Atakan Tekparmak |