<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en-US"><generator uri="https://jekyllrb.com/" version="4.4.1">Jekyll</generator><link href="https://tufalabs.ai/research/feed.xml" rel="self" type="application/atom+xml" /><link href="https://tufalabs.ai/" rel="alternate" type="text/html" hreflang="en-US" /><updated>2026-04-16T10:17:33+02:00</updated><id>https://tufalabs.ai/research/feed.xml</id><title type="html">Tufa Labs</title><subtitle>Public website for Tufa Labs, a small independent research group working on fundamental AI research.</subtitle><entry><title type="html">1st Place in the ARC-AGI-3 Preview Competition</title><link href="https://tufalabs.ai/research/arc3-winning-solution/" rel="alternate" type="text/html" title="1st Place in the ARC-AGI-3 Preview Competition" /><published>2025-08-21T00:50:53+02:00</published><updated>2025-08-21T00:50:53+02:00</updated><id>https://tufalabs.ai/research/ARC3-winning-solution</id><content type="html" xml:base="https://tufalabs.ai/research/arc3-winning-solution/"><![CDATA[<h2 id="introduction">Introduction</h2>

<p>We achieved first place in the ARC-AGI-3 Agent Preview Competition.</p>

<p>Our approach used simple reinforcement learning to tackle the challenge of interactive, turn-based 2D grid games with sparse rewards.</p>

<p>For a detailed breakdown of our methodology, results, and insights from the competition, please read our blog post on Medium.</p>]]></content><author><name>Dries Smit</name></author><summary type="html"><![CDATA[We present our winning solution for the ARC-AGI-3 Agent Preview Competition.]]></summary></entry><entry><title type="html">AlphaWrite: Inference-Time Compute Scaling for Writing</title><link href="https://tufalabs.ai/research/alphawrite/" rel="alternate" type="text/html" title="AlphaWrite: Inference-Time Compute Scaling for Writing" /><published>2025-06-06T00:50:53+02:00</published><updated>2025-06-06T00:50:53+02:00</updated><id>https://tufalabs.ai/research/AlphaWrite</id><content type="html" xml:base="https://tufalabs.ai/research/alphawrite/"><![CDATA[<p>You can try AlphaWrite out here.
<strong>Code Repository</strong>: <a href="https://github.com/tamassimonds/AlphaEvolveWritting">AlphaWrite on GitHub</a></p>

<p>Large language models have demonstrated remarkable improvements in performance through increased inference-time compute on quantitative reasoning tasks, particularly in mathematics and coding. However, the creative domain—where outputs are inherently highly subjective and difficult to evaluate—has seen limited exploration of systematic approaches to scale inference-time compute effectively.</p>

<p>In this work, we introduce Alpha Writing, a novel framework for scaling inference-time compute in creative text generation. Inspired by AlphaEvolve and other evolutionary algorithms, our approach combines iterative story generation with Elo-based evaluation to systematically improve narrative quality. Rather than relying on single-shot generation or simple resampling, Alpha Writing creates a dynamic ecosystem where stories compete, evolve, and improve through multiple generations.</p>

<p>Our method addresses a critical gap in the field: while we can easily scale compute for tasks with clear correctness criteria, creative domains have lacked principled approaches for leveraging additional inference resources. By treating story generation as an evolutionary process guided by pairwise preferences, we demonstrate that creative output quality can be systematically improved through increased compute allocation.</p>

<p>We further demonstrate the scalability of these methods by distilling the enhanced stories back into the base model, creating a stronger foundation for subsequent rounds of Alpha Writing. This recursive cycle—where improved outputs become training data for an enhanced model that can generate even better stories—offers promising potential for self-improving writing models.</p>

<h1 id="methodology">Methodology</h1>

<p><img src="/assets/posts/alphawrite/AlphaWriteProcess.png" alt="image.png" /></p>

<h2 id="overview">Overview</h2>

<p>Alpha Writing employs an evolutionary approach to improve story quality through iterative generation and selection. The process consists of four main stages: (1) diverse initial story generation, (2) pairwise comparison using Elo rankings, (3) evolutionary refinement of top-performing stories, and (4) repetition of stages (2) and (3) across multiple generations to progressively enhance narrative quality.</p>

<h2 id="initial-story-generation">Initial Story Generation</h2>

<p>To establish a diverse starting population, we generate a large corpus of initial stories with systematic variation. Each story is generated with two randomized parameters:</p>

<ul>
  <li><strong>Author style</strong>: The model is prompted to write in the style of different authors</li>
  <li><strong>Theme</strong>: Each generation focuses on a different narrative theme</li>
</ul>

<p>This approach ensures broad exploration of the creative space and prevents early convergence on a single narrative style or structure.</p>

<h2 id="judging-and-elo-ranking">Judging and Elo Ranking</h2>

<p>Stories are evaluated through pairwise comparisons using an LLM judge. The judge is provided with:</p>

<ul>
  <li>A detailed evaluation rubric focusing on narrative quality metrics</li>
  <li>Two stories to compare</li>
  <li>Instructions to select the superior story</li>
</ul>

<p>The rubric improves consistency in judgments by providing clear evaluation criteria. Based on these pairwise comparisons, we update Elo ratings for each story, creating a dynamic ranking system that captures relative quality differences. We use base Elo of 1200 and K-factor of 32. For our experiments we use the same model as the judge and generator</p>

<h2 id="story-evolution">Story Evolution</h2>

<p>After establishing rankings through pairwise comparisons, we implement an evolutionary process to iteratively improve story quality:</p>

<p><strong>1. Selection</strong>: Select top-performing stories as foundation for next generation</p>

<p><strong>2. Variation Generation</strong>: Generate variants using randomly sampled improvement objectives (narrative structure, character development, emotional resonance, dialogue, thematic depth, descriptive detail, plot tension, prose style). Random sampling maintains creative diversity.</p>

<p><strong>3. Population Update</strong>: Retain high-performers, replace lower-ranked stories with variants</p>

<p><strong>4. Re-ranking</strong>: Fresh pairwise comparisons on updated population</p>

<p><strong>5. Iteration</strong>: Repeat across generations, allowing successful elements to propagate</p>

<h2 id="evaluation-protocol">Evaluation Protocol</h2>

<p>Evaluating creative output presents significant challenges due to subjective preferences and high variance in story content. Our evaluation approach includes:</p>

<ul>
  <li><strong>Model selection</strong>: Focus on smaller models where improvements are more pronounced</li>
  <li><strong>Story length</strong>: Restrict to stories under 500 words to enable easier comparison</li>
  <li><strong>Prompt design</strong>: Use open-ended prompts to allow models to demonstrate narrative crafting abilities</li>
  <li><strong>Data collection</strong>: 120 preference comparisons per experiment to establish statistical significance</li>
  <li><strong>Evaluation Protocol:</strong> Evaluators use the same rubric we provide to the LLM judge to score which of the two responses they prefer</li>
</ul>

<p>Initial generations often exhibited fundamental narrative issues including poor story arcs and structural problems, making improvements through evolution particularly noticeable. We compare performance against initial model-generated stories and stories improved through repeated prompting.</p>

<p>We acknowledge that our evaluation methodology, while establishing statistically significant improvements, would benefit from more comprehensive data collection. We simply seek to demonstrate a statistically significant signal that this method works—quantifying the actual improvement is difficult and would require significantly more diverse data collection.</p>

<p>We found quality differences were subtle in opening lines but became pronounced in longer stories, where structural coherence and narrative flow showed clear improvement. However, evaluating these stories remains genuinely difficult—they diverge so dramatically in theme, style, and approach that determining which is “better” becomes largely subjective and dependent on reader preference.</p>

<h3 id="results">Results</h3>

<p>For evaluation we used Llama 3.1 8B and generated 60 initial stories, selected the top 5 performers, and created 5 variants of each. This evolution process was repeated for 5 generations</p>

<p><img src="/assets/posts/alphawrite/AlphaWriteBench.png" alt="image.png" /></p>

<p>Alpha Writing demonstrates substantial improvements in story quality when evaluated through pairwise human preferences. Testing with Llama 3.1 8B revealed:</p>

<ul>
  <li><strong>72% preference rate</strong> over initial story generations (95 % CI 63 % – 79 %)</li>
  <li><strong>62% preference rate</strong> over sequential-prompting baseline (95 % CI 53 % – 70 %)</li>
</ul>

<p>These results indicate that the evolutionary approach significantly outperforms both single-shot generation and traditional inference-time scaling methods for creative writing tasks.</p>

<p><img src="/assets/posts/alphawrite/Demo.png" alt="image.png" /></p>

<h1 id="recursive-self-improvement-through-alphawrite-distillation">Recursive Self-Improvement Through AlphaWrite Distillation</h1>

<p>An intriguing possibility emerges when considering inference scaling techniques like AlphaEvolve or AlphaWrite: could we create a self-improving loop through using inference scaling to improve results then distill back down and repeat?</p>

<h2 id="the-core-concept">The Core Concept</h2>

<p>The process would work as follows:</p>

<ol>
  <li>Apply AlphaWrite techniques to generate improved outputs from the current model</li>
  <li>Distill these enhanced outputs back into training data for the base model</li>
  <li>Reapply AlphaWrite techniques to this improved base, continuing the cycle</li>
</ol>

<h2 id="experiments">Experiments</h2>

<p>We explored this concept through preliminary testing:</p>

<ul>
  <li><strong>Generation Phase</strong>: Ran AlphaWrite with 60 initial questions, top 5 questions per batch, 5 variations of each for 5 generations. Ran process 10 times generating 50 stories in total</li>
  <li><strong>Selection</strong>: Identified the top 10 highest-quality stories of the final batch</li>
  <li><strong>Fine-tuning</strong>: Used these curated stories to fine-tune Llama 3.1 8B</li>
  <li><strong>Iteration</strong>: Repeated the process with the enhanced model</li>
</ul>

<p>This recursive approach theoretically enables continuous self-improvement, where each iteration builds upon the strengths of the previous generation, potentially leading to increasingly sophisticated capabilities without additional human-generated training data.</p>

<h2 id="results-1">Results</h2>

<p><img src="/assets/posts/alphawrite/LLama3.1Results.png" alt="image.png" /></p>

<p>We observed a 56% (95 % CI 47 % – 65 %) preference rate over the base model. While this improvement falls within the statistical significance range for this experiment, collecting sufficient preference data to achieve statistical significance would be prohibitively expensive.</p>

<h2 id="limitations">Limitations</h2>

<p><strong>Prompt Sensitivity</strong>: The quality and diversity of generated stories are highly dependent on the specific prompts used. Our choice of author styles and themes introduces inherent bias that may favor certain narrative approaches over others. Different prompt sets could yield substantially different results.</p>

<p><strong>Evaluation Challenges</strong>: The subjective nature of creative quality makes definitive assessment difficult. Our 120 preference comparisons represent a small sample of possible reader preferences.</p>

<p><strong>Convergence Risks</strong>: Extended evolution could lead to homogenization, where stories converge on particular “winning” formulas rather than maintaining true creative diversity. We observed early signs of this in later generations.</p>

<h3 id="beyond-creative-writing">Beyond Creative Writing</h3>

<p>The Alpha Writing framework extends far beyond narrative fiction. We’ve already employed it in drafting sections of this paper, demonstrating its versatility across writing domains. The approach can be adapted for:</p>

<p><strong>Targeted Generation</strong>: By incorporating specific rubrics, Alpha Writing can optimize individual components of larger works—generating compelling introductions, crafting precise technical explanations, or developing persuasive conclusions. This granular control enables writers to iteratively improve specific weaknesses in their work.</p>

<p><strong>Domain-Specific Applications</strong>: The framework naturally adapts to technical documentation, academic writing, marketing copy, and other specialized formats. Each domain simply requires appropriate evaluation criteria and judge training.</p>

<p><strong>Model Enhancement</strong>: Perhaps most significantly, Alpha Writing offers a systematic approach to improving language models’ general writing capabilities. By generating diverse, high-quality training data through evolutionary refinement, we can potentially bootstrap better foundation models—creating a virtuous cycle where improved models generate even better training data for future iterations.</p>

<p>This positions Alpha Writing not just as a tool for end-users, but as potentially a fundamental technique for advancing the writing capabilities of AI systems themselves.</p>

<h3 id="conclusion">Conclusion</h3>

<p>Alpha Writing demonstrates that creative tasks can benefit from systematic inference-time compute scaling through evolutionary approaches. Our results show consistent improvements over both baseline generation and sequential prompting methods, suggesting that the apparent intractability of scaling compute for creative domains may be addressable through appropriate algorithmic frameworks.</p>

<p><strong>Code Repository</strong>: <a href="https://github.com/tamassimonds/AlphaEvolveWritting">AlphaWrite on GitHub</a></p>

<h3 id="citation">Citation</h3>

<div class="language-bibtex highlighter-rouge"><div class="highlight"><pre class="highlight"><code><span class="nc">@article</span><span class="p">{</span><span class="nl">simonds2025alphawrite</span><span class="p">,</span>
  <span class="na">title</span><span class="p">=</span><span class="s">{AlphaWrite: Inference Time Compute Scaling for Writing}</span><span class="p">,</span>
  <span class="na">author</span><span class="p">=</span><span class="s">{Simonds, Toby}</span><span class="p">,</span>
  <span class="na">journal</span><span class="p">=</span><span class="s">{Tufa Labs Research}</span><span class="p">,</span>
  <span class="na">year</span><span class="p">=</span><span class="s">{2025}</span><span class="p">,</span>
  <span class="na">month</span><span class="p">=</span><span class="s">{June}</span><span class="p">,</span>
  <span class="na">url</span><span class="p">=</span><span class="s">{https://github.com/tamassimonds/AlphaEvolveWritting}</span>
<span class="p">}</span>
</code></pre></div></div>]]></content><author><name>Toby Simonds</name></author><summary type="html"><![CDATA[We introduce AlphaWrite, an inference-time scaling method for creative writing that uses evolutionary generation and ELO-based ranking to improve story quality.]]></summary></entry><entry><title type="html">Self-Rewarding, Self-Improving</title><link href="https://tufalabs.ai/research/self-rewarding/" rel="alternate" type="text/html" title="Self-Rewarding, Self-Improving" /><published>2025-05-12T00:50:53+02:00</published><updated>2025-05-12T00:50:53+02:00</updated><id>https://tufalabs.ai/research/Self-Rewarding</id><content type="html" xml:base="https://tufalabs.ai/research/self-rewarding/"><![CDATA[<h2 id="introduction">Introduction</h2>

<p>For RL to work, we need sufficiently many questions at the correct difficulty level. The
issue is that most difficult problems are very sparse and lack training data. Traditional methods to get around this have involved LADDER,
which solves this by getting the LLM to generate synthetic easier variants of the initial question
in a tree-like structure, allowing the model to use RL to bootstrap itself toward solving
harder problems.</p>]]></content><author><name>Toby Simonds</name></author><summary type="html"><![CDATA[We demonstrate that large language models can autonomously improve by judging their own solutions without reference answers, creating a complete self-learning loop that enhances performance beyond existing benchmarks.]]></summary></entry><entry><title type="html">LLMs for Engineering: Teaching Models to Design High-Powered Rockets</title><link href="https://tufalabs.ai/research/llm-rockets/" rel="alternate" type="text/html" title="LLMs for Engineering: Teaching Models to Design High-Powered Rockets" /><published>2025-04-24T00:50:53+02:00</published><updated>2025-04-24T00:50:53+02:00</updated><id>https://tufalabs.ai/research/LLM-Rockets</id><content type="html" xml:base="https://tufalabs.ai/research/llm-rockets/"><![CDATA[<h2 id="introduction">Introduction</h2>

<p>For RL to work, we need sufficiently many questions at the correct difficulty level. The
issue is that most difficult problems are very sparse and lack training data. Traditional methods to get around this have involved LADDER,
which solves this by getting the LLM to generate synthetic easier variants of the initial question
in a tree-like structure, allowing the model to use RL to bootstrap itself toward solving
harder problems.</p>]]></content><author><name>Toby Simonds</name></author><summary type="html"><![CDATA[We demonstrate that while current SOTA language models struggle with iterative self-improvement in rocket engineering challenges, augmenting them with reinforcement learning unlocks superhuman design capabilities that could revolutionize physical engineering domains.]]></summary></entry><entry><title type="html">Text to RL: Extracting High-Quality RL Questions from Text</title><link href="https://tufalabs.ai/research/textbooks-to-rl/" rel="alternate" type="text/html" title="Text to RL: Extracting High-Quality RL Questions from Text" /><published>2025-03-05T10:50:53+01:00</published><updated>2025-03-05T10:50:53+01:00</updated><id>https://tufalabs.ai/research/textbooks-to-rl</id><content type="html" xml:base="https://tufalabs.ai/research/textbooks-to-rl/"><![CDATA[<h2 id="the-data-challenge-in-modern-rl">The Data Challenge in Modern RL</h2>

<p>Reinforcement learning has demonstrated exceptional potential in solving complex problems, but a critical challenge looms as we scale these systems: we’re facing a severe shortage of high-quality RL questions. Current public datasets max out at around a million questions—below what’s needed to push next-generation RL systems to their full capabilities. This scarcity becomes even more pronounced outside the mathematics domain, where well-structured RL questions are particularly rare in domains like biology and physics. To achieve the full potential of RL, we need to expand our training data by orders of magnitude across diverse knowledge domains.</p>

<h2 id="our-approach-synthetic-question-generation-from-textbooks">Our Approach: Synthetic Question Generation from Textbooks</h2>

<p>We propose a straightforward yet powerful method for generating vast quantities of high-quality RL questions by leveraging existing textbook content. Here’s how it works:</p>

<ol>
  <li><strong>Extraction</strong>: We feed textbooks page by page into an LLM that extracts potential questions and solutions naturally embedded in the educational content.</li>
  <li><strong>Verification</strong>: Each extracted question passes through a verification LLM that confirms solvability using only the provided textbook content.</li>
  <li><strong>Filtering</strong>: We remove questions that depend on visual elements like tables, figures, or diagrams, ensuring our dataset consists of purely text-based, self-contained problems.</li>
</ol>

<p><img src="/assets/posts/textbooks-to-rl/TRL_diagram.png" alt="TRL Diagram" width="400" /></p>

<p>This pipeline allows us to generate around <strong>~3000</strong> valid questions per textbook, creating a scalable approach to building massive RL datasets.</p>

<h2 id="experimental-validation">Experimental Validation</h2>

<p>To validate our approach, we trained Qwen 2.5 7B Instruct on these synthetic questions and measured performance improvements on the MATH benchmark.</p>

<ul>
  <li>Models trained on our synthetic dataset showed a 9% absolute improvement on MATH, increasing performance from 75% to 84%. Performance was still improving; the run simply served as a proof of concept for the effectiveness of synthetic RL questions.</li>
  <li>This significant jump demonstrates that these synthetic RL questions provide a strong reinforcement learning signal</li>
  <li>The results suggest we can effectively bootstrap model performance using generated content</li>
</ul>

<p><img src="/assets/posts/textbooks-to-rl/TTL_bar_image.png" alt="TTL Bar Chart" /></p>

<h2 id="open-source-dataset">Open-Source Dataset</h2>

<p>As part of our commitment to advancing the field, we’re open-sourcing a sample dataset of 100K problems. This dataset provides researchers and developers with immediate access to high-quality synthetic questions for RL training and experimentation.</p>

<h2 id="discussion">Discussion</h2>

<p>As RL training becomes central to advancing model capabilities, the demand for high-quality training questions will continue to grow exponentially. Our textbook extraction approach offers several critical advantages that address current limitations:</p>

<ul>
  <li><strong>Token Efficiency</strong>: Perhaps the most profound implication of our work is the inherent efficiency of RL training with targeted questions compared to traditional next-token prediction. As models advance, structured questions may provide more learning signal per token than standard predictive training. While promising, rigorous ablation studies are needed to quantify this efficiency and determine optimal training approaches.</li>
  <li><strong>Domain Expansion</strong>: While mathematics has benefited from existing question datasets like MATH, fields such as biology, physics, chemistry, and medicine lack comparable resources for RL training. Our extraction pipeline works equally well across disciplines, generating high-quality questions from biology textbooks about protein folding, physics textbooks on quantum mechanics, or medical texts on diagnostic reasoning.</li>
  <li><strong>Adjustable Difficulty</strong>: The pipeline naturally extracts a range of question difficulties, but we can also engineer specific difficulty curves by generating hints alongside questions or by tuning extraction parameters. This enables curriculum learning approaches where models progressively tackle harder problems.</li>
  <li><strong>Customizable Content</strong>: Beyond difficulty tuning, we can target specific subdomains or reasoning types by selecting appropriate textbook sources or by fine-tuning our extraction criteria.</li>
</ul>

<h2 id="conclusion">Conclusion</h2>

<p>Our synthetic question generation approach offers a viable path to scaling RL training data by orders of magnitude. The initial results demonstrate that these synthetic questions drive meaningful performance improvements in trained models.</p>

<p>As we continue to refine this approach, we believe it will become an essential tool for advancing the frontier of reinforcement learning capabilities. The virtually unlimited supply of textbook content gives us confidence that we can generate the billions of diverse, high-quality questions needed for the next generation of RL systems.</p>

<hr />
<h2 id="resources">Resources</h2>

<ul>
  <li>
    <p><strong>Dataset</strong>: <a href="https://huggingface.co/datasets/TamasSimonds/TextbooksToRLQuestions-100k">TextbooksToRLQuestions-100k on Hugging Face</a></p>
  </li>
  <li>
    <p><strong>Code Repository</strong>: <a href="https://github.com/Tufalabs/BeyondNextTokenPrediction">BeyondNextTokenPrediction on GitHub</a></p>
  </li>
</ul>]]></content><author><name>Toby Simonds</name></author><summary type="html"><![CDATA[Turning textbooks into RL Questions]]></summary></entry></feed>