feat(route/thinkingmachines): add news route for Thinking Machines Lab#21609
Open
w3nhao wants to merge 3 commits intoDIYgod:masterfrom
Open
feat(route/thinkingmachines): add news route for Thinking Machines Lab#21609w3nhao wants to merge 3 commits intoDIYgod:masterfrom
w3nhao wants to merge 3 commits intoDIYgod:masterfrom
Conversation
Add route for Thinking Machines Lab (thinkingmachines.ai) news page. Founded by Mira Murati (ex-OpenAI CTO), the lab publishes news about their AI research and products. Closes #0
Contributor
|
Successfully generated as following: http://localhost:1200/thinkingmachines/news - Success ✔️<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:atom="http://www.w3.org/2005/Atom" version="2.0">
<channel>
<title>Thinking Machines Lab - News</title>
<link>https://thinkingmachines.ai/news/</link>
<atom:link href="http://localhost:1200/thinkingmachines/news" rel="self" type="application/rss+xml"></atom:link>
<description>Thinking Machines Lab - News - Powered by RSSHub</description>
<generator>RSSHub</generator>
<webMaster>contact@rsshub.app (RSSHub)</webMaster>
<language>en</language>
<lastBuildDate>Fri, 03 Apr 2026 00:33:44 GMT</lastBuildDate>
<ttl>5</ttl>
<item>
<title>Training LLMs to Predict World Events (Guest Post with Mantic)</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Training LLMs to Predict World Events (Guest Post with Mantic)</h1>
<div class="publish-metadata">
<span class="author">
<a href="https://enjeeneer.io/" rel="me noopener" target="_blank">Scott Jeen</a> and <a href="https://www.linkedin.com/in/matthew-aitchison-16aa8799/" rel="me noopener" target="_blank">Matthew Aitchison</a> in collaboration with others at <a href="https://www.mantic.com/" rel="me noopener" target="_blank">Mantic</a>
</span>
<span>
Mar 19, 2026
</span>
</div>
</div>
<div class="post-cover">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/cover.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
</div>
<article class="content ">
<p><em><a href="https://www.mantic.com/">Mantic</a> have been using Tinker since it launched. This guest post is a technical deep dive on what they have built so far.</em></p>
<p>The top AI forecasting systems are approaching superforecaster-level accuracy on geopolitics and current affairs.<label for="sn1" class="sidenote-number"></label><input type="checkbox" id="sn1" class="margin-toggle"><span class="sidenote"><a href="https://news.polymarket.com/p/its-over">It’s over</a> (Polymarket, 2026)</span> This is exciting because scalable, automated forecasting could significantly improve the quality of decision-making across the economy and in government.</p>
<p>To date, the most successful recipe in forecasting tournaments has been to combine an off-the-shelf LLM (like Gemini 3 or GPT-5) with forecasting-specific context-gathering. These models, to our knowledge, have not been explicitly trained for forecasting. Can we improve the recipe by replacing them with models fine-tuned specifically for forecasting?</p>
<p>We target “judgmental forecasting”: prediction problems that require human-like research and reasoning. Judgmental forecasting is needed for domains like geopolitics, politics, technology, business, and economic policy, where there often isn’t enough data for a standard statistical approach like time-series extrapolation. It was popularized by the book <a href="https://books.google.com/books?hl=en&amp;lr=&amp;id=_lMPDAAAQBAJ&amp;oi=fnd&amp;pg=PA1&amp;ots=5avwaBT55R&amp;sig=B2i7FYispHz5PtB6KVu2Ndt58cE">Superforecasting</a>, and now prediction markets like Polymarket and Kalshi.</p>
<p>In this post, we show it’s possible to significantly improve the forecasting performance of gpt-oss-120b using reinforcement learning. With Tinker, we fine-tune a model on around 10,000 binary questions of the form <em>“Will [event] occur before [date]?”</em>. We reward the model for putting greater probability on the correct real-world outcome.</p>
<p>In a head-to-head contest, the fine-tuned model achieves marginally superior performance to the frontier LLMs (see Figure 1), despite much lower initial performance. We find that providing forecast-specific context increases the gains from fine-tuning.</p>
<p>In the optimal <em>ensemble</em> of different models (which outperforms any single model), Grok 4 and our fine-tuned model are the most important contributors. The fine-tuned model learns a forecasting policy that is as accurate as the frontier LLMs, yet decorrelated from them.</p>
<p>Together, the results demonstrate that on-task training can extend the state-of-the-art in AI forecasting.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_1.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>RL fine-tuning makes gpt-oss-120b competitive with the frontier LLMs on questions from the Metaculus AI Benchmark Q2 2025. Naively predicting 50% on every question would get a score of 0, and perfect foresight would get a score of 100, per the construction of the Metaculus “baseline score”. Naively predicting 18.8% on every question (the rate at which the equivalent questions resolved “yes” in the previous tournament, Q1 2025) yields a score of 22.3, which we use to truncate the Y-axis. Fine-tuning improves gpt-oss-120b‘s score from 38.6 to 45.8, on par with the best general models.</figcaption>
</figure>
<h2 id="the-best-existing-recipe-uses-off-the-shelf-llms">The best existing recipe uses off-the-shelf LLMs</h2>
<p>The past two years have seen considerable progress in AI judgmental forecasting capabilities. In the Metaculus Cup, a major tournament for amateur and professional forecasters, the best AI systems now rival the top humans (Figure 2).</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_2.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Human and AI scores in the Metaculus Cup, a premier forecasting competition. Scores from the top 5 AI forecasters have been steadily improving since they first entered in the Summer of 2024. Mantic first entered in the Summer of 2025 and then in Fall 2025 beat the community prediction and the majority of professional forecasters. These results were without fine-tuning.</figcaption>
</figure>
<p>The trend has been driven by more capable off-the-shelf LLMs, and accelerated by more sophisticated forecasting architectures. Our architecture – which has performed well in recent Metaculus tournaments – consists of two standard phases: (1) a research phase, and (2) a prediction phase (Figure 3).<label for="sn3" class="sidenote-number"></label><input type="checkbox" id="sn3" class="margin-toggle"><span class="sidenote">This two-phase process appears in early work on AI forecasting. See: <a href="https://proceedings.neurips.cc/paper_files/paper/2024/hash/5a5acfd0876c940d81619c1dc60e7748-Abstract-Conference.html">Approaching human-level forecasting with language models</a> (Halawi et al., 2024); <a href="https://arxiv.org/abs/2206.15474">Forecasting Future World Events with Neural Networks</a> (Zou et al, 2022).</span></p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_3.png" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Mantic’s architecture. The research phase takes the forecasting question as input and performs deep research to collect information relevant to the question which goes into the prompt for the prediction LLM. The prediction LLM outputs chain-of-thought reasoning and specifies a probability distribution using specialized tools.</figcaption>
</figure>
<p>The research phase is conducted by deep research agents that collect the context needed to make a good prediction. For example, for the question “Will the United States attack Venezuela before 2026?”, search agents will find information about military buildup in the Caribbean, statements from President Trump, the health of the Venezuelan economy, and so on. The collected research is summarized into a prompt for the prediction phase.</p>
<p>The model’s task at the prediction phase is to use our specialized tools to output a probability distribution. In this post, we consider a canonical type of forecasting question: “Will [event] occur before [date]?”. We instruct the LLM to parameterize a mixture model for when the event will next occur – illustrated in Figure 4. The mixture model defines a cumulative distribution function, and from that we can read off the probability of the event occurring before the date specified in the original question.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_4.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Illustrative mixture model. The LLM selects: the number of components in the mixture, their parameters, and their respective weights. The LLM is prompted to select components capturing different scenarios that could lead to the event occurring. The final prediction is a weighted combination of the components.</figcaption>
</figure>
<p>In past tournaments, we’ve used off-the-shelf models as the prediction LLM. Existing literature has shown promising results from RL fine-tuning small models using a simple architecture.<label for="sn4" class="sidenote-number"></label><input type="checkbox" id="sn4" class="margin-toggle"><span class="sidenote"><a href="https://scholar.google.com/citations?view_op=view_citation&amp;hl=en&amp;user=oQuvQ0sAAAAJ&amp;citation_for_view=oQuvQ0sAAAAJ:u5HHmVD_uO8C">Outcome-based Reinforcement Learning to Predict the Future</a> (Turtel et al, 2025)</span> Can we improve frontier AI forecasting systems through on-task fine-tuning?</p>
<h2 id="training-details">Training details</h2>
<h3 id="datasets">Datasets</h3>
<p>We train the prediction LLM on ~10k questions about whether an event will happen by a given date. The questions are from August 2024 to December 2025 and the model’s knowledge cutoff is prior to that, so the resolution is known to us but not to the model. We generated the training set using an LLM pipeline similar to existing work.<label for="sn5" class="sidenote-number"></label><input type="checkbox" id="sn5" class="margin-toggle"><span class="sidenote"><a href="https://arxiv.org/abs/2601.22444">Automating Forecasting Question Generation and Resolution for AI Evaluation</a> (Bosse et al, 2026); <a href="https://arxiv.org/abs/2601.06336">Future-as-Label: Scalable Supervision from Real-World Outcomes</a> (Turtel et al, 2026)</span> Before training, we run the research phase for each question and store static prompts for the prediction LLM.</p>
<p>We test on unseen questions from the Q2 2025 Metaculus AI Benchmark Tournament.<label for="sn6" class="sidenote-number"></label><input type="checkbox" id="sn6" class="margin-toggle"><span class="sidenote"> The Fall 2025 iteration would have been a more obvious choice (for being more recent) but contains lower quality questions, indicated by less performance differentiated between strong and weak forecasters.</span> We compare models whose knowledge cutoff is before this tournament’s start date. The full list of questions can be accessed on <a href="https://gist.github.com/enjeeneer/86e24a52e6041a3d78e333bcab16984d">Github</a>.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_5.png" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Three example binary event questions from the Metaculus Q2 2025 AI Benchmark.</figcaption>
</figure>
<p>We evaluate using the baseline score, following the Metaculus platform. This is log scoring, i.e. ln(probability assigned to true outcome), rescaled such that 100 is the maximum possible score and 0 is the score for a uniform prediction (in our setting, 50%).</p>
<h3 id="implementation">Implementation</h3>
<p>We run the experiments on <a href="https://thinkingmachines.ai/tinker/">Tinker</a>. Of the models available through the API, we choose to train <a href="https://huggingface.co/openai/gpt-oss-120b">gpt-oss-120b</a> because of its strong initial performance — second only to <a href="https://huggingface.co/moonshotai/Kimi-K2.5">Kimi K2.5</a> — while being cheaper and faster.</p>
<p>We use a standard policy gradient algorithm with <a href="https://arxiv.org/abs/2402.03300">GRPO-style advantage normalisation</a> and <a href="https://arxiv.org/abs/2503.20783">no division by the standard deviation</a>. For rewards we use the Brier score which is <a href="https://sites.stat.washington.edu/raftery/Research/PDF/Gneiting2007jasa.pdf">strictly proper</a>. We found that the Brier score leads to more stable training than the log score, even though the log score is also strictly proper. This could be because the Brier score is bounded in [0, 1] and so produces lower-variance policy gradient estimates.</p>
<p>Open-source RL packages often use <a href="https://arxiv.org/abs/2309.06180">vLLM</a> for the sampling backend and <a href="https://arxiv.org/abs/2304.11277">FSDP</a> for the training backend. These can disagree on token probabilities produced by identical policies, which biases policy gradient estimates and destabilises training. We found these discrepancies to be lower on Tinker’s integrated infrastructure, but further mitigate them with <a href="https://fengyao.notion.site/off-policy-rl">an importance sampling correction on the advantages</a>.</p>
<p>The Brier score reward function is strictly monotonic in the predicted probability for a fixed outcome, so different rollouts almost always produce different rewards. This makes within-group reward ties extremely unlikely and lets us train with a relatively small group size (8) without needing to break ties or induce variance. We use a batch size of 64, as we find that <a href="https://thinkingmachines.ai/blog/lora/">larger batch sizes tend to destabilise training</a>.</p>
<h2 id="results">Results</h2>
<h3 id="fine-tuning-elevates-gpt-oss-120b-to-frontier-llm-performance">Fine-tuning elevates gpt-oss-120b to frontier LLM performance</h3>
<p>The model’s test set score improves through training (Figure 6), moving from an initial score of 38.6 mean baseline points per question (below all frontier models) to a final score of 45.8 mean baseline points per question (marginally above). This demonstrates that on-task forecasting training can provide a large performance uplift.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_6.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Test set baseline score of gpt-oss-120b with and without Mantic research and tools. In the model-only setup, test set performance improves, but never reaches the initial score of gpt-oss-120b with Mantic research and tools. With Mantic research and tools, gpt-oss-120b climbs 7 points through training and marginally exceeds the performance of Gemini 3 Pro. Training continued for further steps but performance no longer improved.</figcaption>
</figure>
<p>The LLM fine-tuned without the benefit of our pre-generated research phase, and without the tools to construct mixture models, gains only 3 points from training instead of 7. This suggests these prerequisites positively influence the optimization dynamics, in addition to improving initialization.</p>
<h3 id="the-fine-tuned-model-is-an-important-member-of-the-optimal-ensemble">The fine-tuned model is an important member of the optimal ensemble</h3>
<p>In human forecasting, there is a well-known “wisdom of the crowd” effect: aggregate forecasts from multiple people often outperforms any one individual. This effect, in part, explains the <a href="https://polymarket.com/accuracy">impressive accuracy of prediction markets</a>. Can we get the same benefit from ensembling the predictions of different LLMs?<label for="sn7" class="sidenote-number"></label><input type="checkbox" id="sn1" class="margin-toggle"><span class="sidenote">See also <a href="https://arxiv.org/abs/2402.19379">Wisdom of the Silicon Crowd: LLM Ensemble Prediction Capabilities Rival Human Crowd Accuracy</a> (Schoenegger et al, 2024)</span></p>
<p>To contribute to an ensemble, an LLM must be sufficiently accurate on its own and also decorrelated from other LLMs in the ensemble. Predictions from most frontier LLMs, while accurate, contribute little diversity to the top-performing model (in our case, Gemini 3 Pro) – Figure 7. Among the frontier LLMs, Grok 4 is the exception: its predictions score well whilst correlating less with other frontier LLMs.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_7.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Mean baseline score on binary event questions from the Metaculus Q2 AIB Benchmark with respect to their Jensson Shannon divergence from Gemini 3 Pro for a suite of closed source and open source LLMs. Marker colour indicates each model’s weight in the optimal 5-sample ensemble. The optimal ensemble consists of fine-tuned gpt-oss-120b (40%), gemini 3 pro (20%), gpt-5 (20%) and grok 4 (20%).</figcaption>
</figure>
<p>The optimal ensemble, on a budget of 5 samples, is 2 samples from our fine-tuned gpt-oss-120b plus 1 sample from each of Gemini 3 Pro, Grok 4, and GPT-5. We can test each model’s contribution by removing it, recomputing the optimal ensemble, and seeing how much the score degrades. We find that Grok 4 is the least replaceable, with fine-tuned gpt-oss-120b in second place. Other models can be replaced with little to no performance degradation (Figure 8).</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_8.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>When selecting an ensemble of frontier and open-source models, Grok 4 and fine-tuned gpt-oss-120b are the least replaceable. Model replaceability is defined as the reduction in score incurred when removing a model from the optimal ensemble. By definition, if a model is not included in the optimal ensemble,there is no cost to removing it.</figcaption>
</figure>
<p>GPT-5 and Gemini 3 Pro make similar predictions, and thus don’t benefit much from ensembling with each other. Both models improve from mixing their predictions with either Grok 4 or the fine-tuned gpt-oss-120b, and Grok 4 also benefits most from mixing with the fine-tuned model. In either optimal 3-way ensemble from Figure 9, the fine-tuned model gets about half of the total weight.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_9.png" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Three-way ensembles. Left: The optimal three-way ensemble between the fine-tuned gpt-oss-120b, Gemini 3 Pro and GPT-5 is weighted 56%, 26% and 18% respectively, depicted by the black star. Right: The optimal three-way ensemble between the fine-tuned gpt-oss-120b, Gemini 3 Pro and Grok 4 is weighted 48, 26% and 26% respectively, depicted by the white star.</figcaption>
</figure>
<h2 id="conclusions-and-next-steps">Conclusions and next steps</h2>
<p>We have shown that we can elevate the forecasting performance of gpt-oss-120b to match frontier LLMs with RL fine-tuning. This work can be extended in many ways, some of which we have already begun exploring:</p>
<ol>
<li><strong>Training larger models.</strong> Tinker enables training larger models with higher initial performance than gpt-oss, specifically Kimi K2.5.</li>
<li><strong>Training on all question formats.</strong> We are already training models on numerical questions such as economic indicators and multiple choice questions such as election results.</li>
<li><strong>Improved question sets.</strong> As the models become stronger forecasters, we need more challenging forecasting questions.</li>
<li><strong>Information retrieval inside the loop.</strong> We could give the prediction LLM tools for information retrieval and include this in the training loop.</li>
</ol>
<h2 id="citation">Citation</h2>
<p>Please cite this work as:</p>
<pre tabindex="0"><code>Jeen, Scott; Aitchison, Matthew; and Mantic, "Training LLMs to Predict World Events",
Thinking Machines Lab: News, Mar 2026.
</code></pre><p>Or use the BibTeX citation:</p>
<pre tabindex="0"><code>@article{scott2026forecasting,
author = {Scott Jeen, Matthew Aitchison, and Mantic},
title = {Training LLMs to Predict World Events},
journal = {Thinking Machines Lab: News},
year = {2026},
note = {https://thinkingmachines.ai/news/training-llms-to-predict-world-events/}
}
</code></pre>
</article>
<div class="paginator">
<a id="post-prev-link" class="link" href="https://thinkingmachines.ai/news/nvidia-partnership/" title="Thinking Machines Lab and NVIDIA Announce Long-Term Gigawatt-Scale Strategic Partnership">prev</a>
<a href="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/#" id="back-to-top" class="link" title="Go to top">back to top</a>
<a id="post-next-link"></a>
</div></description>
<link>https://thinkingmachines.ai/news/training-llms-to-predict-world-events/</link>
<guid isPermaLink="false">https://thinkingmachines.ai/news/training-llms-to-predict-world-events/</guid>
<pubDate>Wed, 18 Mar 2026 16:00:00 GMT</pubDate>
</item>
<item>
<title>Thinking Machines Lab and NVIDIA Announce Long-Term Gigawatt-Scale Strategic Partnership</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Thinking Machines Lab and NVIDIA Announce Long-Term Gigawatt-Scale Strategic Partnership</h1>
<div class="publish-metadata date-only">
<span>
Mar 10, 2026
</span>
</div>
</div>
<div class="post-cover compact">
<img data-zoomable="" src="https://thinkingmachines.ai/news/nvidia-partnership/images/jensen-huang-mira-murati-nvidia-debs-3127-2.jpg" alt="Cover image for Thinking Machines Lab and NVIDIA Announce Long-Term Gigawatt-Scale Strategic Partnership" class="cover-image" fetchpriority="high" decoding="async" loading="eager" referrerpolicy="no-referrer">
<p class="post-cover-caption">Jensen Huang (NVIDIA) and Mira Murati (Thinking Machines)</p>
</div>
<article class="content ">
<p>Thinking Machines Lab and NVIDIA announced today a multi-year strategic partnership to deploy at least one gigawatt of next-generation NVIDIA Vera Rubin systems to support Thinking Machines’ frontier model training and platforms delivering customizable AI at scale. Deployment on NVIDIA’s Vera Rubin platform is targeted for early next year.</p>
<p>The partnership also includes an effort to design training and serving systems for NVIDIA architectures and broaden access to frontier AI and open models for enterprises, research institutions, and the scientific community.</p>
<p>NVIDIA has also made a significant investment in Thinking Machines Lab to support the company’s long-term growth.</p>
<p>“AI is the most powerful knowledge discovery instrument in human history,” said Jensen Huang, founder and CEO of NVIDIA. “Thinking Machines has brought together a world-class team to advance the frontier of AI. We are thrilled to partner with Thinking Machines to realize their exciting vision for the future of AI.”</p>
<p>“NVIDIA’s technology is the foundation on which the entire field is built,” said Mira Murati, cofounder and CEO of Thinking Machines. “This partnership accelerates our capacity to build AI that people can shape and make their own, as it shapes human potential in turn.”</p>
<p>Building powerful AI systems that are understandable, customizable, and collaborative demands advances in research, design, and infrastructure at scale. This partnership provides that foundation, with the shared aim of ensuring that the most transformative technology of our time expands human capability.</p>
<img data-zoomable="" src="https://thinkingmachines.ai/news/nvidia-partnership/cover.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
</article>
<div class="paginator">
<a id="post-prev-link" class="link" href="https://thinkingmachines.ai/news/tinker-general-availability/" title="Tinker: General Availability and Vision Input">prev</a>
<a href="https://thinkingmachines.ai/news/nvidia-partnership/#" id="back-to-top" class="link" title="Go to top">back to top</a>
<a id="post-next-link" class="link" href="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/" title="Training LLMs to Predict World Events (Guest Post with Mantic)">next</a>
</div></description>
<link>https://thinkingmachines.ai/news/nvidia-partnership/</link>
<guid isPermaLink="false">https://thinkingmachines.ai/news/nvidia-partnership/</guid>
<pubDate>Mon, 09 Mar 2026 16:00:00 GMT</pubDate>
</item>
<item>
<title>Tinker: General Availability and Vision Input</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Tinker: General Availability and Vision Input</h1>
<div class="publish-metadata date-only">
<span>
Dec 12, 2025
</span>
</div>
</div>
<article class="content ">
<p>Today we are announcing four updates to Tinker:</p>
<ul>
<li>No more waitlist</li>
<li>New reasoning model: Kimi K2 Thinking</li>
<li>New inference interface that is compatible with the OpenAI API</li>
<li>Vision input support with Qwen3-VL</li>
</ul>
<h2 id="general-availability">General availability</h2>
<p>The waitlist is over! Everybody can use Tinker now; <a href="https://auth.thinkingmachines.ai/sign-up">sign up here</a> to get started. See the <a href="https://thinkingmachines.ai/tinker/">Tinker homepage</a> for available models and pricing, and check out the <a href="https://github.com/thinking-machines-lab/tinker-cookbook">Tinker cookbook</a> for code examples.</p>
<h2 id="more-reasoning-with-kimi-k2-thinking">More reasoning with Kimi K2 Thinking</h2>
<p>Users can now fine-tune Kimi K2 Thinking on Tinker. With a trillion parameters, Kimi K2 is the largest model in our lineup so far. It is built for long chains of reasoning and tool use.</p>
<h2 id="openai-api-compatible-sampling">OpenAI API-compatible sampling</h2>
<p>Tinker has a standard function for inference:</p>
<div class="highlight"><pre tabindex="0" class="chroma"><code class="language-python" data-lang="python"><span class="line"><span class="cl"><span class="n">prompt</span> <span class="o">=</span> <span class="n">types</span><span class="o">.</span><span class="n">ModelInput</span><span class="o">.</span><span class="n">from_ints</span><span class="p">(</span><span class="n">tokenizer</span><span class="o">.</span><span class="n">encode</span><span class="p">(</span><span class="s2">"The capital of France is"</span><span class="p">,))</span>
</span></span><span class="line"><span class="cl"><span class="n">params</span> <span class="o">=</span> <span class="n">types</span><span class="o">.</span><span class="n">SamplingParams</span><span class="p">(</span><span class="n">max_tokens</span><span class="o">=</span><span class="mi">20</span><span class="p">,</span> <span class="n">temperature</span><span class="o">=</span><span class="mf">0.0</span><span class="p">,</span> <span class="n">stop</span><span class="o">=</span><span class="p">[</span><span class="s2">"</span><span class="se">\n</span><span class="s2">"</span><span class="p">])</span>
</span></span><span class="line"><span class="cl"><span class="n">future</span> <span class="o">=</span> <span class="n">sampling_client</span><span class="o">.</span><span class="n">sample</span><span class="p">(</span><span class="n">prompt</span><span class="o">=</span><span class="n">prompt</span><span class="p">,</span> <span class="n">sampling_params</span><span class="o">=</span><span class="n">params</span><span class="p">)</span>
</span></span></code></pre></div><p>With this release, we have added OpenAI API-compatible scaffolding for quickly sampling from a model by specifying a path, even while it’s still training. This also means Tinker can now plug-and-play with any OpenAI API-compatible platform. See more information in our <a href="https://tinker-docs.thinkingmachines.ai/compatible-apis/openai">Tinker documentation</a>.</p>
<div class="highlight"><pre tabindex="0" class="chroma"><code class="language-python" data-lang="python"><span class="line"><span class="cl"><span class="n">response</span> <span class="o">=</span> <span class="n">openai_client</span><span class="o">.</span><span class="n">completions</span><span class="o">.</span><span class="n">create</span><span class="p">(</span>
</span></span><span class="line"><span class="cl"> <span class="n">model</span><span class="o">=</span><span class="s2">"tinker://0034d8c9-0a88-52a9-b2b7-bce7cb1e6fef:train:0/sampler_weights/000080"</span><span class="p">,</span>
</span></span><span class="line"><span class="cl"> <span class="n">prompt</span><span class="o">=</span><span class="s2">"The capital of France is"</span><span class="p">,</span>
</span></span><span class="line"><span class="cl"> <span class="n">max_tokens</span><span class="o">=</span><span class="mi">20</span><span class="p">,</span>
</span></span><span class="line"><span class="cl"> <span class="n">temperature</span><span class="o">=</span><span class="mf">0.0</span><span class="p">,</span>
</span></span><span class="line"><span class="cl"> <span class="n">stop</span><span class="o">=</span><span class="p">[</span><span class="s2">"</span><span class="se">\n</span><span class="s2">"</span><span class="p">],</span>
</span></span><span class="line"><span class="cl"><span class="p">)</span>
</span></span></code></pre></div><h2 id="vision-input-with-qwen3-vl">Vision input with Qwen3-VL</h2>
<p>We’ve added two vision models to Tinker: Qwen3-VL-30B-A3B-Instruct and Qwen3-VL-235B-A22B-Instruct. With these, users can process pictures, screenshots, and diagrams for a variety of applications.</p>
<p>To input images, just interleave together an <a href="https://tinker-docs.thinkingmachines.ai/api-reference/types#imagechunk-objects">ImageChunk</a> – consisting of your image, saved as bytes – with text chunks. For example:</p>
<div class="highlight"><pre tabindex="0" class="chroma"><code class="language-python" data-lang="python"><span class="line"><span class="cl"><span class="n">model_input</span> <span class="o">=</span> <span class="n">tinker</span><span class="o">.</span><span class="n">ModelInput</span><span class="p">(</span><span class="n">chunks</span><span class="o">=</span><span class="p">[</span>
</span></span><span class="line"><span class="cl"> <span class="n">tinker</span><span class="o">.</span><span class="n">types</span><span class="o">.</span><span class="n">ImageChunk</span><span class="p">(</span><span class="n">data</span><span class="o">=</span><span class="n">image_data</span><span class="p">,</span> <span class="nb">format</span><span class="o">=</span><span class="s2">"png"</span><span class="p">),</span>
</span></span><span class="line"><span class="cl"> <span class="n">tinker</span><span class="o">.</span><span class="n">types</span><span class="o">.</span><span class="n">EncodedTextChunk</span><span class="p">(</span><span class="n">tokens</span><span class="o">=</span><span class="n">tokenizer</span><span class="o">.</span><span class="n">encode</span><span class="p">(</span><span class="s2">"What is this?"</span><span class="p">)),</span>
</span></span><span class="line"><span class="cl"><span class="p">])</span>
</span></span></code></pre></div><p>These vision inputs can be used in a variety of applications out-of-the-box, including SFT and RL finetuning.</p>
<p>To demonstrate vision understanding in action, we are sharing <a href="https://github.com/thinking-machines-lab/tinker-cookbook/tree/main/tinker_cookbook/recipes/vlm_classifier">a new cookbook recipe for fine-tuning VLMs as image classifiers</a>. Qwen3-VL-235B-A22B-Instruct obtains reasonable accuracy even given just one example per class; performance improves with more labeled data.</p>
<h2 id="training-image-classifiers-with-tinker">Training image classifiers with Tinker</h2>
<p>To showcase Tinker’s new vision capabilities, we finetuned <a href="https://huggingface.co/Qwen/Qwen3-VL-235B-A22B-Instruct">Qwen3-VL-235B-A22B-Instruct</a> to classify images on four classic datasets:</p>
<ul>
<li><a href="https://data.caltech.edu/records/mzrjq-6wc02">Caltech 101</a>, a dataset of 101 general object categories.</li>
<li><a href="https://www.kaggle.com/datasets/eduardo4jesus/stanford-cars-dataset">Stanford Cars</a>, a dataset of car makes, models, and years.</li>
<li><a href="https://www.robots.ox.ac.uk/~vgg/data/flowers/102/">Oxford Flowers</a>, a dataset of flower species.</li>
<li><a href="https://www.robots.ox.ac.uk/~vgg/data/pets/">Oxford Pets</a>, a dataset of pet breeds.</li>
</ul>
<p>Since Qwen3-VL is a language model, we frame classification as text generation: given an image, the model outputs the class name. We compare this approach against a traditional vision baseline of finetuning a vision-only model — DINOv2-base. <a href="https://arxiv.org/pdf/2304.07193">DINOv2</a> is a self-supervised vision transformer that was trained to encode images, and is commonly used as a backbone for pure computer vision tasks. For DINOv2, we add a classification head that predicts a distribution over all N classes. Both models are fine-tuned with LoRA.</p>
<p>Labeled image data is scarce for many real-world use cases, so data efficiency is the primary measure we look at. We show the classification accuracy when sweeping across the number of labeled examples per class, starting with just a single one.</p>
<figure id="fig:qwen-v-dino">
<img data-zoomable="" src="https://thinkingmachines.ai/news/tinker-general-availability/images/vlm-graphs.png" referrerpolicy="no-referrer">
<figcaption>Comparison of fine-tuned Qwen3-VL-235-A22B and DINOv2 performance on simple image classification tasks.</figcaption>
</figure>
<p>In the limited-data regime, Qwen3-VL-235-A22B outperforms DINOv2. Not only is it a bigger model, but as a VLM, it also comes with language knowledge out-of-the-box (i.e. what a “golden retriever” or “sunflower” is). This general language-and-vision capability of Qwen3-VL makes it readily available for vision tasks beyond classification.</p>
<h2 id="happy-holidays">Happy Holidays</h2>
<p>Tinker exists to enable builders and researchers to train and customize state-of-the-art models. As always, we look forward to seeing what you build with Tinker. Happy holidays!</p>
</article>
<div class="paginator">
<a id="post-prev-link" class="link" href="https://thinkingmachines.ai/news/call-for-community-projects/" title="Tinker: Call for Community Projects">prev</a>
<a href="https://thinkingmachines.ai/news/tinker-general-availability/#" id="back-to-top" class="link" title="Go to top">back to top</a>
<a id="post-next-link" class="link" href="https://thinkingmachines.ai/news/nvidia-partnership/" title="Thinking Machines Lab and NVIDIA Announce Long-Term Gigawatt-Scale Strategic Partnership">next</a>
</div></description>
<link>https://thinkingmachines.ai/news/tinker-general-availability/</link>
<guid isPermaLink="false">https://thinkingmachines.ai/news/tinker-general-availability/</guid>
<pubDate>Thu, 11 Dec 2025 16:00:00 GMT</pubDate>
</item>
<item>
<title>Tinker: Call for Community Projects</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Tinker: Call for Community Projects</h1>
<div class="publish-metadata date-only">
<span>
Nov 7, 2025
</span>
</div>
</div>
<article class="content ">
<p>We launched <a href="https://thinkingmachines.ai/tinker/">Tinker</a> to enable builders and researchers to train models their own way, whether they’re conducting studies or customizing models for new applications. We plan to publish regular roundups of the coolest projects from the Tinker community, and <strong>we invite you to submit what you’ve been Tinkering on to be featured on our blog</strong>.</p>
<p>Below are some broad suggestions for what we hope to see from the Tinker featured projects, and some specific research directions we would particularly love to see pursued.</p>
<h2 id="guidelines-for-tinker-featured-projects">Guidelines for Tinker Featured Projects</h2>
<p>We’re interested in featuring ML research projects, AI-enabled research in other domains, custom models, and other contributions. Some examples:</p>
<ul>
<li>A reimplementation of a research project or tech report using Tinker, such as papers that compare algorithmic recipes or datasets.</li>
<li>Original research in machine learning, such as exploring new approaches to training or optimization or applying novel benchmarks and evaluations.</li>
<li>Research in a non-AI field that uses fine-tuned models, such as the work on mathematical theorem provers and chemistry models we <a href="https://thinkingmachines.ai/news/announcing-tinker/#:~:text=Groups%20at%20Princeton%2C%20Stanford">highlighted previously</a>.</li>
<li>Product prototypes built with Tinker, demoing a model that does something fresh or delightful.</li>
<li>Novel datasets and task environments for training models.</li>
<li>High-level libraries built on top of Tinker that enable less experienced practitioners to perform fine-tuning effectively.</li>
<li>Infrastructure contributions, such as a clean self-hosted implementation of the Tinker training API.</li>
</ul>
<p>Your submission should include a write-up and, preferably, an open-source release of your code. We encourage you to focus on rigor and clear evaluation in your write-ups: crisp charts, raw output examples, clear comparisons to alternative approaches or models on relevant benchmarks and metrics. Tinkering is experimenting — we want to feature diligent work and transparent results over novelty or hype.</p>
<p>Please send your projects and any related questions to <a href="https://thinkingmachines.ai/cdn-cgi/l/email-protection#2c58454247495e6c584445424745424b414d4f444542495f024d45135f594e46494f58116a494d58595e4948091e1c7c5e4346494f58091e1c"><span class="__cf_email__" data-cfemail="295d4047424c5b695d4140474240474e44484a4140474c5a074840">[email&nbsp;protected]</span></a> with “Featured Project” in the subject line.</p>
<h2 id="suggested-research-projects">Suggested research projects</h2>
<p>Here are some research directions that we would personally love to see explored and that Tinker can enable real progress on. We have <a href="https://github.com/thinking-machines-lab/tinker-project-ideas">created a repo</a> with detailed motivation and guidelines for each; we’ll be adding more resources to it over time to help researchers get started. We expect most project ideas to surprise us, but this short list could serve as inspiration.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/replicate-cai-with-base-models.md">Replicating Constitutional AI, starting from the base model.</a></strong> Though RLAIF is widely used, it’s most often bootstrapped from existing instruction-tuned models. This makes it difficult to separate the impact of the constitution from the impact of the data-generating model that interprets it. A study of Constitutional AI with and without instruction-tuned models in the pipeline would shed light on the use of constitutions and RLAIF.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/noisy-student.md">RLVR with Noisy student.</a></strong> Noisy student self-distillation was a popular technique in an earlier era of machine learning for making use of large unlabeled datasets, but it hasn’t been adapted widely to LLMs. One possible adaptation is to start RLVR with a small labeled training set and a large unlabeled one, then have the student apply labels to the latter set after each RL run and iterate.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/on-policy-context-distillation.md">On-Policy Context Distillation.</a></strong> Context distillation trains a student model with empty context on a teacher model with long and detailed context. Prior work used off-policy distillation — fine-tuning on teacher samples. We have found that <a href="https://thinkingmachines.ai/blog/on-policy-distillation/">on-policy distillation</a> is often much more effective; it would be useful to compare the two approaches for context distillation in particular.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/memorization-empirical-study.md">RL memory test.</a></strong> Our <a href="https://thinkingmachines.ai/blog/lora/">post on LoRA</a> presented theoretical arguments on the rate of information acquisition by both SFT and RL. You can set up a toy environment where RL must learn a completely random number sequence, to compare the empirical learning rate under various reward functions to the theoretical estimate.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/direct-rl-on-pairwise-judge.md">Direct RL on pairwise judge.</a></strong> RLHF and RLAIF use datasets of pairwise preferences, which are used to train a reward model, which is then used in RL. As an alternative “direct” approach, we can do RL using a prompted model that does pairwise comparisons, without training the reward model. It would be interesting to do experiments comparing the direct and indirect approaches.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/replicate-open-character-training.md">Replicate Open Character Training.</a></strong> Replicate the recent paper on <a href="https://arxiv.org/abs/2511.01689">Open Character Training</a> using Tinker.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/gan-joke-generation.md">GAN for jokes.</a></strong> In domains such as humor, it is easier to curate a human-vetted set of demonstrations than to train a reliable judge or reward model. Try implementing GAN-style training for a joke evaluator and joke generator that can craft a joke with a requested subject and keywords.</p>
<h2 id="tips-for-high-quality-ml-experiments">Tips for high-quality ML experiments</h2>
<p>In closing, we want to offer a few guidelines for running quality ML studies, the same guidelines we strive to adhere to internally when running experiments and documenting the results.</p>
<p>We encourage researchers to apply multiple analyses for examining each result. When creating datasets or environments, we recommend training a range of models and applying different evals. When developing novel methods, we suggest comparing to simpler baseline methods and sweeping hyperparameters that performance is sensitive to, particularly learning rate.</p>
<p>We’d love to see your reasoning in the write-up: assumptions you made, how your approach diverges from previous reports, and what motivated each change. We hope to see examples of the raw data and model rollouts, along with the summarized results. Finally, we appreciate crisp and detailed write-ups with clean and well-labeled charts and illustrations of the inner workings of the methods used.</p>
<p>We are excited to see what our community creates with Tinker, and hope that our featured projects will inspire your own work.</p>
</article>
<div class="paginator">
<a id="post-prev-link" class="link" href="https://thinkingmachines.ai/news/tinker-research-and-teaching-grants/" title="Tinker: Announcing Research and Teaching Grants">prev</a>
<a href="https://thinkingmachines.ai/news/call-for-community-projects/#" id="back-to-top" class="link" title="Go to top">back to top</a>
<a id="post-next-link" class="link" href="https://thinkingmachines.ai/news/tinker-general-availability/" title="Tinker: General Availability and Vision Input">next</a>
</div></description>
<link>https://thinkingmachines.ai/news/call-for-community-projects/</link>
<guid isPermaLink="false">https://thinkingmachines.ai/news/call-for-community-projects/</guid>
<pubDate>Thu, 06 Nov 2025 16:00:00 GMT</pubDate>
</item>
<item>
<title>Tinker: Announcing Research and Teaching Grants</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Tinker: Announcing Research and Teaching Grants</h1>
<div class="publish-metadata date-only">
<span>
Oct 29, 2025
</span>
</div>
</div>
<article class="content ">
<p>We launched <a href="https://thinkingmachines.ai/news/announcing-tinker/">Tinker</a> nearly one month ago. Since then, researchers across academia and non-profits have been using Tinker to train custom models and advance their research.</p>
<p>Today, we’re launching research and teaching grants for Tinker access. As part of our commitment to open and collaborative science, we want to make it as easy as possible for students and scholars to use Tinker. If your research or teaching involves training open-weight LLMs, we encourage you to apply.</p>
<p>We’re offering two types of grants to support your work:</p>
<ul>
<li>
<p><strong>Teaching Grants:</strong> We provide $250 in free credits per student for academic classes using Tinker, whether you’re integrating it into an assignment or enabling students to use Tinker for self-directed projects, this is sized to support your entire class for the duration of the course.</p>
</li>
<li>
<p><strong>Research Grants:</strong> We provide grants starting at $5,000 to support research projects and open-source software that uses Tinker.</p>
</li>
</ul>
<p>A selection of early grants we have awarded:</p>
<ul>
<li>
<p>Diyi Yang’s <a href="https://web.stanford.edu/class/cs329x/">Stanford class on Human-Centered LLMs</a> uses Tinker to compare different approaches for training personalized LLMs that capture unique writing styles and align with user habits.</p>
</li>
<li>
<p>Aviral Kumar and Katerina Fragkiadaki’s <a href="https://cmudeeprl.github.io/703website_f25/">CMU class on Deep RL</a> will use Tinker to enable class projects to experiment with state-of-the-art methods for training LLM and VLM based policies via RL.</p>
</li>
<li>
<p><a href="https://chemistry.stanford.edu/people/grant-m-rotskoff">Grant Rotskoff</a>’s lab at Stanford is fine-tuning small-molecule chemistry models with Tinker to help solve problems in computational chemistry.</p>
</li>
</ul>
<p><strong>Instructors</strong>, please apply for teaching grants <a href="https://form.typeform.com/to/JgPkuMvB"><strong>here.</strong></a></p>
<p><strong>Researchers</strong>, please apply for research grants <a href="https://form.typeform.com/to/E9wVFZJJ"><strong>here.</strong></a></p>
<p>We’re assessing applications on a rolling basis and will aim to respond within a week of your application.</p>
</article>
<div class="paginator">
<a id="post-prev-link" class="link" href="https://thinkingmachines.ai/news/announcing-tinker/" title="Announcing Tinker">prev</a>
<a href="https://thinkingmachines.ai/news/tinker-research-and-teaching-grants/#" id="back-to-top" class="link" title="Go to top">back to top</a>
<a id="post-next-link" class="link" href="https://thinkingmachines.ai/news/call-for-community-projects/" title="Tinker: Call for Community Projects">next</a>
</div></description>
<link>https://thinkingmachines.ai/news/tinker-research-and-teaching-grants/</link>
<guid isPermaLink="false">https://thinkingmachines.ai/news/tinker-research-and-teaching-grants/</guid>
<pubDate>Tue, 28 Oct 2025 16:00:00 GMT</pubDate>
</item>
<item>
<title>Announcing Tinker</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Announcing Tinker</h1>
<div class="publish-metadata date-only">
<span>
Oct 1, 2025
</span>
</div>
</div>
<div class="post-cover">
<img data-zoomable="" src="https://thinkingmachines.ai/news/announcing-tinker/svgs/tinker-cover.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
</div>
<article class="content ">
<p class="image-caption" style="text-align: center; margin-top: -1rem; margin-bottom: 2rem; font-size: 0.9rem; color: var(--fg-muted, #666);">
<a href="https://www.computerhistory.org/collections/catalog/X39.81/" target="_blank" rel="noopener">TinkerToy Computer</a> invented by <a href="https://en.wikipedia.org/wiki/Danny_Hillis" target="_blank" rel="noopener">Daniel Hillis</a> and <a href="https://en.wikipedia.org/wiki/Brian_Silverman" target="_blank" rel="noopener">Brian Silverman</a>
</p>
<p>Today, we are launching <a href="https://thinkingmachines.ai/tinker">Tinker</a>, a flexible API for fine-tuning language models. It empowers researchers and hackers to experiment with models by giving them control over the algorithms and data while we handle the complexity of distributed training. Tinker advances our mission of enabling more people to do research on cutting-edge models and customize them to their needs.</p>
<p>Tinker lets you fine-tune a range of large and small open-weight models, including large mixture-of-experts models such as Qwen-235B-A22B. Switching from a small model to a large one is as simple as changing a single string in your Python code.</p>
<p>Tinker is a managed service that runs on our internal clusters and training infrastructure. We handle scheduling, resource allocation, and failure recovery. This allows you to get small or large runs started immediately, without worrying about managing infrastructure. We use LoRA so that we can share the same pool of compute between multiple training runs, lowering costs.</p>
<p>Tinker’s API gives you low-level primitives like <code>forward_backward</code> and <code>sample</code>, which can be used to express most common post-training methods. Even so, achieving good results requires getting many details right. That’s why we’re releasing an open-source library, the <a href="http://github.com/thinking-machines-lab/tinker-cookbook">Tinker Cookbook</a>, with modern implementations of post-training methods that run on top of the Tinker API.</p>
<p>Groups at Princeton, Stanford, Berkeley, and Redwood Research have already been using Tinker:</p>
<ul>
<li>The <a href="https://blog.goedel-prover.com/">Princeton Goedel Team</a> trained mathematical theorem provers</li>
<li>The <a href="https://statmech.stanford.edu/">Rotskoff Chemistry group</a> at Stanford fine-tuned a model to complete chemistry reasoning tasks</li>
<li><a href="https://sky.cs.berkeley.edu/project/skyrl/">Berkeley’s SkyRL group</a> ran experiments on a custom async off-policy RL training loop with multi-agents and multi-turn tool-use.</li>
<li><a href="https://www.redwoodresearch.org/">Redwood Re |
Contributor
|
Successfully generated as following: http://localhost:1200/thinkingmachines/news - Success ✔️<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:atom="http://www.w3.org/2005/Atom" version="2.0">
<channel>
<title>Thinking Machines Lab - News</title>
<link>https://thinkingmachines.ai/news/</link>
<atom:link href="http://localhost:1200/thinkingmachines/news" rel="self" type="application/rss+xml"></atom:link>
<description>Thinking Machines Lab - News - Powered by RSSHub</description>
<generator>RSSHub</generator>
<webMaster>contact@rsshub.app (RSSHub)</webMaster>
<language>en</language>
<lastBuildDate>Fri, 03 Apr 2026 01:22:08 GMT</lastBuildDate>
<ttl>5</ttl>
<item>
<title>Training LLMs to Predict World Events (Guest Post with Mantic)</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Training LLMs to Predict World Events (Guest Post with Mantic)</h1>
<div class="publish-metadata">
<span class="author">
<a href="https://enjeeneer.io/" rel="me noopener" target="_blank">Scott Jeen</a> and <a href="https://www.linkedin.com/in/matthew-aitchison-16aa8799/" rel="me noopener" target="_blank">Matthew Aitchison</a> in collaboration with others at <a href="https://www.mantic.com/" rel="me noopener" target="_blank">Mantic</a>
</span>
<span>
Mar 19, 2026
</span>
</div>
</div>
<div class="post-cover">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/cover.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
</div>
<article class="content ">
<p><em><a href="https://www.mantic.com/">Mantic</a> have been using Tinker since it launched. This guest post is a technical deep dive on what they have built so far.</em></p>
<p>The top AI forecasting systems are approaching superforecaster-level accuracy on geopolitics and current affairs.<label for="sn1" class="sidenote-number"></label><input type="checkbox" id="sn1" class="margin-toggle"><span class="sidenote"><a href="https://news.polymarket.com/p/its-over">It’s over</a> (Polymarket, 2026)</span> This is exciting because scalable, automated forecasting could significantly improve the quality of decision-making across the economy and in government.</p>
<p>To date, the most successful recipe in forecasting tournaments has been to combine an off-the-shelf LLM (like Gemini 3 or GPT-5) with forecasting-specific context-gathering. These models, to our knowledge, have not been explicitly trained for forecasting. Can we improve the recipe by replacing them with models fine-tuned specifically for forecasting?</p>
<p>We target “judgmental forecasting”: prediction problems that require human-like research and reasoning. Judgmental forecasting is needed for domains like geopolitics, politics, technology, business, and economic policy, where there often isn’t enough data for a standard statistical approach like time-series extrapolation. It was popularized by the book <a href="https://books.google.com/books?hl=en&amp;lr=&amp;id=_lMPDAAAQBAJ&amp;oi=fnd&amp;pg=PA1&amp;ots=5avwaBT55R&amp;sig=B2i7FYispHz5PtB6KVu2Ndt58cE">Superforecasting</a>, and now prediction markets like Polymarket and Kalshi.</p>
<p>In this post, we show it’s possible to significantly improve the forecasting performance of gpt-oss-120b using reinforcement learning. With Tinker, we fine-tune a model on around 10,000 binary questions of the form <em>“Will [event] occur before [date]?”</em>. We reward the model for putting greater probability on the correct real-world outcome.</p>
<p>In a head-to-head contest, the fine-tuned model achieves marginally superior performance to the frontier LLMs (see Figure 1), despite much lower initial performance. We find that providing forecast-specific context increases the gains from fine-tuning.</p>
<p>In the optimal <em>ensemble</em> of different models (which outperforms any single model), Grok 4 and our fine-tuned model are the most important contributors. The fine-tuned model learns a forecasting policy that is as accurate as the frontier LLMs, yet decorrelated from them.</p>
<p>Together, the results demonstrate that on-task training can extend the state-of-the-art in AI forecasting.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_1.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>RL fine-tuning makes gpt-oss-120b competitive with the frontier LLMs on questions from the Metaculus AI Benchmark Q2 2025. Naively predicting 50% on every question would get a score of 0, and perfect foresight would get a score of 100, per the construction of the Metaculus “baseline score”. Naively predicting 18.8% on every question (the rate at which the equivalent questions resolved “yes” in the previous tournament, Q1 2025) yields a score of 22.3, which we use to truncate the Y-axis. Fine-tuning improves gpt-oss-120b‘s score from 38.6 to 45.8, on par with the best general models.</figcaption>
</figure>
<h2 id="the-best-existing-recipe-uses-off-the-shelf-llms">The best existing recipe uses off-the-shelf LLMs</h2>
<p>The past two years have seen considerable progress in AI judgmental forecasting capabilities. In the Metaculus Cup, a major tournament for amateur and professional forecasters, the best AI systems now rival the top humans (Figure 2).</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_2.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Human and AI scores in the Metaculus Cup, a premier forecasting competition. Scores from the top 5 AI forecasters have been steadily improving since they first entered in the Summer of 2024. Mantic first entered in the Summer of 2025 and then in Fall 2025 beat the community prediction and the majority of professional forecasters. These results were without fine-tuning.</figcaption>
</figure>
<p>The trend has been driven by more capable off-the-shelf LLMs, and accelerated by more sophisticated forecasting architectures. Our architecture – which has performed well in recent Metaculus tournaments – consists of two standard phases: (1) a research phase, and (2) a prediction phase (Figure 3).<label for="sn3" class="sidenote-number"></label><input type="checkbox" id="sn3" class="margin-toggle"><span class="sidenote">This two-phase process appears in early work on AI forecasting. See: <a href="https://proceedings.neurips.cc/paper_files/paper/2024/hash/5a5acfd0876c940d81619c1dc60e7748-Abstract-Conference.html">Approaching human-level forecasting with language models</a> (Halawi et al., 2024); <a href="https://arxiv.org/abs/2206.15474">Forecasting Future World Events with Neural Networks</a> (Zou et al, 2022).</span></p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_3.png" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Mantic’s architecture. The research phase takes the forecasting question as input and performs deep research to collect information relevant to the question which goes into the prompt for the prediction LLM. The prediction LLM outputs chain-of-thought reasoning and specifies a probability distribution using specialized tools.</figcaption>
</figure>
<p>The research phase is conducted by deep research agents that collect the context needed to make a good prediction. For example, for the question “Will the United States attack Venezuela before 2026?”, search agents will find information about military buildup in the Caribbean, statements from President Trump, the health of the Venezuelan economy, and so on. The collected research is summarized into a prompt for the prediction phase.</p>
<p>The model’s task at the prediction phase is to use our specialized tools to output a probability distribution. In this post, we consider a canonical type of forecasting question: “Will [event] occur before [date]?”. We instruct the LLM to parameterize a mixture model for when the event will next occur – illustrated in Figure 4. The mixture model defines a cumulative distribution function, and from that we can read off the probability of the event occurring before the date specified in the original question.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_4.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Illustrative mixture model. The LLM selects: the number of components in the mixture, their parameters, and their respective weights. The LLM is prompted to select components capturing different scenarios that could lead to the event occurring. The final prediction is a weighted combination of the components.</figcaption>
</figure>
<p>In past tournaments, we’ve used off-the-shelf models as the prediction LLM. Existing literature has shown promising results from RL fine-tuning small models using a simple architecture.<label for="sn4" class="sidenote-number"></label><input type="checkbox" id="sn4" class="margin-toggle"><span class="sidenote"><a href="https://scholar.google.com/citations?view_op=view_citation&amp;hl=en&amp;user=oQuvQ0sAAAAJ&amp;citation_for_view=oQuvQ0sAAAAJ:u5HHmVD_uO8C">Outcome-based Reinforcement Learning to Predict the Future</a> (Turtel et al, 2025)</span> Can we improve frontier AI forecasting systems through on-task fine-tuning?</p>
<h2 id="training-details">Training details</h2>
<h3 id="datasets">Datasets</h3>
<p>We train the prediction LLM on ~10k questions about whether an event will happen by a given date. The questions are from August 2024 to December 2025 and the model’s knowledge cutoff is prior to that, so the resolution is known to us but not to the model. We generated the training set using an LLM pipeline similar to existing work.<label for="sn5" class="sidenote-number"></label><input type="checkbox" id="sn5" class="margin-toggle"><span class="sidenote"><a href="https://arxiv.org/abs/2601.22444">Automating Forecasting Question Generation and Resolution for AI Evaluation</a> (Bosse et al, 2026); <a href="https://arxiv.org/abs/2601.06336">Future-as-Label: Scalable Supervision from Real-World Outcomes</a> (Turtel et al, 2026)</span> Before training, we run the research phase for each question and store static prompts for the prediction LLM.</p>
<p>We test on unseen questions from the Q2 2025 Metaculus AI Benchmark Tournament.<label for="sn6" class="sidenote-number"></label><input type="checkbox" id="sn6" class="margin-toggle"><span class="sidenote"> The Fall 2025 iteration would have been a more obvious choice (for being more recent) but contains lower quality questions, indicated by less performance differentiated between strong and weak forecasters.</span> We compare models whose knowledge cutoff is before this tournament’s start date. The full list of questions can be accessed on <a href="https://gist.github.com/enjeeneer/86e24a52e6041a3d78e333bcab16984d">Github</a>.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_5.png" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Three example binary event questions from the Metaculus Q2 2025 AI Benchmark.</figcaption>
</figure>
<p>We evaluate using the baseline score, following the Metaculus platform. This is log scoring, i.e. ln(probability assigned to true outcome), rescaled such that 100 is the maximum possible score and 0 is the score for a uniform prediction (in our setting, 50%).</p>
<h3 id="implementation">Implementation</h3>
<p>We run the experiments on <a href="https://thinkingmachines.ai/tinker/">Tinker</a>. Of the models available through the API, we choose to train <a href="https://huggingface.co/openai/gpt-oss-120b">gpt-oss-120b</a> because of its strong initial performance — second only to <a href="https://huggingface.co/moonshotai/Kimi-K2.5">Kimi K2.5</a> — while being cheaper and faster.</p>
<p>We use a standard policy gradient algorithm with <a href="https://arxiv.org/abs/2402.03300">GRPO-style advantage normalisation</a> and <a href="https://arxiv.org/abs/2503.20783">no division by the standard deviation</a>. For rewards we use the Brier score which is <a href="https://sites.stat.washington.edu/raftery/Research/PDF/Gneiting2007jasa.pdf">strictly proper</a>. We found that the Brier score leads to more stable training than the log score, even though the log score is also strictly proper. This could be because the Brier score is bounded in [0, 1] and so produces lower-variance policy gradient estimates.</p>
<p>Open-source RL packages often use <a href="https://arxiv.org/abs/2309.06180">vLLM</a> for the sampling backend and <a href="https://arxiv.org/abs/2304.11277">FSDP</a> for the training backend. These can disagree on token probabilities produced by identical policies, which biases policy gradient estimates and destabilises training. We found these discrepancies to be lower on Tinker’s integrated infrastructure, but further mitigate them with <a href="https://fengyao.notion.site/off-policy-rl">an importance sampling correction on the advantages</a>.</p>
<p>The Brier score reward function is strictly monotonic in the predicted probability for a fixed outcome, so different rollouts almost always produce different rewards. This makes within-group reward ties extremely unlikely and lets us train with a relatively small group size (8) without needing to break ties or induce variance. We use a batch size of 64, as we find that <a href="https://thinkingmachines.ai/blog/lora/">larger batch sizes tend to destabilise training</a>.</p>
<h2 id="results">Results</h2>
<h3 id="fine-tuning-elevates-gpt-oss-120b-to-frontier-llm-performance">Fine-tuning elevates gpt-oss-120b to frontier LLM performance</h3>
<p>The model’s test set score improves through training (Figure 6), moving from an initial score of 38.6 mean baseline points per question (below all frontier models) to a final score of 45.8 mean baseline points per question (marginally above). This demonstrates that on-task forecasting training can provide a large performance uplift.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_6.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Test set baseline score of gpt-oss-120b with and without Mantic research and tools. In the model-only setup, test set performance improves, but never reaches the initial score of gpt-oss-120b with Mantic research and tools. With Mantic research and tools, gpt-oss-120b climbs 7 points through training and marginally exceeds the performance of Gemini 3 Pro. Training continued for further steps but performance no longer improved.</figcaption>
</figure>
<p>The LLM fine-tuned without the benefit of our pre-generated research phase, and without the tools to construct mixture models, gains only 3 points from training instead of 7. This suggests these prerequisites positively influence the optimization dynamics, in addition to improving initialization.</p>
<h3 id="the-fine-tuned-model-is-an-important-member-of-the-optimal-ensemble">The fine-tuned model is an important member of the optimal ensemble</h3>
<p>In human forecasting, there is a well-known “wisdom of the crowd” effect: aggregate forecasts from multiple people often outperforms any one individual. This effect, in part, explains the <a href="https://polymarket.com/accuracy">impressive accuracy of prediction markets</a>. Can we get the same benefit from ensembling the predictions of different LLMs?<label for="sn7" class="sidenote-number"></label><input type="checkbox" id="sn1" class="margin-toggle"><span class="sidenote">See also <a href="https://arxiv.org/abs/2402.19379">Wisdom of the Silicon Crowd: LLM Ensemble Prediction Capabilities Rival Human Crowd Accuracy</a> (Schoenegger et al, 2024)</span></p>
<p>To contribute to an ensemble, an LLM must be sufficiently accurate on its own and also decorrelated from other LLMs in the ensemble. Predictions from most frontier LLMs, while accurate, contribute little diversity to the top-performing model (in our case, Gemini 3 Pro) – Figure 7. Among the frontier LLMs, Grok 4 is the exception: its predictions score well whilst correlating less with other frontier LLMs.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_7.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Mean baseline score on binary event questions from the Metaculus Q2 AIB Benchmark with respect to their Jensson Shannon divergence from Gemini 3 Pro for a suite of closed source and open source LLMs. Marker colour indicates each model’s weight in the optimal 5-sample ensemble. The optimal ensemble consists of fine-tuned gpt-oss-120b (40%), gemini 3 pro (20%), gpt-5 (20%) and grok 4 (20%).</figcaption>
</figure>
<p>The optimal ensemble, on a budget of 5 samples, is 2 samples from our fine-tuned gpt-oss-120b plus 1 sample from each of Gemini 3 Pro, Grok 4, and GPT-5. We can test each model’s contribution by removing it, recomputing the optimal ensemble, and seeing how much the score degrades. We find that Grok 4 is the least replaceable, with fine-tuned gpt-oss-120b in second place. Other models can be replaced with little to no performance degradation (Figure 8).</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_8.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>When selecting an ensemble of frontier and open-source models, Grok 4 and fine-tuned gpt-oss-120b are the least replaceable. Model replaceability is defined as the reduction in score incurred when removing a model from the optimal ensemble. By definition, if a model is not included in the optimal ensemble,there is no cost to removing it.</figcaption>
</figure>
<p>GPT-5 and Gemini 3 Pro make similar predictions, and thus don’t benefit much from ensembling with each other. Both models improve from mixing their predictions with either Grok 4 or the fine-tuned gpt-oss-120b, and Grok 4 also benefits most from mixing with the fine-tuned model. In either optimal 3-way ensemble from Figure 9, the fine-tuned model gets about half of the total weight.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_9.png" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Three-way ensembles. Left: The optimal three-way ensemble between the fine-tuned gpt-oss-120b, Gemini 3 Pro and GPT-5 is weighted 56%, 26% and 18% respectively, depicted by the black star. Right: The optimal three-way ensemble between the fine-tuned gpt-oss-120b, Gemini 3 Pro and Grok 4 is weighted 48, 26% and 26% respectively, depicted by the white star.</figcaption>
</figure>
<h2 id="conclusions-and-next-steps">Conclusions and next steps</h2>
<p>We have shown that we can elevate the forecasting performance of gpt-oss-120b to match frontier LLMs with RL fine-tuning. This work can be extended in many ways, some of which we have already begun exploring:</p>
<ol>
<li><strong>Training larger models.</strong> Tinker enables training larger models with higher initial performance than gpt-oss, specifically Kimi K2.5.</li>
<li><strong>Training on all question formats.</strong> We are already training models on numerical questions such as economic indicators and multiple choice questions such as election results.</li>
<li><strong>Improved question sets.</strong> As the models become stronger forecasters, we need more challenging forecasting questions.</li>
<li><strong>Information retrieval inside the loop.</strong> We could give the prediction LLM tools for information retrieval and include this in the training loop.</li>
</ol>
<h2 id="citation">Citation</h2>
<p>Please cite this work as:</p>
<pre tabindex="0"><code>Jeen, Scott; Aitchison, Matthew; and Mantic, "Training LLMs to Predict World Events",
Thinking Machines Lab: News, Mar 2026.
</code></pre><p>Or use the BibTeX citation:</p>
<pre tabindex="0"><code>@article{scott2026forecasting,
author = {Scott Jeen, Matthew Aitchison, and Mantic},
title = {Training LLMs to Predict World Events},
journal = {Thinking Machines Lab: News},
year = {2026},
note = {https://thinkingmachines.ai/news/training-llms-to-predict-world-events/}
}
</code></pre>
</article>
<div class="paginator">
<a id="post-prev-link" class="link" href="https://thinkingmachines.ai/news/nvidia-partnership/" title="Thinking Machines Lab and NVIDIA Announce Long-Term Gigawatt-Scale Strategic Partnership">prev</a>
<a href="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/#" id="back-to-top" class="link" title="Go to top">back to top</a>
<a id="post-next-link"></a>
</div></description>
<link>https://thinkingmachines.ai/news/training-llms-to-predict-world-events/</link>
<guid isPermaLink="false">https://thinkingmachines.ai/news/training-llms-to-predict-world-events/</guid>
<pubDate>Wed, 18 Mar 2026 16:00:00 GMT</pubDate>
</item>
<item>
<title>Thinking Machines Lab and NVIDIA Announce Long-Term Gigawatt-Scale Strategic Partnership</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Thinking Machines Lab and NVIDIA Announce Long-Term Gigawatt-Scale Strategic Partnership</h1>
<div class="publish-metadata date-only">
<span>
Mar 10, 2026
</span>
</div>
</div>
<div class="post-cover compact">
<img data-zoomable="" src="https://thinkingmachines.ai/news/nvidia-partnership/images/jensen-huang-mira-murati-nvidia-debs-3127-2.jpg" alt="Cover image for Thinking Machines Lab and NVIDIA Announce Long-Term Gigawatt-Scale Strategic Partnership" class="cover-image" fetchpriority="high" decoding="async" loading="eager" referrerpolicy="no-referrer">
<p class="post-cover-caption">Jensen Huang (NVIDIA) and Mira Murati (Thinking Machines)</p>
</div>
<article class="content ">
<p>Thinking Machines Lab and NVIDIA announced today a multi-year strategic partnership to deploy at least one gigawatt of next-generation NVIDIA Vera Rubin systems to support Thinking Machines’ frontier model training and platforms delivering customizable AI at scale. Deployment on NVIDIA’s Vera Rubin platform is targeted for early next year.</p>
<p>The partnership also includes an effort to design training and serving systems for NVIDIA architectures and broaden access to frontier AI and open models for enterprises, research institutions, and the scientific community.</p>
<p>NVIDIA has also made a significant investment in Thinking Machines Lab to support the company’s long-term growth.</p>
<p>“AI is the most powerful knowledge discovery instrument in human history,” said Jensen Huang, founder and CEO of NVIDIA. “Thinking Machines has brought together a world-class team to advance the frontier of AI. We are thrilled to partner with Thinking Machines to realize their exciting vision for the future of AI.”</p>
<p>“NVIDIA’s technology is the foundation on which the entire field is built,” said Mira Murati, cofounder and CEO of Thinking Machines. “This partnership accelerates our capacity to build AI that people can shape and make their own, as it shapes human potential in turn.”</p>
<p>Building powerful AI systems that are understandable, customizable, and collaborative demands advances in research, design, and infrastructure at scale. This partnership provides that foundation, with the shared aim of ensuring that the most transformative technology of our time expands human capability.</p>
<img data-zoomable="" src="https://thinkingmachines.ai/news/nvidia-partnership/cover.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
</article>
<div class="paginator">
<a id="post-prev-link" class="link" href="https://thinkingmachines.ai/news/tinker-general-availability/" title="Tinker: General Availability and Vision Input">prev</a>
<a href="https://thinkingmachines.ai/news/nvidia-partnership/#" id="back-to-top" class="link" title="Go to top">back to top</a>
<a id="post-next-link" class="link" href="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/" title="Training LLMs to Predict World Events (Guest Post with Mantic)">next</a>
</div></description>
<link>https://thinkingmachines.ai/news/nvidia-partnership/</link>
<guid isPermaLink="false">https://thinkingmachines.ai/news/nvidia-partnership/</guid>
<pubDate>Mon, 09 Mar 2026 16:00:00 GMT</pubDate>
</item>
<item>
<title>Tinker: General Availability and Vision Input</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Tinker: General Availability and Vision Input</h1>
<div class="publish-metadata date-only">
<span>
Dec 12, 2025
</span>
</div>
</div>
<article class="content ">
<p>Today we are announcing four updates to Tinker:</p>
<ul>
<li>No more waitlist</li>
<li>New reasoning model: Kimi K2 Thinking</li>
<li>New inference interface that is compatible with the OpenAI API</li>
<li>Vision input support with Qwen3-VL</li>
</ul>
<h2 id="general-availability">General availability</h2>
<p>The waitlist is over! Everybody can use Tinker now; <a href="https://auth.thinkingmachines.ai/sign-up">sign up here</a> to get started. See the <a href="https://thinkingmachines.ai/tinker/">Tinker homepage</a> for available models and pricing, and check out the <a href="https://github.com/thinking-machines-lab/tinker-cookbook">Tinker cookbook</a> for code examples.</p>
<h2 id="more-reasoning-with-kimi-k2-thinking">More reasoning with Kimi K2 Thinking</h2>
<p>Users can now fine-tune Kimi K2 Thinking on Tinker. With a trillion parameters, Kimi K2 is the largest model in our lineup so far. It is built for long chains of reasoning and tool use.</p>
<h2 id="openai-api-compatible-sampling">OpenAI API-compatible sampling</h2>
<p>Tinker has a standard function for inference:</p>
<div class="highlight"><pre tabindex="0" class="chroma"><code class="language-python" data-lang="python"><span class="line"><span class="cl"><span class="n">prompt</span> <span class="o">=</span> <span class="n">types</span><span class="o">.</span><span class="n">ModelInput</span><span class="o">.</span><span class="n">from_ints</span><span class="p">(</span><span class="n">tokenizer</span><span class="o">.</span><span class="n">encode</span><span class="p">(</span><span class="s2">"The capital of France is"</span><span class="p">,))</span>
</span></span><span class="line"><span class="cl"><span class="n">params</span> <span class="o">=</span> <span class="n">types</span><span class="o">.</span><span class="n">SamplingParams</span><span class="p">(</span><span class="n">max_tokens</span><span class="o">=</span><span class="mi">20</span><span class="p">,</span> <span class="n">temperature</span><span class="o">=</span><span class="mf">0.0</span><span class="p">,</span> <span class="n">stop</span><span class="o">=</span><span class="p">[</span><span class="s2">"</span><span class="se">\n</span><span class="s2">"</span><span class="p">])</span>
</span></span><span class="line"><span class="cl"><span class="n">future</span> <span class="o">=</span> <span class="n">sampling_client</span><span class="o">.</span><span class="n">sample</span><span class="p">(</span><span class="n">prompt</span><span class="o">=</span><span class="n">prompt</span><span class="p">,</span> <span class="n">sampling_params</span><span class="o">=</span><span class="n">params</span><span class="p">)</span>
</span></span></code></pre></div><p>With this release, we have added OpenAI API-compatible scaffolding for quickly sampling from a model by specifying a path, even while it’s still training. This also means Tinker can now plug-and-play with any OpenAI API-compatible platform. See more information in our <a href="https://tinker-docs.thinkingmachines.ai/compatible-apis/openai">Tinker documentation</a>.</p>
<div class="highlight"><pre tabindex="0" class="chroma"><code class="language-python" data-lang="python"><span class="line"><span class="cl"><span class="n">response</span> <span class="o">=</span> <span class="n">openai_client</span><span class="o">.</span><span class="n">completions</span><span class="o">.</span><span class="n">create</span><span class="p">(</span>
</span></span><span class="line"><span class="cl"> <span class="n">model</span><span class="o">=</span><span class="s2">"tinker://0034d8c9-0a88-52a9-b2b7-bce7cb1e6fef:train:0/sampler_weights/000080"</span><span class="p">,</span>
</span></span><span class="line"><span class="cl"> <span class="n">prompt</span><span class="o">=</span><span class="s2">"The capital of France is"</span><span class="p">,</span>
</span></span><span class="line"><span class="cl"> <span class="n">max_tokens</span><span class="o">=</span><span class="mi">20</span><span class="p">,</span>
</span></span><span class="line"><span class="cl"> <span class="n">temperature</span><span class="o">=</span><span class="mf">0.0</span><span class="p">,</span>
</span></span><span class="line"><span class="cl"> <span class="n">stop</span><span class="o">=</span><span class="p">[</span><span class="s2">"</span><span class="se">\n</span><span class="s2">"</span><span class="p">],</span>
</span></span><span class="line"><span class="cl"><span class="p">)</span>
</span></span></code></pre></div><h2 id="vision-input-with-qwen3-vl">Vision input with Qwen3-VL</h2>
<p>We’ve added two vision models to Tinker: Qwen3-VL-30B-A3B-Instruct and Qwen3-VL-235B-A22B-Instruct. With these, users can process pictures, screenshots, and diagrams for a variety of applications.</p>
<p>To input images, just interleave together an <a href="https://tinker-docs.thinkingmachines.ai/api-reference/types#imagechunk-objects">ImageChunk</a> – consisting of your image, saved as bytes – with text chunks. For example:</p>
<div class="highlight"><pre tabindex="0" class="chroma"><code class="language-python" data-lang="python"><span class="line"><span class="cl"><span class="n">model_input</span> <span class="o">=</span> <span class="n">tinker</span><span class="o">.</span><span class="n">ModelInput</span><span class="p">(</span><span class="n">chunks</span><span class="o">=</span><span class="p">[</span>
</span></span><span class="line"><span class="cl"> <span class="n">tinker</span><span class="o">.</span><span class="n">types</span><span class="o">.</span><span class="n">ImageChunk</span><span class="p">(</span><span class="n">data</span><span class="o">=</span><span class="n">image_data</span><span class="p">,</span> <span class="nb">format</span><span class="o">=</span><span class="s2">"png"</span><span class="p">),</span>
</span></span><span class="line"><span class="cl"> <span class="n">tinker</span><span class="o">.</span><span class="n">types</span><span class="o">.</span><span class="n">EncodedTextChunk</span><span class="p">(</span><span class="n">tokens</span><span class="o">=</span><span class="n">tokenizer</span><span class="o">.</span><span class="n">encode</span><span class="p">(</span><span class="s2">"What is this?"</span><span class="p">)),</span>
</span></span><span class="line"><span class="cl"><span class="p">])</span>
</span></span></code></pre></div><p>These vision inputs can be used in a variety of applications out-of-the-box, including SFT and RL finetuning.</p>
<p>To demonstrate vision understanding in action, we are sharing <a href="https://github.com/thinking-machines-lab/tinker-cookbook/tree/main/tinker_cookbook/recipes/vlm_classifier">a new cookbook recipe for fine-tuning VLMs as image classifiers</a>. Qwen3-VL-235B-A22B-Instruct obtains reasonable accuracy even given just one example per class; performance improves with more labeled data.</p>
<h2 id="training-image-classifiers-with-tinker">Training image classifiers with Tinker</h2>
<p>To showcase Tinker’s new vision capabilities, we finetuned <a href="https://huggingface.co/Qwen/Qwen3-VL-235B-A22B-Instruct">Qwen3-VL-235B-A22B-Instruct</a> to classify images on four classic datasets:</p>
<ul>
<li><a href="https://data.caltech.edu/records/mzrjq-6wc02">Caltech 101</a>, a dataset of 101 general object categories.</li>
<li><a href="https://www.kaggle.com/datasets/eduardo4jesus/stanford-cars-dataset">Stanford Cars</a>, a dataset of car makes, models, and years.</li>
<li><a href="https://www.robots.ox.ac.uk/~vgg/data/flowers/102/">Oxford Flowers</a>, a dataset of flower species.</li>
<li><a href="https://www.robots.ox.ac.uk/~vgg/data/pets/">Oxford Pets</a>, a dataset of pet breeds.</li>
</ul>
<p>Since Qwen3-VL is a language model, we frame classification as text generation: given an image, the model outputs the class name. We compare this approach against a traditional vision baseline of finetuning a vision-only model — DINOv2-base. <a href="https://arxiv.org/pdf/2304.07193">DINOv2</a> is a self-supervised vision transformer that was trained to encode images, and is commonly used as a backbone for pure computer vision tasks. For DINOv2, we add a classification head that predicts a distribution over all N classes. Both models are fine-tuned with LoRA.</p>
<p>Labeled image data is scarce for many real-world use cases, so data efficiency is the primary measure we look at. We show the classification accuracy when sweeping across the number of labeled examples per class, starting with just a single one.</p>
<figure id="fig:qwen-v-dino">
<img data-zoomable="" src="https://thinkingmachines.ai/news/tinker-general-availability/images/vlm-graphs.png" referrerpolicy="no-referrer">
<figcaption>Comparison of fine-tuned Qwen3-VL-235-A22B and DINOv2 performance on simple image classification tasks.</figcaption>
</figure>
<p>In the limited-data regime, Qwen3-VL-235-A22B outperforms DINOv2. Not only is it a bigger model, but as a VLM, it also comes with language knowledge out-of-the-box (i.e. what a “golden retriever” or “sunflower” is). This general language-and-vision capability of Qwen3-VL makes it readily available for vision tasks beyond classification.</p>
<h2 id="happy-holidays">Happy Holidays</h2>
<p>Tinker exists to enable builders and researchers to train and customize state-of-the-art models. As always, we look forward to seeing what you build with Tinker. Happy holidays!</p>
</article>
<div class="paginator">
<a id="post-prev-link" class="link" href="https://thinkingmachines.ai/news/call-for-community-projects/" title="Tinker: Call for Community Projects">prev</a>
<a href="https://thinkingmachines.ai/news/tinker-general-availability/#" id="back-to-top" class="link" title="Go to top">back to top</a>
<a id="post-next-link" class="link" href="https://thinkingmachines.ai/news/nvidia-partnership/" title="Thinking Machines Lab and NVIDIA Announce Long-Term Gigawatt-Scale Strategic Partnership">next</a>
</div></description>
<link>https://thinkingmachines.ai/news/tinker-general-availability/</link>
<guid isPermaLink="false">https://thinkingmachines.ai/news/tinker-general-availability/</guid>
<pubDate>Thu, 11 Dec 2025 16:00:00 GMT</pubDate>
</item>
<item>
<title>Tinker: Call for Community Projects</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Tinker: Call for Community Projects</h1>
<div class="publish-metadata date-only">
<span>
Nov 7, 2025
</span>
</div>
</div>
<article class="content ">
<p>We launched <a href="https://thinkingmachines.ai/tinker/">Tinker</a> to enable builders and researchers to train models their own way, whether they’re conducting studies or customizing models for new applications. We plan to publish regular roundups of the coolest projects from the Tinker community, and <strong>we invite you to submit what you’ve been Tinkering on to be featured on our blog</strong>.</p>
<p>Below are some broad suggestions for what we hope to see from the Tinker featured projects, and some specific research directions we would particularly love to see pursued.</p>
<h2 id="guidelines-for-tinker-featured-projects">Guidelines for Tinker Featured Projects</h2>
<p>We’re interested in featuring ML research projects, AI-enabled research in other domains, custom models, and other contributions. Some examples:</p>
<ul>
<li>A reimplementation of a research project or tech report using Tinker, such as papers that compare algorithmic recipes or datasets.</li>
<li>Original research in machine learning, such as exploring new approaches to training or optimization or applying novel benchmarks and evaluations.</li>
<li>Research in a non-AI field that uses fine-tuned models, such as the work on mathematical theorem provers and chemistry models we <a href="https://thinkingmachines.ai/news/announcing-tinker/#:~:text=Groups%20at%20Princeton%2C%20Stanford">highlighted previously</a>.</li>
<li>Product prototypes built with Tinker, demoing a model that does something fresh or delightful.</li>
<li>Novel datasets and task environments for training models.</li>
<li>High-level libraries built on top of Tinker that enable less experienced practitioners to perform fine-tuning effectively.</li>
<li>Infrastructure contributions, such as a clean self-hosted implementation of the Tinker training API.</li>
</ul>
<p>Your submission should include a write-up and, preferably, an open-source release of your code. We encourage you to focus on rigor and clear evaluation in your write-ups: crisp charts, raw output examples, clear comparisons to alternative approaches or models on relevant benchmarks and metrics. Tinkering is experimenting — we want to feature diligent work and transparent results over novelty or hype.</p>
<p>Please send your projects and any related questions to <a href="https://thinkingmachines.ai/cdn-cgi/l/email-protection#0f7b6661646a7d4f7b67666164666168626e6c6766616a7c216e66307c7a6d656a6c7b32496a6e7b7a7d6a6b2a3d3f5f7d60656a6c7b2a3d3f"><span class="__cf_email__" data-cfemail="4f3b2621242a3d0f3b27262124262128222e2c2726212a3c612e26">[email&nbsp;protected]</span></a> with “Featured Project” in the subject line.</p>
<h2 id="suggested-research-projects">Suggested research projects</h2>
<p>Here are some research directions that we would personally love to see explored and that Tinker can enable real progress on. We have <a href="https://github.com/thinking-machines-lab/tinker-project-ideas">created a repo</a> with detailed motivation and guidelines for each; we’ll be adding more resources to it over time to help researchers get started. We expect most project ideas to surprise us, but this short list could serve as inspiration.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/replicate-cai-with-base-models.md">Replicating Constitutional AI, starting from the base model.</a></strong> Though RLAIF is widely used, it’s most often bootstrapped from existing instruction-tuned models. This makes it difficult to separate the impact of the constitution from the impact of the data-generating model that interprets it. A study of Constitutional AI with and without instruction-tuned models in the pipeline would shed light on the use of constitutions and RLAIF.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/noisy-student.md">RLVR with Noisy student.</a></strong> Noisy student self-distillation was a popular technique in an earlier era of machine learning for making use of large unlabeled datasets, but it hasn’t been adapted widely to LLMs. One possible adaptation is to start RLVR with a small labeled training set and a large unlabeled one, then have the student apply labels to the latter set after each RL run and iterate.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/on-policy-context-distillation.md">On-Policy Context Distillation.</a></strong> Context distillation trains a student model with empty context on a teacher model with long and detailed context. Prior work used off-policy distillation — fine-tuning on teacher samples. We have found that <a href="https://thinkingmachines.ai/blog/on-policy-distillation/">on-policy distillation</a> is often much more effective; it would be useful to compare the two approaches for context distillation in particular.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/memorization-empirical-study.md">RL memory test.</a></strong> Our <a href="https://thinkingmachines.ai/blog/lora/">post on LoRA</a> presented theoretical arguments on the rate of information acquisition by both SFT and RL. You can set up a toy environment where RL must learn a completely random number sequence, to compare the empirical learning rate under various reward functions to the theoretical estimate.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/direct-rl-on-pairwise-judge.md">Direct RL on pairwise judge.</a></strong> RLHF and RLAIF use datasets of pairwise preferences, which are used to train a reward model, which is then used in RL. As an alternative “direct” approach, we can do RL using a prompted model that does pairwise comparisons, without training the reward model. It would be interesting to do experiments comparing the direct and indirect approaches.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/replicate-open-character-training.md">Replicate Open Character Training.</a></strong> Replicate the recent paper on <a href="https://arxiv.org/abs/2511.01689">Open Character Training</a> using Tinker.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/gan-joke-generation.md">GAN for jokes.</a></strong> In domains such as humor, it is easier to curate a human-vetted set of demonstrations than to train a reliable judge or reward model. Try implementing GAN-style training for a joke evaluator and joke generator that can craft a joke with a requested subject and keywords.</p>
<h2 id="tips-for-high-quality-ml-experiments">Tips for high-quality ML experiments</h2>
<p>In closing, we want to offer a few guidelines for running quality ML studies, the same guidelines we strive to adhere to internally when running experiments and documenting the results.</p>
<p>We encourage researchers to apply multiple analyses for examining each result. When creating datasets or environments, we recommend training a range of models and applying different evals. When developing novel methods, we suggest comparing to simpler baseline methods and sweeping hyperparameters that performance is sensitive to, particularly learning rate.</p>
<p>We’d love to see your reasoning in the write-up: assumptions you made, how your approach diverges from previous reports, and what motivated each change. We hope to see examples of the raw data and model rollouts, along with the summarized results. Finally, we appreciate crisp and detailed write-ups with clean and well-labeled charts and illustrations of the inner workings of the methods used.</p>
<p>We are excited to see what our community creates with Tinker, and hope that our featured projects will inspire your own work.</p>
</article>
<div class="paginator">
<a id="post-prev-link" class="link" href="https://thinkingmachines.ai/news/tinker-research-and-teaching-grants/" title="Tinker: Announcing Research and Teaching Grants">prev</a>
<a href="https://thinkingmachines.ai/news/call-for-community-projects/#" id="back-to-top" class="link" title="Go to top">back to top</a>
<a id="post-next-link" class="link" href="https://thinkingmachines.ai/news/tinker-general-availability/" title="Tinker: General Availability and Vision Input">next</a>
</div></description>
<link>https://thinkingmachines.ai/news/call-for-community-projects/</link>
<guid isPermaLink="false">https://thinkingmachines.ai/news/call-for-community-projects/</guid>
<pubDate>Thu, 06 Nov 2025 16:00:00 GMT</pubDate>
</item>
<item>
<title>Tinker: Announcing Research and Teaching Grants</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Tinker: Announcing Research and Teaching Grants</h1>
<div class="publish-metadata date-only">
<span>
Oct 29, 2025
</span>
</div>
</div>
<article class="content ">
<p>We launched <a href="https://thinkingmachines.ai/news/announcing-tinker/">Tinker</a> nearly one month ago. Since then, researchers across academia and non-profits have been using Tinker to train custom models and advance their research.</p>
<p>Today, we’re launching research and teaching grants for Tinker access. As part of our commitment to open and collaborative science, we want to make it as easy as possible for students and scholars to use Tinker. If your research or teaching involves training open-weight LLMs, we encourage you to apply.</p>
<p>We’re offering two types of grants to support your work:</p>
<ul>
<li>
<p><strong>Teaching Grants:</strong> We provide $250 in free credits per student for academic classes using Tinker, whether you’re integrating it into an assignment or enabling students to use Tinker for self-directed projects, this is sized to support your entire class for the duration of the course.</p>
</li>
<li>
<p><strong>Research Grants:</strong> We provide grants starting at $5,000 to support research projects and open-source software that uses Tinker.</p>
</li>
</ul>
<p>A selection of early grants we have awarded:</p>
<ul>
<li>
<p>Diyi Yang’s <a href="https://web.stanford.edu/class/cs329x/">Stanford class on Human-Centered LLMs</a> uses Tinker to compare different approaches for training personalized LLMs that capture unique writing styles and align with user habits.</p>
</li>
<li>
<p>Aviral Kumar and Katerina Fragkiadaki’s <a href="https://cmudeeprl.github.io/703website_f25/">CMU class on Deep RL</a> will use Tinker to enable class projects to experiment with state-of-the-art methods for training LLM and VLM based policies via RL.</p>
</li>
<li>
<p><a href="https://chemistry.stanford.edu/people/grant-m-rotskoff">Grant Rotskoff</a>’s lab at Stanford is fine-tuning small-molecule chemistry models with Tinker to help solve problems in computational chemistry.</p>
</li>
</ul>
<p><strong>Instructors</strong>, please apply for teaching grants <a href="https://form.typeform.com/to/JgPkuMvB"><strong>here.</strong></a></p>
<p><strong>Researchers</strong>, please apply for research grants <a href="https://form.typeform.com/to/E9wVFZJJ"><strong>here.</strong></a></p>
<p>We’re assessing applications on a rolling basis and will aim to respond within a week of your application.</p>
</article>
<div class="paginator">
<a id="post-prev-link" class="link" href="https://thinkingmachines.ai/news/announcing-tinker/" title="Announcing Tinker">prev</a>
<a href="https://thinkingmachines.ai/news/tinker-research-and-teaching-grants/#" id="back-to-top" class="link" title="Go to top">back to top</a>
<a id="post-next-link" class="link" href="https://thinkingmachines.ai/news/call-for-community-projects/" title="Tinker: Call for Community Projects">next</a>
</div></description>
<link>https://thinkingmachines.ai/news/tinker-research-and-teaching-grants/</link>
<guid isPermaLink="false">https://thinkingmachines.ai/news/tinker-research-and-teaching-grants/</guid>
<pubDate>Tue, 28 Oct 2025 16:00:00 GMT</pubDate>
</item>
<item>
<title>Announcing Tinker</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Announcing Tinker</h1>
<div class="publish-metadata date-only">
<span>
Oct 1, 2025
</span>
</div>
</div>
<div class="post-cover">
<img data-zoomable="" src="https://thinkingmachines.ai/news/announcing-tinker/svgs/tinker-cover.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
</div>
<article class="content ">
<p class="image-caption" style="text-align: center; margin-top: -1rem; margin-bottom: 2rem; font-size: 0.9rem; color: var(--fg-muted, #666);">
<a href="https://www.computerhistory.org/collections/catalog/X39.81/" target="_blank" rel="noopener">TinkerToy Computer</a> invented by <a href="https://en.wikipedia.org/wiki/Danny_Hillis" target="_blank" rel="noopener">Daniel Hillis</a> and <a href="https://en.wikipedia.org/wiki/Brian_Silverman" target="_blank" rel="noopener">Brian Silverman</a>
</p>
<p>Today, we are launching <a href="https://thinkingmachines.ai/tinker">Tinker</a>, a flexible API for fine-tuning language models. It empowers researchers and hackers to experiment with models by giving them control over the algorithms and data while we handle the complexity of distributed training. Tinker advances our mission of enabling more people to do research on cutting-edge models and customize them to their needs.</p>
<p>Tinker lets you fine-tune a range of large and small open-weight models, including large mixture-of-experts models such as Qwen-235B-A22B. Switching from a small model to a large one is as simple as changing a single string in your Python code.</p>
<p>Tinker is a managed service that runs on our internal clusters and training infrastructure. We handle scheduling, resource allocation, and failure recovery. This allows you to get small or large runs started immediately, without worrying about managing infrastructure. We use LoRA so that we can share the same pool of compute between multiple training runs, lowering costs.</p>
<p>Tinker’s API gives you low-level primitives like <code>forward_backward</code> and <code>sample</code>, which can be used to express most common post-training methods. Even so, achieving good results requires getting many details right. That’s why we’re releasing an open-source library, the <a href="http://github.com/thinking-machines-lab/tinker-cookbook">Tinker Cookbook</a>, with modern implementations of post-training methods that run on top of the Tinker API.</p>
<p>Groups at Princeton, Stanford, Berkeley, and Redwood Research have already been using Tinker:</p>
<ul>
<li>The <a href="https://blog.goedel-prover.com/">Princeton Goedel Team</a> trained mathematical theorem provers</li>
<li>The <a href="https://statmech.stanford.edu/">Rotskoff Chemistry group</a> at Stanford fine-tuned a model to complete chemistry reasoning tasks</li>
<li><a href="https://sky.cs.berkeley.edu/project/skyrl/">Berkeley’s SkyRL group</a> ran experiments on a custom async off-policy RL training loop with multi-agents and multi-turn tool-use.</li>
<li><a href="https://www.redwoodresearch.org/">Redwood Re |
Contributor
|
Successfully generated as following: http://localhost:1200/thinkingmachines/news - Success ✔️<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:atom="http://www.w3.org/2005/Atom" version="2.0">
<channel>
<title>Thinking Machines Lab - News</title>
<link>https://thinkingmachines.ai/news/</link>
<atom:link href="http://localhost:1200/thinkingmachines/news" rel="self" type="application/rss+xml"></atom:link>
<description>Thinking Machines Lab - News - Powered by RSSHub</description>
<generator>RSSHub</generator>
<webMaster>contact@rsshub.app (RSSHub)</webMaster>
<language>en</language>
<lastBuildDate>Fri, 03 Apr 2026 01:28:04 GMT</lastBuildDate>
<ttl>5</ttl>
<item>
<title>Training LLMs to Predict World Events (Guest Post with Mantic)</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Training LLMs to Predict World Events (Guest Post with Mantic)</h1>
<div class="publish-metadata">
<span class="author">
<a href="https://enjeeneer.io/" rel="me noopener" target="_blank">Scott Jeen</a> and <a href="https://www.linkedin.com/in/matthew-aitchison-16aa8799/" rel="me noopener" target="_blank">Matthew Aitchison</a> in collaboration with others at <a href="https://www.mantic.com/" rel="me noopener" target="_blank">Mantic</a>
</span>
<span>
Mar 19, 2026
</span>
</div>
</div>
<div class="post-cover">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/cover.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
</div>
<article class="content ">
<p><em><a href="https://www.mantic.com/">Mantic</a> have been using Tinker since it launched. This guest post is a technical deep dive on what they have built so far.</em></p>
<p>The top AI forecasting systems are approaching superforecaster-level accuracy on geopolitics and current affairs.<label for="sn1" class="sidenote-number"></label><input type="checkbox" id="sn1" class="margin-toggle"><span class="sidenote"><a href="https://news.polymarket.com/p/its-over">It’s over</a> (Polymarket, 2026)</span> This is exciting because scalable, automated forecasting could significantly improve the quality of decision-making across the economy and in government.</p>
<p>To date, the most successful recipe in forecasting tournaments has been to combine an off-the-shelf LLM (like Gemini 3 or GPT-5) with forecasting-specific context-gathering. These models, to our knowledge, have not been explicitly trained for forecasting. Can we improve the recipe by replacing them with models fine-tuned specifically for forecasting?</p>
<p>We target “judgmental forecasting”: prediction problems that require human-like research and reasoning. Judgmental forecasting is needed for domains like geopolitics, politics, technology, business, and economic policy, where there often isn’t enough data for a standard statistical approach like time-series extrapolation. It was popularized by the book <a href="https://books.google.com/books?hl=en&amp;lr=&amp;id=_lMPDAAAQBAJ&amp;oi=fnd&amp;pg=PA1&amp;ots=5avwaBT55R&amp;sig=B2i7FYispHz5PtB6KVu2Ndt58cE">Superforecasting</a>, and now prediction markets like Polymarket and Kalshi.</p>
<p>In this post, we show it’s possible to significantly improve the forecasting performance of gpt-oss-120b using reinforcement learning. With Tinker, we fine-tune a model on around 10,000 binary questions of the form <em>“Will [event] occur before [date]?”</em>. We reward the model for putting greater probability on the correct real-world outcome.</p>
<p>In a head-to-head contest, the fine-tuned model achieves marginally superior performance to the frontier LLMs (see Figure 1), despite much lower initial performance. We find that providing forecast-specific context increases the gains from fine-tuning.</p>
<p>In the optimal <em>ensemble</em> of different models (which outperforms any single model), Grok 4 and our fine-tuned model are the most important contributors. The fine-tuned model learns a forecasting policy that is as accurate as the frontier LLMs, yet decorrelated from them.</p>
<p>Together, the results demonstrate that on-task training can extend the state-of-the-art in AI forecasting.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_1.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>RL fine-tuning makes gpt-oss-120b competitive with the frontier LLMs on questions from the Metaculus AI Benchmark Q2 2025. Naively predicting 50% on every question would get a score of 0, and perfect foresight would get a score of 100, per the construction of the Metaculus “baseline score”. Naively predicting 18.8% on every question (the rate at which the equivalent questions resolved “yes” in the previous tournament, Q1 2025) yields a score of 22.3, which we use to truncate the Y-axis. Fine-tuning improves gpt-oss-120b‘s score from 38.6 to 45.8, on par with the best general models.</figcaption>
</figure>
<h2 id="the-best-existing-recipe-uses-off-the-shelf-llms">The best existing recipe uses off-the-shelf LLMs</h2>
<p>The past two years have seen considerable progress in AI judgmental forecasting capabilities. In the Metaculus Cup, a major tournament for amateur and professional forecasters, the best AI systems now rival the top humans (Figure 2).</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_2.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Human and AI scores in the Metaculus Cup, a premier forecasting competition. Scores from the top 5 AI forecasters have been steadily improving since they first entered in the Summer of 2024. Mantic first entered in the Summer of 2025 and then in Fall 2025 beat the community prediction and the majority of professional forecasters. These results were without fine-tuning.</figcaption>
</figure>
<p>The trend has been driven by more capable off-the-shelf LLMs, and accelerated by more sophisticated forecasting architectures. Our architecture – which has performed well in recent Metaculus tournaments – consists of two standard phases: (1) a research phase, and (2) a prediction phase (Figure 3).<label for="sn3" class="sidenote-number"></label><input type="checkbox" id="sn3" class="margin-toggle"><span class="sidenote">This two-phase process appears in early work on AI forecasting. See: <a href="https://proceedings.neurips.cc/paper_files/paper/2024/hash/5a5acfd0876c940d81619c1dc60e7748-Abstract-Conference.html">Approaching human-level forecasting with language models</a> (Halawi et al., 2024); <a href="https://arxiv.org/abs/2206.15474">Forecasting Future World Events with Neural Networks</a> (Zou et al, 2022).</span></p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_3.png" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Mantic’s architecture. The research phase takes the forecasting question as input and performs deep research to collect information relevant to the question which goes into the prompt for the prediction LLM. The prediction LLM outputs chain-of-thought reasoning and specifies a probability distribution using specialized tools.</figcaption>
</figure>
<p>The research phase is conducted by deep research agents that collect the context needed to make a good prediction. For example, for the question “Will the United States attack Venezuela before 2026?”, search agents will find information about military buildup in the Caribbean, statements from President Trump, the health of the Venezuelan economy, and so on. The collected research is summarized into a prompt for the prediction phase.</p>
<p>The model’s task at the prediction phase is to use our specialized tools to output a probability distribution. In this post, we consider a canonical type of forecasting question: “Will [event] occur before [date]?”. We instruct the LLM to parameterize a mixture model for when the event will next occur – illustrated in Figure 4. The mixture model defines a cumulative distribution function, and from that we can read off the probability of the event occurring before the date specified in the original question.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_4.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Illustrative mixture model. The LLM selects: the number of components in the mixture, their parameters, and their respective weights. The LLM is prompted to select components capturing different scenarios that could lead to the event occurring. The final prediction is a weighted combination of the components.</figcaption>
</figure>
<p>In past tournaments, we’ve used off-the-shelf models as the prediction LLM. Existing literature has shown promising results from RL fine-tuning small models using a simple architecture.<label for="sn4" class="sidenote-number"></label><input type="checkbox" id="sn4" class="margin-toggle"><span class="sidenote"><a href="https://scholar.google.com/citations?view_op=view_citation&amp;hl=en&amp;user=oQuvQ0sAAAAJ&amp;citation_for_view=oQuvQ0sAAAAJ:u5HHmVD_uO8C">Outcome-based Reinforcement Learning to Predict the Future</a> (Turtel et al, 2025)</span> Can we improve frontier AI forecasting systems through on-task fine-tuning?</p>
<h2 id="training-details">Training details</h2>
<h3 id="datasets">Datasets</h3>
<p>We train the prediction LLM on ~10k questions about whether an event will happen by a given date. The questions are from August 2024 to December 2025 and the model’s knowledge cutoff is prior to that, so the resolution is known to us but not to the model. We generated the training set using an LLM pipeline similar to existing work.<label for="sn5" class="sidenote-number"></label><input type="checkbox" id="sn5" class="margin-toggle"><span class="sidenote"><a href="https://arxiv.org/abs/2601.22444">Automating Forecasting Question Generation and Resolution for AI Evaluation</a> (Bosse et al, 2026); <a href="https://arxiv.org/abs/2601.06336">Future-as-Label: Scalable Supervision from Real-World Outcomes</a> (Turtel et al, 2026)</span> Before training, we run the research phase for each question and store static prompts for the prediction LLM.</p>
<p>We test on unseen questions from the Q2 2025 Metaculus AI Benchmark Tournament.<label for="sn6" class="sidenote-number"></label><input type="checkbox" id="sn6" class="margin-toggle"><span class="sidenote"> The Fall 2025 iteration would have been a more obvious choice (for being more recent) but contains lower quality questions, indicated by less performance differentiated between strong and weak forecasters.</span> We compare models whose knowledge cutoff is before this tournament’s start date. The full list of questions can be accessed on <a href="https://gist.github.com/enjeeneer/86e24a52e6041a3d78e333bcab16984d">Github</a>.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_5.png" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Three example binary event questions from the Metaculus Q2 2025 AI Benchmark.</figcaption>
</figure>
<p>We evaluate using the baseline score, following the Metaculus platform. This is log scoring, i.e. ln(probability assigned to true outcome), rescaled such that 100 is the maximum possible score and 0 is the score for a uniform prediction (in our setting, 50%).</p>
<h3 id="implementation">Implementation</h3>
<p>We run the experiments on <a href="https://thinkingmachines.ai/tinker/">Tinker</a>. Of the models available through the API, we choose to train <a href="https://huggingface.co/openai/gpt-oss-120b">gpt-oss-120b</a> because of its strong initial performance — second only to <a href="https://huggingface.co/moonshotai/Kimi-K2.5">Kimi K2.5</a> — while being cheaper and faster.</p>
<p>We use a standard policy gradient algorithm with <a href="https://arxiv.org/abs/2402.03300">GRPO-style advantage normalisation</a> and <a href="https://arxiv.org/abs/2503.20783">no division by the standard deviation</a>. For rewards we use the Brier score which is <a href="https://sites.stat.washington.edu/raftery/Research/PDF/Gneiting2007jasa.pdf">strictly proper</a>. We found that the Brier score leads to more stable training than the log score, even though the log score is also strictly proper. This could be because the Brier score is bounded in [0, 1] and so produces lower-variance policy gradient estimates.</p>
<p>Open-source RL packages often use <a href="https://arxiv.org/abs/2309.06180">vLLM</a> for the sampling backend and <a href="https://arxiv.org/abs/2304.11277">FSDP</a> for the training backend. These can disagree on token probabilities produced by identical policies, which biases policy gradient estimates and destabilises training. We found these discrepancies to be lower on Tinker’s integrated infrastructure, but further mitigate them with <a href="https://fengyao.notion.site/off-policy-rl">an importance sampling correction on the advantages</a>.</p>
<p>The Brier score reward function is strictly monotonic in the predicted probability for a fixed outcome, so different rollouts almost always produce different rewards. This makes within-group reward ties extremely unlikely and lets us train with a relatively small group size (8) without needing to break ties or induce variance. We use a batch size of 64, as we find that <a href="https://thinkingmachines.ai/blog/lora/">larger batch sizes tend to destabilise training</a>.</p>
<h2 id="results">Results</h2>
<h3 id="fine-tuning-elevates-gpt-oss-120b-to-frontier-llm-performance">Fine-tuning elevates gpt-oss-120b to frontier LLM performance</h3>
<p>The model’s test set score improves through training (Figure 6), moving from an initial score of 38.6 mean baseline points per question (below all frontier models) to a final score of 45.8 mean baseline points per question (marginally above). This demonstrates that on-task forecasting training can provide a large performance uplift.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_6.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Test set baseline score of gpt-oss-120b with and without Mantic research and tools. In the model-only setup, test set performance improves, but never reaches the initial score of gpt-oss-120b with Mantic research and tools. With Mantic research and tools, gpt-oss-120b climbs 7 points through training and marginally exceeds the performance of Gemini 3 Pro. Training continued for further steps but performance no longer improved.</figcaption>
</figure>
<p>The LLM fine-tuned without the benefit of our pre-generated research phase, and without the tools to construct mixture models, gains only 3 points from training instead of 7. This suggests these prerequisites positively influence the optimization dynamics, in addition to improving initialization.</p>
<h3 id="the-fine-tuned-model-is-an-important-member-of-the-optimal-ensemble">The fine-tuned model is an important member of the optimal ensemble</h3>
<p>In human forecasting, there is a well-known “wisdom of the crowd” effect: aggregate forecasts from multiple people often outperforms any one individual. This effect, in part, explains the <a href="https://polymarket.com/accuracy">impressive accuracy of prediction markets</a>. Can we get the same benefit from ensembling the predictions of different LLMs?<label for="sn7" class="sidenote-number"></label><input type="checkbox" id="sn1" class="margin-toggle"><span class="sidenote">See also <a href="https://arxiv.org/abs/2402.19379">Wisdom of the Silicon Crowd: LLM Ensemble Prediction Capabilities Rival Human Crowd Accuracy</a> (Schoenegger et al, 2024)</span></p>
<p>To contribute to an ensemble, an LLM must be sufficiently accurate on its own and also decorrelated from other LLMs in the ensemble. Predictions from most frontier LLMs, while accurate, contribute little diversity to the top-performing model (in our case, Gemini 3 Pro) – Figure 7. Among the frontier LLMs, Grok 4 is the exception: its predictions score well whilst correlating less with other frontier LLMs.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_7.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Mean baseline score on binary event questions from the Metaculus Q2 AIB Benchmark with respect to their Jensson Shannon divergence from Gemini 3 Pro for a suite of closed source and open source LLMs. Marker colour indicates each model’s weight in the optimal 5-sample ensemble. The optimal ensemble consists of fine-tuned gpt-oss-120b (40%), gemini 3 pro (20%), gpt-5 (20%) and grok 4 (20%).</figcaption>
</figure>
<p>The optimal ensemble, on a budget of 5 samples, is 2 samples from our fine-tuned gpt-oss-120b plus 1 sample from each of Gemini 3 Pro, Grok 4, and GPT-5. We can test each model’s contribution by removing it, recomputing the optimal ensemble, and seeing how much the score degrades. We find that Grok 4 is the least replaceable, with fine-tuned gpt-oss-120b in second place. Other models can be replaced with little to no performance degradation (Figure 8).</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_8.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>When selecting an ensemble of frontier and open-source models, Grok 4 and fine-tuned gpt-oss-120b are the least replaceable. Model replaceability is defined as the reduction in score incurred when removing a model from the optimal ensemble. By definition, if a model is not included in the optimal ensemble,there is no cost to removing it.</figcaption>
</figure>
<p>GPT-5 and Gemini 3 Pro make similar predictions, and thus don’t benefit much from ensembling with each other. Both models improve from mixing their predictions with either Grok 4 or the fine-tuned gpt-oss-120b, and Grok 4 also benefits most from mixing with the fine-tuned model. In either optimal 3-way ensemble from Figure 9, the fine-tuned model gets about half of the total weight.</p>
<figure style="margin-top: 2rem; margin-bottom: 2.3rem;">
<img data-zoomable="" src="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/svgs/figure_9.png" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
<figcaption>Three-way ensembles. Left: The optimal three-way ensemble between the fine-tuned gpt-oss-120b, Gemini 3 Pro and GPT-5 is weighted 56%, 26% and 18% respectively, depicted by the black star. Right: The optimal three-way ensemble between the fine-tuned gpt-oss-120b, Gemini 3 Pro and Grok 4 is weighted 48, 26% and 26% respectively, depicted by the white star.</figcaption>
</figure>
<h2 id="conclusions-and-next-steps">Conclusions and next steps</h2>
<p>We have shown that we can elevate the forecasting performance of gpt-oss-120b to match frontier LLMs with RL fine-tuning. This work can be extended in many ways, some of which we have already begun exploring:</p>
<ol>
<li><strong>Training larger models.</strong> Tinker enables training larger models with higher initial performance than gpt-oss, specifically Kimi K2.5.</li>
<li><strong>Training on all question formats.</strong> We are already training models on numerical questions such as economic indicators and multiple choice questions such as election results.</li>
<li><strong>Improved question sets.</strong> As the models become stronger forecasters, we need more challenging forecasting questions.</li>
<li><strong>Information retrieval inside the loop.</strong> We could give the prediction LLM tools for information retrieval and include this in the training loop.</li>
</ol>
<h2 id="citation">Citation</h2>
<p>Please cite this work as:</p>
<pre tabindex="0"><code>Jeen, Scott; Aitchison, Matthew; and Mantic, "Training LLMs to Predict World Events",
Thinking Machines Lab: News, Mar 2026.
</code></pre><p>Or use the BibTeX citation:</p>
<pre tabindex="0"><code>@article{scott2026forecasting,
author = {Scott Jeen, Matthew Aitchison, and Mantic},
title = {Training LLMs to Predict World Events},
journal = {Thinking Machines Lab: News},
year = {2026},
note = {https://thinkingmachines.ai/news/training-llms-to-predict-world-events/}
}
</code></pre>
</article>
<div class="paginator">
<a id="post-prev-link" class="link" href="https://thinkingmachines.ai/news/nvidia-partnership/" title="Thinking Machines Lab and NVIDIA Announce Long-Term Gigawatt-Scale Strategic Partnership">prev</a>
<a href="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/#" id="back-to-top" class="link" title="Go to top">back to top</a>
<a id="post-next-link"></a>
</div></description>
<link>https://thinkingmachines.ai/news/training-llms-to-predict-world-events/</link>
<guid isPermaLink="false">https://thinkingmachines.ai/news/training-llms-to-predict-world-events/</guid>
<pubDate>Wed, 18 Mar 2026 16:00:00 GMT</pubDate>
</item>
<item>
<title>Thinking Machines Lab and NVIDIA Announce Long-Term Gigawatt-Scale Strategic Partnership</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Thinking Machines Lab and NVIDIA Announce Long-Term Gigawatt-Scale Strategic Partnership</h1>
<div class="publish-metadata date-only">
<span>
Mar 10, 2026
</span>
</div>
</div>
<div class="post-cover compact">
<img data-zoomable="" src="https://thinkingmachines.ai/news/nvidia-partnership/images/jensen-huang-mira-murati-nvidia-debs-3127-2.jpg" alt="Cover image for Thinking Machines Lab and NVIDIA Announce Long-Term Gigawatt-Scale Strategic Partnership" class="cover-image" fetchpriority="high" decoding="async" loading="eager" referrerpolicy="no-referrer">
<p class="post-cover-caption">Jensen Huang (NVIDIA) and Mira Murati (Thinking Machines)</p>
</div>
<article class="content ">
<p>Thinking Machines Lab and NVIDIA announced today a multi-year strategic partnership to deploy at least one gigawatt of next-generation NVIDIA Vera Rubin systems to support Thinking Machines’ frontier model training and platforms delivering customizable AI at scale. Deployment on NVIDIA’s Vera Rubin platform is targeted for early next year.</p>
<p>The partnership also includes an effort to design training and serving systems for NVIDIA architectures and broaden access to frontier AI and open models for enterprises, research institutions, and the scientific community.</p>
<p>NVIDIA has also made a significant investment in Thinking Machines Lab to support the company’s long-term growth.</p>
<p>“AI is the most powerful knowledge discovery instrument in human history,” said Jensen Huang, founder and CEO of NVIDIA. “Thinking Machines has brought together a world-class team to advance the frontier of AI. We are thrilled to partner with Thinking Machines to realize their exciting vision for the future of AI.”</p>
<p>“NVIDIA’s technology is the foundation on which the entire field is built,” said Mira Murati, cofounder and CEO of Thinking Machines. “This partnership accelerates our capacity to build AI that people can shape and make their own, as it shapes human potential in turn.”</p>
<p>Building powerful AI systems that are understandable, customizable, and collaborative demands advances in research, design, and infrastructure at scale. This partnership provides that foundation, with the shared aim of ensuring that the most transformative technology of our time expands human capability.</p>
<img data-zoomable="" src="https://thinkingmachines.ai/news/nvidia-partnership/cover.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
</article>
<div class="paginator">
<a id="post-prev-link" class="link" href="https://thinkingmachines.ai/news/tinker-general-availability/" title="Tinker: General Availability and Vision Input">prev</a>
<a href="https://thinkingmachines.ai/news/nvidia-partnership/#" id="back-to-top" class="link" title="Go to top">back to top</a>
<a id="post-next-link" class="link" href="https://thinkingmachines.ai/news/training-llms-to-predict-world-events/" title="Training LLMs to Predict World Events (Guest Post with Mantic)">next</a>
</div></description>
<link>https://thinkingmachines.ai/news/nvidia-partnership/</link>
<guid isPermaLink="false">https://thinkingmachines.ai/news/nvidia-partnership/</guid>
<pubDate>Mon, 09 Mar 2026 16:00:00 GMT</pubDate>
</item>
<item>
<title>Tinker: General Availability and Vision Input</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Tinker: General Availability and Vision Input</h1>
<div class="publish-metadata date-only">
<span>
Dec 12, 2025
</span>
</div>
</div>
<article class="content ">
<p>Today we are announcing four updates to Tinker:</p>
<ul>
<li>No more waitlist</li>
<li>New reasoning model: Kimi K2 Thinking</li>
<li>New inference interface that is compatible with the OpenAI API</li>
<li>Vision input support with Qwen3-VL</li>
</ul>
<h2 id="general-availability">General availability</h2>
<p>The waitlist is over! Everybody can use Tinker now; <a href="https://auth.thinkingmachines.ai/sign-up">sign up here</a> to get started. See the <a href="https://thinkingmachines.ai/tinker/">Tinker homepage</a> for available models and pricing, and check out the <a href="https://github.com/thinking-machines-lab/tinker-cookbook">Tinker cookbook</a> for code examples.</p>
<h2 id="more-reasoning-with-kimi-k2-thinking">More reasoning with Kimi K2 Thinking</h2>
<p>Users can now fine-tune Kimi K2 Thinking on Tinker. With a trillion parameters, Kimi K2 is the largest model in our lineup so far. It is built for long chains of reasoning and tool use.</p>
<h2 id="openai-api-compatible-sampling">OpenAI API-compatible sampling</h2>
<p>Tinker has a standard function for inference:</p>
<div class="highlight"><pre tabindex="0" class="chroma"><code class="language-python" data-lang="python"><span class="line"><span class="cl"><span class="n">prompt</span> <span class="o">=</span> <span class="n">types</span><span class="o">.</span><span class="n">ModelInput</span><span class="o">.</span><span class="n">from_ints</span><span class="p">(</span><span class="n">tokenizer</span><span class="o">.</span><span class="n">encode</span><span class="p">(</span><span class="s2">"The capital of France is"</span><span class="p">,))</span>
</span></span><span class="line"><span class="cl"><span class="n">params</span> <span class="o">=</span> <span class="n">types</span><span class="o">.</span><span class="n">SamplingParams</span><span class="p">(</span><span class="n">max_tokens</span><span class="o">=</span><span class="mi">20</span><span class="p">,</span> <span class="n">temperature</span><span class="o">=</span><span class="mf">0.0</span><span class="p">,</span> <span class="n">stop</span><span class="o">=</span><span class="p">[</span><span class="s2">"</span><span class="se">\n</span><span class="s2">"</span><span class="p">])</span>
</span></span><span class="line"><span class="cl"><span class="n">future</span> <span class="o">=</span> <span class="n">sampling_client</span><span class="o">.</span><span class="n">sample</span><span class="p">(</span><span class="n">prompt</span><span class="o">=</span><span class="n">prompt</span><span class="p">,</span> <span class="n">sampling_params</span><span class="o">=</span><span class="n">params</span><span class="p">)</span>
</span></span></code></pre></div><p>With this release, we have added OpenAI API-compatible scaffolding for quickly sampling from a model by specifying a path, even while it’s still training. This also means Tinker can now plug-and-play with any OpenAI API-compatible platform. See more information in our <a href="https://tinker-docs.thinkingmachines.ai/compatible-apis/openai">Tinker documentation</a>.</p>
<div class="highlight"><pre tabindex="0" class="chroma"><code class="language-python" data-lang="python"><span class="line"><span class="cl"><span class="n">response</span> <span class="o">=</span> <span class="n">openai_client</span><span class="o">.</span><span class="n">completions</span><span class="o">.</span><span class="n">create</span><span class="p">(</span>
</span></span><span class="line"><span class="cl"> <span class="n">model</span><span class="o">=</span><span class="s2">"tinker://0034d8c9-0a88-52a9-b2b7-bce7cb1e6fef:train:0/sampler_weights/000080"</span><span class="p">,</span>
</span></span><span class="line"><span class="cl"> <span class="n">prompt</span><span class="o">=</span><span class="s2">"The capital of France is"</span><span class="p">,</span>
</span></span><span class="line"><span class="cl"> <span class="n">max_tokens</span><span class="o">=</span><span class="mi">20</span><span class="p">,</span>
</span></span><span class="line"><span class="cl"> <span class="n">temperature</span><span class="o">=</span><span class="mf">0.0</span><span class="p">,</span>
</span></span><span class="line"><span class="cl"> <span class="n">stop</span><span class="o">=</span><span class="p">[</span><span class="s2">"</span><span class="se">\n</span><span class="s2">"</span><span class="p">],</span>
</span></span><span class="line"><span class="cl"><span class="p">)</span>
</span></span></code></pre></div><h2 id="vision-input-with-qwen3-vl">Vision input with Qwen3-VL</h2>
<p>We’ve added two vision models to Tinker: Qwen3-VL-30B-A3B-Instruct and Qwen3-VL-235B-A22B-Instruct. With these, users can process pictures, screenshots, and diagrams for a variety of applications.</p>
<p>To input images, just interleave together an <a href="https://tinker-docs.thinkingmachines.ai/api-reference/types#imagechunk-objects">ImageChunk</a> – consisting of your image, saved as bytes – with text chunks. For example:</p>
<div class="highlight"><pre tabindex="0" class="chroma"><code class="language-python" data-lang="python"><span class="line"><span class="cl"><span class="n">model_input</span> <span class="o">=</span> <span class="n">tinker</span><span class="o">.</span><span class="n">ModelInput</span><span class="p">(</span><span class="n">chunks</span><span class="o">=</span><span class="p">[</span>
</span></span><span class="line"><span class="cl"> <span class="n">tinker</span><span class="o">.</span><span class="n">types</span><span class="o">.</span><span class="n">ImageChunk</span><span class="p">(</span><span class="n">data</span><span class="o">=</span><span class="n">image_data</span><span class="p">,</span> <span class="nb">format</span><span class="o">=</span><span class="s2">"png"</span><span class="p">),</span>
</span></span><span class="line"><span class="cl"> <span class="n">tinker</span><span class="o">.</span><span class="n">types</span><span class="o">.</span><span class="n">EncodedTextChunk</span><span class="p">(</span><span class="n">tokens</span><span class="o">=</span><span class="n">tokenizer</span><span class="o">.</span><span class="n">encode</span><span class="p">(</span><span class="s2">"What is this?"</span><span class="p">)),</span>
</span></span><span class="line"><span class="cl"><span class="p">])</span>
</span></span></code></pre></div><p>These vision inputs can be used in a variety of applications out-of-the-box, including SFT and RL finetuning.</p>
<p>To demonstrate vision understanding in action, we are sharing <a href="https://github.com/thinking-machines-lab/tinker-cookbook/tree/main/tinker_cookbook/recipes/vlm_classifier">a new cookbook recipe for fine-tuning VLMs as image classifiers</a>. Qwen3-VL-235B-A22B-Instruct obtains reasonable accuracy even given just one example per class; performance improves with more labeled data.</p>
<h2 id="training-image-classifiers-with-tinker">Training image classifiers with Tinker</h2>
<p>To showcase Tinker’s new vision capabilities, we finetuned <a href="https://huggingface.co/Qwen/Qwen3-VL-235B-A22B-Instruct">Qwen3-VL-235B-A22B-Instruct</a> to classify images on four classic datasets:</p>
<ul>
<li><a href="https://data.caltech.edu/records/mzrjq-6wc02">Caltech 101</a>, a dataset of 101 general object categories.</li>
<li><a href="https://www.kaggle.com/datasets/eduardo4jesus/stanford-cars-dataset">Stanford Cars</a>, a dataset of car makes, models, and years.</li>
<li><a href="https://www.robots.ox.ac.uk/~vgg/data/flowers/102/">Oxford Flowers</a>, a dataset of flower species.</li>
<li><a href="https://www.robots.ox.ac.uk/~vgg/data/pets/">Oxford Pets</a>, a dataset of pet breeds.</li>
</ul>
<p>Since Qwen3-VL is a language model, we frame classification as text generation: given an image, the model outputs the class name. We compare this approach against a traditional vision baseline of finetuning a vision-only model — DINOv2-base. <a href="https://arxiv.org/pdf/2304.07193">DINOv2</a> is a self-supervised vision transformer that was trained to encode images, and is commonly used as a backbone for pure computer vision tasks. For DINOv2, we add a classification head that predicts a distribution over all N classes. Both models are fine-tuned with LoRA.</p>
<p>Labeled image data is scarce for many real-world use cases, so data efficiency is the primary measure we look at. We show the classification accuracy when sweeping across the number of labeled examples per class, starting with just a single one.</p>
<figure id="fig:qwen-v-dino">
<img data-zoomable="" src="https://thinkingmachines.ai/news/tinker-general-availability/images/vlm-graphs.png" referrerpolicy="no-referrer">
<figcaption>Comparison of fine-tuned Qwen3-VL-235-A22B and DINOv2 performance on simple image classification tasks.</figcaption>
</figure>
<p>In the limited-data regime, Qwen3-VL-235-A22B outperforms DINOv2. Not only is it a bigger model, but as a VLM, it also comes with language knowledge out-of-the-box (i.e. what a “golden retriever” or “sunflower” is). This general language-and-vision capability of Qwen3-VL makes it readily available for vision tasks beyond classification.</p>
<h2 id="happy-holidays">Happy Holidays</h2>
<p>Tinker exists to enable builders and researchers to train and customize state-of-the-art models. As always, we look forward to seeing what you build with Tinker. Happy holidays!</p>
</article>
<div class="paginator">
<a id="post-prev-link" class="link" href="https://thinkingmachines.ai/news/call-for-community-projects/" title="Tinker: Call for Community Projects">prev</a>
<a href="https://thinkingmachines.ai/news/tinker-general-availability/#" id="back-to-top" class="link" title="Go to top">back to top</a>
<a id="post-next-link" class="link" href="https://thinkingmachines.ai/news/nvidia-partnership/" title="Thinking Machines Lab and NVIDIA Announce Long-Term Gigawatt-Scale Strategic Partnership">next</a>
</div></description>
<link>https://thinkingmachines.ai/news/tinker-general-availability/</link>
<guid isPermaLink="false">https://thinkingmachines.ai/news/tinker-general-availability/</guid>
<pubDate>Thu, 11 Dec 2025 16:00:00 GMT</pubDate>
</item>
<item>
<title>Tinker: Call for Community Projects</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Tinker: Call for Community Projects</h1>
<div class="publish-metadata date-only">
<span>
Nov 7, 2025
</span>
</div>
</div>
<article class="content ">
<p>We launched <a href="https://thinkingmachines.ai/tinker/">Tinker</a> to enable builders and researchers to train models their own way, whether they’re conducting studies or customizing models for new applications. We plan to publish regular roundups of the coolest projects from the Tinker community, and <strong>we invite you to submit what you’ve been Tinkering on to be featured on our blog</strong>.</p>
<p>Below are some broad suggestions for what we hope to see from the Tinker featured projects, and some specific research directions we would particularly love to see pursued.</p>
<h2 id="guidelines-for-tinker-featured-projects">Guidelines for Tinker Featured Projects</h2>
<p>We’re interested in featuring ML research projects, AI-enabled research in other domains, custom models, and other contributions. Some examples:</p>
<ul>
<li>A reimplementation of a research project or tech report using Tinker, such as papers that compare algorithmic recipes or datasets.</li>
<li>Original research in machine learning, such as exploring new approaches to training or optimization or applying novel benchmarks and evaluations.</li>
<li>Research in a non-AI field that uses fine-tuned models, such as the work on mathematical theorem provers and chemistry models we <a href="https://thinkingmachines.ai/news/announcing-tinker/#:~:text=Groups%20at%20Princeton%2C%20Stanford">highlighted previously</a>.</li>
<li>Product prototypes built with Tinker, demoing a model that does something fresh or delightful.</li>
<li>Novel datasets and task environments for training models.</li>
<li>High-level libraries built on top of Tinker that enable less experienced practitioners to perform fine-tuning effectively.</li>
<li>Infrastructure contributions, such as a clean self-hosted implementation of the Tinker training API.</li>
</ul>
<p>Your submission should include a write-up and, preferably, an open-source release of your code. We encourage you to focus on rigor and clear evaluation in your write-ups: crisp charts, raw output examples, clear comparisons to alternative approaches or models on relevant benchmarks and metrics. Tinkering is experimenting — we want to feature diligent work and transparent results over novelty or hype.</p>
<p>Please send your projects and any related questions to <a href="https://thinkingmachines.ai/cdn-cgi/l/email-protection#56223f383d332416223e3f383d3f38313b37353e3f38332578373f692523343c3335226b10333722232433327364660624393c333522736466"><span class="__cf_email__" data-cfemail="06726f686d637446726e6f686d6f68616b67656e6f68637528676f">[email&nbsp;protected]</span></a> with “Featured Project” in the subject line.</p>
<h2 id="suggested-research-projects">Suggested research projects</h2>
<p>Here are some research directions that we would personally love to see explored and that Tinker can enable real progress on. We have <a href="https://github.com/thinking-machines-lab/tinker-project-ideas">created a repo</a> with detailed motivation and guidelines for each; we’ll be adding more resources to it over time to help researchers get started. We expect most project ideas to surprise us, but this short list could serve as inspiration.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/replicate-cai-with-base-models.md">Replicating Constitutional AI, starting from the base model.</a></strong> Though RLAIF is widely used, it’s most often bootstrapped from existing instruction-tuned models. This makes it difficult to separate the impact of the constitution from the impact of the data-generating model that interprets it. A study of Constitutional AI with and without instruction-tuned models in the pipeline would shed light on the use of constitutions and RLAIF.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/noisy-student.md">RLVR with Noisy student.</a></strong> Noisy student self-distillation was a popular technique in an earlier era of machine learning for making use of large unlabeled datasets, but it hasn’t been adapted widely to LLMs. One possible adaptation is to start RLVR with a small labeled training set and a large unlabeled one, then have the student apply labels to the latter set after each RL run and iterate.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/on-policy-context-distillation.md">On-Policy Context Distillation.</a></strong> Context distillation trains a student model with empty context on a teacher model with long and detailed context. Prior work used off-policy distillation — fine-tuning on teacher samples. We have found that <a href="https://thinkingmachines.ai/blog/on-policy-distillation/">on-policy distillation</a> is often much more effective; it would be useful to compare the two approaches for context distillation in particular.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/memorization-empirical-study.md">RL memory test.</a></strong> Our <a href="https://thinkingmachines.ai/blog/lora/">post on LoRA</a> presented theoretical arguments on the rate of information acquisition by both SFT and RL. You can set up a toy environment where RL must learn a completely random number sequence, to compare the empirical learning rate under various reward functions to the theoretical estimate.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/direct-rl-on-pairwise-judge.md">Direct RL on pairwise judge.</a></strong> RLHF and RLAIF use datasets of pairwise preferences, which are used to train a reward model, which is then used in RL. As an alternative “direct” approach, we can do RL using a prompted model that does pairwise comparisons, without training the reward model. It would be interesting to do experiments comparing the direct and indirect approaches.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/replicate-open-character-training.md">Replicate Open Character Training.</a></strong> Replicate the recent paper on <a href="https://arxiv.org/abs/2511.01689">Open Character Training</a> using Tinker.</p>
<p><strong><a href="https://github.com/thinking-machines-lab/tinker-project-ideas/blob/main/gan-joke-generation.md">GAN for jokes.</a></strong> In domains such as humor, it is easier to curate a human-vetted set of demonstrations than to train a reliable judge or reward model. Try implementing GAN-style training for a joke evaluator and joke generator that can craft a joke with a requested subject and keywords.</p>
<h2 id="tips-for-high-quality-ml-experiments">Tips for high-quality ML experiments</h2>
<p>In closing, we want to offer a few guidelines for running quality ML studies, the same guidelines we strive to adhere to internally when running experiments and documenting the results.</p>
<p>We encourage researchers to apply multiple analyses for examining each result. When creating datasets or environments, we recommend training a range of models and applying different evals. When developing novel methods, we suggest comparing to simpler baseline methods and sweeping hyperparameters that performance is sensitive to, particularly learning rate.</p>
<p>We’d love to see your reasoning in the write-up: assumptions you made, how your approach diverges from previous reports, and what motivated each change. We hope to see examples of the raw data and model rollouts, along with the summarized results. Finally, we appreciate crisp and detailed write-ups with clean and well-labeled charts and illustrations of the inner workings of the methods used.</p>
<p>We are excited to see what our community creates with Tinker, and hope that our featured projects will inspire your own work.</p>
</article>
<div class="paginator">
<a id="post-prev-link" class="link" href="https://thinkingmachines.ai/news/tinker-research-and-teaching-grants/" title="Tinker: Announcing Research and Teaching Grants">prev</a>
<a href="https://thinkingmachines.ai/news/call-for-community-projects/#" id="back-to-top" class="link" title="Go to top">back to top</a>
<a id="post-next-link" class="link" href="https://thinkingmachines.ai/news/tinker-general-availability/" title="Tinker: General Availability and Vision Input">next</a>
</div></description>
<link>https://thinkingmachines.ai/news/call-for-community-projects/</link>
<guid isPermaLink="false">https://thinkingmachines.ai/news/call-for-community-projects/</guid>
<pubDate>Thu, 06 Nov 2025 16:00:00 GMT</pubDate>
</item>
<item>
<title>Tinker: Announcing Research and Teaching Grants</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Tinker: Announcing Research and Teaching Grants</h1>
<div class="publish-metadata date-only">
<span>
Oct 29, 2025
</span>
</div>
</div>
<article class="content ">
<p>We launched <a href="https://thinkingmachines.ai/news/announcing-tinker/">Tinker</a> nearly one month ago. Since then, researchers across academia and non-profits have been using Tinker to train custom models and advance their research.</p>
<p>Today, we’re launching research and teaching grants for Tinker access. As part of our commitment to open and collaborative science, we want to make it as easy as possible for students and scholars to use Tinker. If your research or teaching involves training open-weight LLMs, we encourage you to apply.</p>
<p>We’re offering two types of grants to support your work:</p>
<ul>
<li>
<p><strong>Teaching Grants:</strong> We provide $250 in free credits per student for academic classes using Tinker, whether you’re integrating it into an assignment or enabling students to use Tinker for self-directed projects, this is sized to support your entire class for the duration of the course.</p>
</li>
<li>
<p><strong>Research Grants:</strong> We provide grants starting at $5,000 to support research projects and open-source software that uses Tinker.</p>
</li>
</ul>
<p>A selection of early grants we have awarded:</p>
<ul>
<li>
<p>Diyi Yang’s <a href="https://web.stanford.edu/class/cs329x/">Stanford class on Human-Centered LLMs</a> uses Tinker to compare different approaches for training personalized LLMs that capture unique writing styles and align with user habits.</p>
</li>
<li>
<p>Aviral Kumar and Katerina Fragkiadaki’s <a href="https://cmudeeprl.github.io/703website_f25/">CMU class on Deep RL</a> will use Tinker to enable class projects to experiment with state-of-the-art methods for training LLM and VLM based policies via RL.</p>
</li>
<li>
<p><a href="https://chemistry.stanford.edu/people/grant-m-rotskoff">Grant Rotskoff</a>’s lab at Stanford is fine-tuning small-molecule chemistry models with Tinker to help solve problems in computational chemistry.</p>
</li>
</ul>
<p><strong>Instructors</strong>, please apply for teaching grants <a href="https://form.typeform.com/to/JgPkuMvB"><strong>here.</strong></a></p>
<p><strong>Researchers</strong>, please apply for research grants <a href="https://form.typeform.com/to/E9wVFZJJ"><strong>here.</strong></a></p>
<p>We’re assessing applications on a rolling basis and will aim to respond within a week of your application.</p>
</article>
<div class="paginator">
<a id="post-prev-link" class="link" href="https://thinkingmachines.ai/news/announcing-tinker/" title="Announcing Tinker">prev</a>
<a href="https://thinkingmachines.ai/news/tinker-research-and-teaching-grants/#" id="back-to-top" class="link" title="Go to top">back to top</a>
<a id="post-next-link" class="link" href="https://thinkingmachines.ai/news/call-for-community-projects/" title="Tinker: Call for Community Projects">next</a>
</div></description>
<link>https://thinkingmachines.ai/news/tinker-research-and-teaching-grants/</link>
<guid isPermaLink="false">https://thinkingmachines.ai/news/tinker-research-and-teaching-grants/</guid>
<pubDate>Tue, 28 Oct 2025 16:00:00 GMT</pubDate>
</item>
<item>
<title>Announcing Tinker</title>
<description><div class="post-heading news-post">
<h1 class="post-title">Announcing Tinker</h1>
<div class="publish-metadata date-only">
<span>
Oct 1, 2025
</span>
</div>
</div>
<div class="post-cover">
<img data-zoomable="" src="https://thinkingmachines.ai/news/announcing-tinker/svgs/tinker-cover.svg" style="display: block; width: 100%; max-width: 800px; height: auto; margin: 0 auto;" loading="lazy" alt="" referrerpolicy="no-referrer">
</div>
<article class="content ">
<p class="image-caption" style="text-align: center; margin-top: -1rem; margin-bottom: 2rem; font-size: 0.9rem; color: var(--fg-muted, #666);">
<a href="https://www.computerhistory.org/collections/catalog/X39.81/" target="_blank" rel="noopener">TinkerToy Computer</a> invented by <a href="https://en.wikipedia.org/wiki/Danny_Hillis" target="_blank" rel="noopener">Daniel Hillis</a> and <a href="https://en.wikipedia.org/wiki/Brian_Silverman" target="_blank" rel="noopener">Brian Silverman</a>
</p>
<p>Today, we are launching <a href="https://thinkingmachines.ai/tinker">Tinker</a>, a flexible API for fine-tuning language models. It empowers researchers and hackers to experiment with models by giving them control over the algorithms and data while we handle the complexity of distributed training. Tinker advances our mission of enabling more people to do research on cutting-edge models and customize them to their needs.</p>
<p>Tinker lets you fine-tune a range of large and small open-weight models, including large mixture-of-experts models such as Qwen-235B-A22B. Switching from a small model to a large one is as simple as changing a single string in your Python code.</p>
<p>Tinker is a managed service that runs on our internal clusters and training infrastructure. We handle scheduling, resource allocation, and failure recovery. This allows you to get small or large runs started immediately, without worrying about managing infrastructure. We use LoRA so that we can share the same pool of compute between multiple training runs, lowering costs.</p>
<p>Tinker’s API gives you low-level primitives like <code>forward_backward</code> and <code>sample</code>, which can be used to express most common post-training methods. Even so, achieving good results requires getting many details right. That’s why we’re releasing an open-source library, the <a href="http://github.com/thinking-machines-lab/tinker-cookbook">Tinker Cookbook</a>, with modern implementations of post-training methods that run on top of the Tinker API.</p>
<p>Groups at Princeton, Stanford, Berkeley, and Redwood Research have already been using Tinker:</p>
<ul>
<li>The <a href="https://blog.goedel-prover.com/">Princeton Goedel Team</a> trained mathematical theorem provers</li>
<li>The <a href="https://statmech.stanford.edu/">Rotskoff Chemistry group</a> at Stanford fine-tuned a model to complete chemistry reasoning tasks</li>
<li><a href="https://sky.cs.berkeley.edu/project/skyrl/">Berkeley’s SkyRL group</a> ran experiments on a custom async off-policy RL training loop with multi-agents and multi-turn tool-use.</li>
<li><a href="https://www.redwoodresearch.org/">Redwood Re |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Involved Issue / 该 PR 相关 Issue
N/A - New route
Example for the Proposed Route(s) / 路由地址示例
New RSS Route Checklist / 新 RSS 路由检查表
parseDatewithMMM D, YYYYformat)PuppeteerNote / 说明
Add RSS feed for Thinking Machines Lab news page. Founded by Mira Murati (ex-OpenAI CTO), the lab publishes news about their AI research, products (Tinker), and partnerships.
The route scrapes the news listing page using cheerio, extracting article titles, dates, and links from
<time class="desktop-time">and<div class="post-title">elements. Full article content is fetched and cached for each entry.