Tree of Thoughts Improves AI Reasoning and Logic By Nine Times

Language models are increasingly being deployed for general problem solving across a wide range of tasks, but are still confined to token-level, left-to-right decision-making processes during inference. This means they can fall short in tasks that require exploration, strategic lookahead, or where initial decisions play a pivotal role. Deepmind researchers introduce a new framework for…
Tree of Thoughts Improves AI Reasoning and Logic By Nine Times

Language models are increasingly being deployed for general problem solving across a wide range of tasks, but are still confined to token-level, left-to-right decision-making processes during inference. This means they can fall short in tasks that require exploration, strategic lookahead, or where initial decisions play a pivotal role. Deepmind researchers introduce a new framework for language model inference, “Tree of Thoughts” (ToT), which generalizes over the popular “Chain of Thought” approach to prompting language models, and enables exploration over coherent units of text (“thoughts”) that serve as intermediate step toward problem solving.

Tree of thought allows multiple step analysis like chain of thought and allows multiple comparisons of different multiple step analysis. Tree of thought allows increased options after each step and allows the system to restart the at the first or earlier steps to look again for new options. It then finds the best option after multiple searchs for best of different analytical options.

Tree of Thoughts (ToT) allows LMs (Language Models) to perform deliberate decision making by considering multiple different reasoning paths and self-evaluating choices to decide the next course of action, as well as looking ahead or backtracking when necessary to make global choices. The experiments show that ToT significantly enhances language models’ problem-solving abilities on three novel tasks requiring non-trivial planning or search: Game of 24, Creative Writing, and Mini Crosswords.

For instance, in Game of 24, while GPT-4 with chain-of-thought prompting only solved 4% of tasks, our method achieved a success rate of 74%.

Tree of Thought is an improvement of basic input and output, chain of thought and self-consistency with chain of thought.

The Tree-of-Thought approach extends existing planning formulations by considering multiple potentially feasible plans simultaneously at each problem-solving step, and proceeding with the most promising ones. The integration between thought sampling and value feedback organically integrates planning and decision-making mechanisms, enabling effective search inside a solution tree. Traditional decision-making procedures usually require training dedicated reward and policy models as in reinforcement learning whereas we use the LM itself to provide the value estimates for decision making.

The Tree-of-Thought formulation is more versatile and handles challenging tasks on which GPT-4 only achieves very low accuracy with standard prompts.

Deliberate search such as ToT might not be necessary for many existing tasks that GPT-4 already excels at, and as an initial step this work only explores three relatively simple tasks that challenges GPT-4 and calls of better search and planning abilities incorporated with LMs. However, as we begin to deploy LMs for more real-world decision making applications (e.g. coding, data analysis, robotics, etc.), more complex tasks could emerge and present new opportunities to study these research questions. Also, search methods like ToT requires more resources (e.g. GPT-4 API cost) than sampling methods in order to improve task performances, but the modular flexibility of ToT allows users to customize such performance-cost tradeoffs.

Read More

Total
0
Shares
Leave a Reply

Your email address will not be published.

Related Posts
Bjarki and Thomas Harrington Rawle skewer online personas on ‘I Wish I Was A Model’
Read More

Bjarki and Thomas Harrington Rawle skewer online personas on ‘I Wish I Was A Model’

Stock footage and AI techniques combine on Rawle’s sardonic visual for Bjarki’s new single. Bjarki’s ‘I Wish I Was A Model’ is the latest release on Differance, the new label from the Icelandic techno producer and visual artist Thomas Harrington Rawle. The track’s visual is an extension of Rawle’s previous work in his Care More…
Tesla SVP Drew Baglino Discusses Longterm Energy Transition
Read More

Tesla SVP Drew Baglino Discusses Longterm Energy Transition

Tesla Senior Vice President, Powertrain and Energy Engineering, Drew Baglino discusses the long term issues building the battery and electric vehicle supply chain. This was at a Stanford University event. Drew talked with Yi Cui, Professor of Materials Science and Engineering and Director of the Stanford Precourt Institute for Energy and Will Chueh, Associate Professor…
Flying to Jupiter in 30 Days at 0.1% of Lightspeed
Read More

Flying to Jupiter in 30 Days at 0.1% of Lightspeed

The Jupiter Observing Velocity Experiment (JOVE) is a solar-powered technology demonstration of rapid flight to outer solar system targets, performing a flyby of the Jovian magnetosphere 30 days after launch. This is achieved using a magnetic drag device to accelerate with the solar wind plasma. This “Wind Rider” propulsion system can potentially also decelerate against…
Who Picks Up the Pieces of Silicon Valley Bank?
Read More

Who Picks Up the Pieces of Silicon Valley Bank?

Silicon Valley Bank (SVB) failed on Friday. It was the 16th largest bank in the USA. SVB gave loans to startups and the terms of the loans required the startup to keep deposits with SVB. Venture Capitalists for decades would require that startups use SVB and other “standard” best practice legal and financial structures. SVB…