Close Menu
    Facebook X (Twitter) Instagram
    Facebook Instagram YouTube
    Crypto Go Lore News
    Subscribe
    Sunday, June 8
    • Home
    • Market Analysis
    • Latest
      • Bitcoin News
      • Ethereum News
      • Altcoin News
      • Blockchain News
      • NFT News
      • Market Analysis
      • Mining News
      • Technology
      • Videos
    • Trending Cryptos
    • AI News
    • Market Cap List
    • Mining
    • Trading
    • Contact
    Crypto Go Lore News
    Home»AI News»With Quiet-STaR, language models learn to think before speaking
    AI News

    With Quiet-STaR, language models learn to think before speaking

    CryptoExpertBy CryptoExpertMarch 27, 2024No Comments6 Mins Read
    Share Facebook Twitter Pinterest Copy Link LinkedIn Tumblr Email VKontakte Telegram
    With Quiet-STaR, language models learn to think before speaking
    Share
    Facebook Twitter Pinterest Email Copy Link
    fiverr


    Join us in Atlanta on April 10th and explore the landscape of security workforce. We will explore the vision, benefits, and use cases of AI for security teams. Request an invite here.

    Humans are gifted with the ability to reason: “if” and “why” and the ability to “read between the lines” and infer unstated information are all critical to our problem-solving capabilities. 

    Up until now, AI models have, naturally, struggled in this area. But researchers from Stanford University and Notbad AI, Inc., have now revealed that they have taught AI models to think before they respond to prompts — just as (most) people consider what to say before speaking. 

    The researchers have introduced Quiet-STaR — an extension of the Self-Taught Reasoner (STaR) model — which is trained on a wide corpus of internet data and learns to generate rationales at each token to explain future text and improve predictions.

    itrust

    Quiet-STaR was applied to Mistral 7B, showing improvements to zero-shot direct reasoning abilities on the CommonsenseQA question-answering challenge (from 36.3% base to 47.2%) and the GSM8K grade school math word problems dataset (from 5.9% base to 10.9%). And, these improvements consistently increased with the number of tokens used in the model’s “internal thoughts.”

    VB Event

    The AI Impact Tour – Atlanta

    Continuing our tour, we’re headed to Atlanta for the AI Impact Tour stop on April 10th. This exclusive, invite-only event, in partnership with Microsoft, will feature discussions on how generative AI is transforming the security workforce. Space is limited, so request an invite today.

    Request an invite

    “Quiet-STaR marks a step towards LMs that can learn to reason in a more general and scalable way,” the researchers write. 

    Where AI reasoning has so far come up short

    Previous methods that have helped language models learn from their reasoning have been more hyper-focused and less generalized: AIs have been trained to solve individual tasks or predefined sets of tasks that rely on carefully curated datasets. 

    For instance, a pre-trained language model fine-tuned to output on human reasoning traces before answering multiple-choice questions outperformed an AI trained directly on answers, the Quiet-STaR developers pointed out. Other models, when provided with “scaffolding,” can generate chain-of-thought solutions without additional supervision. Further, researchers have “forced” models to use chain-of-thought reasoning by preventing them from answering unless completely confident. 

    “However, once again, these approaches only work for a question-answer dataset,” the Stanford University and Notbad AI, Inc., researchers contend. 

    STaR, particularly, proved that models could “bootstrap” their reasoning abilities on question-answering datasets. They could sample rationales to attempt to answer questions, train on those rationales if they led to correct answers and repeat iteratively to solve more and more difficult problems. 

    However, the Quiet-STaR researchers point out, that training from curated datasets limits the “scale and generalizability” of rationales. High-quality datasets will “inherently only ever cover a subset of reasoning tasks.”

    Inferring rationales from few-shot examples in question-answering is a “highly-constrained setting,” the researchers assert. “Ideally, a language model could instead learn to infer unstated rationales in arbitrary text.”

    By extending STaR, “we allow the LM to learn from the diverse tasks present in the language. To our knowledge, this is the first work explicitly training LMs to reason generally from text, rather than on curated reasoning tasks or collections of reasoning tasks.”

    ‘Quietly’ thinking

    The Stanford University and Notbad AI, Inc. researchers refer to their technique as Quiet-STaR because it applies STaR “quietly.” 

    The method generates many inner thoughts in parallel, at every token, to explain future text before responding to a prompt (i.e., the process of “thinking”). When the AI finally answers, it produces a mixture of predictions with and without rationales. 

    The REINFORCE algorithm was then applied; in reinforcement learning, this collects samples in an episode to update policy parameters as well as start-of-thought and end-of-thought embeddings. Researchers explain that this helps increase the likelihood that the AI will accurately predict future text. As part of this, the model also discards incorrect predictions. 

    “By iteratively optimizing these parameters, Quiet-STaR trains the model to generate more useful rationales throughout training,” the researchers write. 

    Because their goal was generalist reasoning, they used a zero-shot prompt (“Let’s think step by step”) without in-context examples. Quiet-STaR was applied to Mistral 7B using the web text datasets OpenWebMath and Colossal Clean Crawled Corpus. 

    “Quiet-STaR… allows a model to think quietly at every token, with a distribution trained to be useful,” researchers write. 

    They add that, “by training on the rich spectrum of reasoning tasks implicit in diverse web text, rather than narrowly specializing for particular datasets, Quiet-STaR points the way to more robust and adaptable language models.”

    Closing the gap between model and human reasoning capabilities

    Notably, researchers created a parallel sampling algorithm that generates rationales from all tokens in a string. This allowed the tokens to “pay attention to themselves,” all preceding tokens with the same thought and the preceding text. This allows for “continuations of all of the thoughts in parallel,” and each inference call generates an additional token for all tokens. 

    Researchers introduced custom meta-tokens at the beginning and the end of each thought. <|startofthought|> and <|endofthought|> were initialized with the em dash, ”—”, which is often used to denote a pause. 

    “Intuitively, the start thought tokens can be understood as putting the model into a ‘thinking mode,’” the researchers explain, “and the end thought token can be understood as telling the model when it’s done thinking.”

    The next step incorporated what’s known as a “mixing head,” a “shallow” multilayer perceptron. This helped researchers retrospectively determine how much to incorporate the next-token prediction from a given thought into the current next-token prediction.

    Finally, researchers optimized parameters to increase the likelihood of more probable future text. Reinforcement techniques provide a “learning signal” to rationales based on their impact on future predictions. To help reduce variance, researchers also introduced a “teacher forcing” trick, which ensures that neural networks stay as close as possible to ground truth sequences. 

    Ultimately, “Quiet-STaR represents a step towards language models that can learn to reason in a general and scalable way,” the researchers conclude. “Future work can build on these insights to further close the gap between language model and human-like reasoning capabilities.”

    VB Daily

    Stay in the know! Get the latest news in your inbox daily

    By subscribing, you agree to VentureBeat’s Terms of Service.

    Thanks for subscribing. Check out more VB newsletters here.

    An error occured.



    Source link

    Tokenmetrics
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Telegram Copy Link
    CryptoExpert
    • Website

    Related Posts

    AI News

    Privacy is the most fundamental aspect of human rights! #ai #ainews #chatgpt #openai #technews

    June 7, 2025
    AI News

    Test your AI knowledge | Fun AI Quiz for beginners & Developers

    June 6, 2025
    AI News

    Struggling with One Part? Let AI Guide You, Not Replace You #ai #shorts #homework

    June 5, 2025
    AI News

    Nude photo dikhai parliament me #news #nude #ai #parliament #newsupdate #foryou #shortsvideo #short

    June 4, 2025
    AI News

    Top 10 AI Tools in 2025 🔥 | Life-Changing Tools for Beginners | AI Use at 55 Story

    June 3, 2025
    AI News

    What if the characters knew they were fake? 🤯 #ai #shorts #veo3 #aigenerated

    June 2, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Recommended
    Editors Picks

    Patent hoarder sues BTC miners over Bitcoin using its IP

    June 8, 2025

    How a 91% Audit Score Signals DeFi’s Maturing Moment

    June 8, 2025

    TRUMP Meme Coin is Unlikely to Recover Soon – Here’s Why

    June 8, 2025

    Privacy is the most fundamental aspect of human rights! #ai #ainews #chatgpt #openai #technews

    June 7, 2025
    Latest Posts

    We are a leading platform dedicated to delivering authoritative insights, news, and resources on cryptocurrencies and blockchain technology. At Crypto Go Lore News, our mission is to empower individuals and businesses with reliable, actionable, and up-to-date information about the cryptocurrency ecosystem. We aim to bridge the gap between complex blockchain technology and practical understanding, fostering a more informed global community.

    Latest Posts

    Patent hoarder sues BTC miners over Bitcoin using its IP

    June 8, 2025

    How a 91% Audit Score Signals DeFi’s Maturing Moment

    June 8, 2025

    TRUMP Meme Coin is Unlikely to Recover Soon – Here’s Why

    June 8, 2025
    Newsletter

    Subscribe to Updates

    Get the latest Crypto news from Crypto Golore News about crypto around the world.

    Facebook Instagram YouTube
    • Contact
    • Privacy Policy
    • Terms Of Service
    • Social Media Disclaimer
    • DMCA Compliance
    • Anti-Spam Policy
    © 2025 CryptoGoLoreNews. All rights reserved by CryptoGoLoreNews.

    Type above and press Enter to search. Press Esc to cancel.

    bitcoin
    Bitcoin (BTC) $ 105,398.24
    ethereum
    Ethereum (ETH) $ 2,514.75
    tether
    Tether (USDT) $ 1.00
    xrp
    XRP (XRP) $ 2.21
    bnb
    BNB (BNB) $ 649.61
    solana
    Solana (SOL) $ 148.92
    usd-coin
    USDC (USDC) $ 0.999971
    dogecoin
    Dogecoin (DOGE) $ 0.182704
    tron
    TRON (TRX) $ 0.28603
    cardano
    Cardano (ADA) $ 0.659727