Close Menu
    Facebook X (Twitter) Instagram
    Facebook Instagram YouTube
    Crypto Go Lore News
    Subscribe
    Monday, June 9
    • Home
    • Market Analysis
    • Latest
      • Bitcoin News
      • Ethereum News
      • Altcoin News
      • Blockchain News
      • NFT News
      • Market Analysis
      • Mining News
      • Technology
      • Videos
    • Trending Cryptos
    • AI News
    • Market Cap List
    • Mining
    • Trading
    • Contact
    Crypto Go Lore News
    Home»AI News»What Exactly Do Large Language Models Really Understand?
    AI News

    What Exactly Do Large Language Models Really Understand?

    CryptoExpertBy CryptoExpertMarch 1, 2024No Comments6 Mins Read
    Share Facebook Twitter Pinterest Copy Link LinkedIn Tumblr Email VKontakte Telegram
    What Exactly Do Large Language Models Really Understand?
    Share
    Facebook Twitter Pinterest Email Copy Link
    Paxful


    understanding may emerge with scale

    Dall-E

    Ever since these models appeared, researchers and philosophers have debated whether they are complex ‘stochastic parrots’ or whether they are indeed capable of understanding and will eventually be smarter than us.

    It’s an important question: if foundation models – the catch-all term – can truly understand in the way humans do, it might be a step towards superintelligence, and even a form of artificial sentience.

    The debate has been muddy because people often talk past each other due to differing definitions of ‘meaning’ and ‘understanding.’ But researchers at Amazon Web Services (AWS) have proposed using some uncontroversial definitions to help align the discussion – and according to those definitions, they believe foundation model understanding is not only possible but inevitable as these models scale.

    Stefano Soatto

    AWS

    “These definitions are useful in the sense that they align with definitions that people have used in mathematical logic, as well as mathematical model theory, as well as epistemology, dynamical systems, and linguistics and so on and so forth,” said Stefano Soatto, Amazon Web Services’ vice president and distinguished scientist, who coauthored a paper on the subject.

    For ‘meaning,’ the researchers use a definition first proposed by the 17th century philosopher Antoine Arnauld, which defined it as the relationship between linguistic form (the text the models are trained on) and the thing or concept that linguistic form points to (which is external to language). The word ‘tree’ is not a tree, but it points to the concept of ‘tree.’ The meaning of the word tree is its relationship with the concept ‘tree.’

    okex

    Soatto and his fellow applied scientist Matthew Trager explain in a paper published earlier this month that these meanings can be represented as either vectors in an abstract space, or probability distributions in the same space. This allows us to discuss notions of proximity (which meanings are like each other), and to some extent entailment (what implies what) and causality (what follows what).

    They define ‘understanding,’ meanwhile, as equivalence classes of expressions, which form a representation of that meaning – or the relationship between the word and what it refers to. Large Foundation Models represent such equivalence classes, viewed as either vectors or distribution of continuations. This allows them to reason and operate on the meaning without storing every detail, they said.

    Consider the number pi (π). Pi is an infinite, non-repeating decimal number (3.14159265358979…). No one could possibly memorize or store in their brain the entire infinite sequence of digits. But we can still understand and work with the concept of pi.

    How? By having a representation in our minds that captures the essential properties of pi – that it is the ratio of a circle’s circumference to its diameter, that it starts with 3.14, that it goes on forever without repeating, etc. With that mental representation, we don’t need to recall the entire unending string of digits.

    “Our representation is an abstract concept, just like the meaning of a sentence,” Soatto said in a phone call.

    More importantly, this representation allows us to answer specific questions and do things with the concept of pi. For example, if I asked you what the millionth digit of pi is, you wouldn’t have to recite a million digits to prove you understand pi. Your mental representation would allow you to work it out or look it up, using a finite amount of brainpower and time.

    So, understanding an abstract concept means forming a mental model or representation that captures the relationship between language describing a concept and key properties of that concept. That mental model or representation can be reasoned with and queried using our limited human memory, mental processing power, and access to data – rather than trying to literally store an infinite amount of raw data. Building a concise yet powerful representation, according to this view, demonstrates true understanding.

    with scale, the researchers believe, these models begin to understand

    Models trained as next-token predictors learn stochastic patterns. But over time, they identify relationships – meanings – between all the data they’ve seen and, as they scale, develop their own internal languages describing those relationships. Inner languages emerge when a model reaches hundreds of billions of parameters, Soatto says.

    Humans acquire language using only around 1.5 megabytes of data (less than a 1980s floppy disk). “Humans are proof that you don’t need specialized logic hardware or dedicated hardware with tight logical constraints to do math,” Soatto said.

    But Trager and Soatto suggest that large-scale models must “find their way” around their large “brains” to shed redundant information and improve relational discovery. By doing so, these models can enhance their ability to discover meaningful relationships within the data they are trained on.

    As that relational discovery improves with scale, the researchers believe these models begin to understand; in other words, form representations of those relationships, of those meanings, that they then can operate on. Even if they do not currently understand in this way, the researchers say, they likely will with continued scaling.

    There is a concept in biology of a ‘critical period,’ when an organism’s nervous system is particularly receptive to learning specific skills or traits. The period is marked by high levels of brain plasticity before neural connections become more fixed.

    Foundation models also have a critical period, the researchers believe, during which they need to accumulate a significant amount of information. The model can then discard some of this accumulated information while enhancing its ability to identify relationships among the data it encounters during inference.

    CLOSE-UP SQUIRREL HOLDING WALNUT

    getty

    It’s not clear when in the evolutionary tree that happens in biology – it’s not like a light switch between understanding meaning and not. It’s a spectrum that emerges with the scale of neurons in the brain. An amoeba probably doesn’t have it; a squirrel likely has some. As machines scale, these researchers believe that understanding of meaning, under their definition, will emerge as it does in biology.

    “Scale plays a key role in this emergent phenomenon,” Soatto said. “Where does it stop? Right now, there is no obvious end in sight.”

    The problem, he said, is that we may never definitively know when understanding emerges.

    With machine learning models at scale, you can prove that (a) they can represent abstract concepts, such as pi, (b) they currently do not, at least not reliably, and (c) if and when they do, we cannot know for sure. This is no different from humans: When we teach students, how do we know they “understood” the material? We administer an exam: if they fail, we know they did not. If they pass, do we know they got it? Had we asked one more question, would they have answered correctly? We won’t know until we ask”

    Yet, unlike humans, these large models are not opaque systems. We can measure, audit, and calibrate them. This is a key advantage over the opacity of the human mind and holds the promise of understanding not only how these properties emerge in a foundation model, but how they emerge in the human mind.



    Source link

    Betfury
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Telegram Copy Link
    CryptoExpert
    • Website

    Related Posts

    AI News

    Learn CSS Easily with AI _ Step-by-Step Guide for Beginners _ai _aitools _css _aicoding#viral#shorts

    June 8, 2025
    AI News

    Privacy is the most fundamental aspect of human rights! #ai #ainews #chatgpt #openai #technews

    June 7, 2025
    AI News

    Test your AI knowledge | Fun AI Quiz for beginners & Developers

    June 6, 2025
    AI News

    Struggling with One Part? Let AI Guide You, Not Replace You #ai #shorts #homework

    June 5, 2025
    AI News

    Nude photo dikhai parliament me #news #nude #ai #parliament #newsupdate #foryou #shortsvideo #short

    June 4, 2025
    AI News

    Top 10 AI Tools in 2025 🔥 | Life-Changing Tools for Beginners | AI Use at 55 Story

    June 3, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Recommended
    Editors Picks

    Over 60% of Pump.fun wallets lost money: report

    June 9, 2025

    Circle rejected Ripple’s $5 billion buyout — now valued at over $20 billion after NYSE debut

    June 8, 2025

    Bitcoin at $104K, but falling MVRV ratio hints at short-term correction

    June 8, 2025

    Learn CSS Easily with AI _ Step-by-Step Guide for Beginners _ai _aitools _css _aicoding#viral#shorts

    June 8, 2025
    Latest Posts

    We are a leading platform dedicated to delivering authoritative insights, news, and resources on cryptocurrencies and blockchain technology. At Crypto Go Lore News, our mission is to empower individuals and businesses with reliable, actionable, and up-to-date information about the cryptocurrency ecosystem. We aim to bridge the gap between complex blockchain technology and practical understanding, fostering a more informed global community.

    Latest Posts

    Over 60% of Pump.fun wallets lost money: report

    June 9, 2025

    Circle rejected Ripple’s $5 billion buyout — now valued at over $20 billion after NYSE debut

    June 8, 2025

    Bitcoin at $104K, but falling MVRV ratio hints at short-term correction

    June 8, 2025
    Newsletter

    Subscribe to Updates

    Get the latest Crypto news from Crypto Golore News about crypto around the world.

    Facebook Instagram YouTube
    • Contact
    • Privacy Policy
    • Terms Of Service
    • Social Media Disclaimer
    • DMCA Compliance
    • Anti-Spam Policy
    © 2025 CryptoGoLoreNews. All rights reserved by CryptoGoLoreNews.

    Type above and press Enter to search. Press Esc to cancel.

    bitcoin
    Bitcoin (BTC) $ 105,700.32
    ethereum
    Ethereum (ETH) $ 2,502.71
    tether
    Tether (USDT) $ 1.00
    xrp
    XRP (XRP) $ 2.25
    bnb
    BNB (BNB) $ 652.14
    solana
    Solana (SOL) $ 152.44
    usd-coin
    USDC (USDC) $ 1.00
    dogecoin
    Dogecoin (DOGE) $ 0.183256
    tron
    TRON (TRX) $ 0.282318
    cardano
    Cardano (ADA) $ 0.667175