Close Menu
    Facebook X (Twitter) Instagram
    Facebook Instagram YouTube
    Crypto Go Lore News
    Subscribe
    Sunday, June 8
    • Home
    • Market Analysis
    • Latest
      • Bitcoin News
      • Ethereum News
      • Altcoin News
      • Blockchain News
      • NFT News
      • Market Analysis
      • Mining News
      • Technology
      • Videos
    • Trending Cryptos
    • AI News
    • Market Cap List
    • Mining
    • Trading
    • Contact
    Crypto Go Lore News
    Home»AI News»Google AI Introduces an Efficient Machine Learning Method to Scale Transformer-based Large Language Models (LLMs) to Infinitely Long Inputs
    AI News

    Google AI Introduces an Efficient Machine Learning Method to Scale Transformer-based Large Language Models (LLMs) to Infinitely Long Inputs

    CryptoExpertBy CryptoExpertApril 15, 2024No Comments4 Mins Read
    Share Facebook Twitter Pinterest Copy Link LinkedIn Tumblr Email VKontakte Telegram
    Google AI Introduces an Efficient Machine Learning Method to Scale Transformer-based Large Language Models (LLMs) to Infinitely Long Inputs
    Share
    Facebook Twitter Pinterest Email Copy Link
    Ledger


    Memory is significant for intelligence as it helps to recall past experiences and apply them to current situations. However, because of the way their attention mechanism works, both conventional Transformer models and Transformer-based Large Language Models (LLMs) have limitations when it comes to context-dependent memory. The memory consumption and computation time of this attention mechanism are both quadratic in complexity.

    Compressive memory systems present a viable substitute, with the objective of being more efficient and scalable for managing very lengthy sequences. Compressive memory systems keep storage and computation costs in check by maintaining a constant number of parameters for storing and retrieving information, in contrast to classical attention mechanisms that need memory to expand with the duration of the input sequence. 

    The goal of this system’s parameter adjustment process is to assimilate new information into memory while maintaining its retrievability. However, an efficient compressive memory method that strikes a compromise between simplicity and quality has not yet been adopted by existing LLMs.

    To overcome these limitations, a team of researchers from Google has proposed a unique solution that allows Transformer LLMs to handle arbitrarily lengthy inputs with a constrained memory footprint and computing power. A key component of their approach is an attention mechanism known as Infini-attention, which combines long-term linear attention and masked local attention into a single Transformer block and includes compressive memory in the conventional attention process.

    okex

    The primary breakthrough of Infini-attention is its capacity to effectively manage memory while processing lengthy sequences. The model can store and recall data with a fixed set of parameters by using compressive memory, which eliminates the requirement for memory to expand with the length of the input sequence. This keeps computing costs within reasonable bounds and helps control memory consumption.

    The team has shared that this method has shown to be effective in a number of tasks, such as book summarising tasks with input sequences of 500,000 tokens, passkey context block retrieval for sequences up to 1 million tokens in length, and long-context language modeling benchmarks. LLMs of sizes ranging from 1 billion to 8 billion parameters have been used to solve these tasks. 

    The ability to include minimal bounded memory parameters, that is, to limit and anticipate the model’s memory requirements, is one of this approach’s main advantages. Also, fast streaming inference for LLMs has been made possible by the suggested approach, which makes it possible to analyze sequential input efficiently in real-time or almost real-time circumstances. 

    The team has summarized their primary contributions as follows,

    The team has presented Infini-attention, a unique attention mechanism that blends local causal attention with long-term compressive memory. This method is both useful and effective since it effectively represents contextual dependencies over both short and long distances. 

    The standard scaled dot-product attention mechanism needs only be slightly altered to accommodate infini-attention. This enables plug-and-play continuous pre-training and long-context adaptation, and makes incorporation into current Transformer structures simple. 

    The method keeps constrained memory and computational resources while allowing Transformer-based LLMs to accommodate endlessly long contexts. The approach guarantees optimal resource utilization by processing very long inputs in a streaming mode, which enables LLMs to function well in large-scale data real-world applications.

    In conclusion, this study is a major step forward for LLMs, allowing for the efficient handling of very long inputs in terms of computation and memory utilization.

    Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. Join our Telegram Channel, Discord Channel, and LinkedIn Group.

    If you like our work, you will love our newsletter..

    Don’t Forget to join our 40k+ ML SubReddit

    Want to get in front of 1.5 Million AI Audience? Work with us here

    Tanya Malhotra is a final year undergrad from the University of Petroleum & Energy Studies, Dehradun, pursuing BTech in Computer Science Engineering with a specialization in Artificial Intelligence and Machine Learning.She is a Data Science enthusiast with good analytical and critical thinking, along with an ardent interest in acquiring new skills, leading groups, and managing work in an organized manner.

    🐝 Join the Fastest Growing AI Research Newsletter Read by Researchers from Google + NVIDIA + Meta + Stanford + MIT + Microsoft and many others…



    Source link

    Tokenmetrics
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Telegram Copy Link
    CryptoExpert
    • Website

    Related Posts

    AI News

    Privacy is the most fundamental aspect of human rights! #ai #ainews #chatgpt #openai #technews

    June 7, 2025
    AI News

    Test your AI knowledge | Fun AI Quiz for beginners & Developers

    June 6, 2025
    AI News

    Struggling with One Part? Let AI Guide You, Not Replace You #ai #shorts #homework

    June 5, 2025
    AI News

    Nude photo dikhai parliament me #news #nude #ai #parliament #newsupdate #foryou #shortsvideo #short

    June 4, 2025
    AI News

    Top 10 AI Tools in 2025 🔥 | Life-Changing Tools for Beginners | AI Use at 55 Story

    June 3, 2025
    AI News

    What if the characters knew they were fake? 🤯 #ai #shorts #veo3 #aigenerated

    June 2, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Recommended
    Editors Picks

    Privacy is the most fundamental aspect of human rights! #ai #ainews #chatgpt #openai #technews

    June 7, 2025

    Pumpfun pe memecoin kaise bnaye #crypto #guide

    June 7, 2025

    Bitcoin-News on mining-guide.com

    June 7, 2025

    NFT artist relives ‘crypto tax nightmare’ in new song

    June 7, 2025
    Latest Posts

    We are a leading platform dedicated to delivering authoritative insights, news, and resources on cryptocurrencies and blockchain technology. At Crypto Go Lore News, our mission is to empower individuals and businesses with reliable, actionable, and up-to-date information about the cryptocurrency ecosystem. We aim to bridge the gap between complex blockchain technology and practical understanding, fostering a more informed global community.

    Latest Posts

    Privacy is the most fundamental aspect of human rights! #ai #ainews #chatgpt #openai #technews

    June 7, 2025

    Pumpfun pe memecoin kaise bnaye #crypto #guide

    June 7, 2025

    Bitcoin-News on mining-guide.com

    June 7, 2025
    Newsletter

    Subscribe to Updates

    Get the latest Crypto news from Crypto Golore News about crypto around the world.

    Facebook Instagram YouTube
    • Contact
    • Privacy Policy
    • Terms Of Service
    • Social Media Disclaimer
    • DMCA Compliance
    • Anti-Spam Policy
    © 2025 CryptoGoLoreNews. All rights reserved by CryptoGoLoreNews.

    Type above and press Enter to search. Press Esc to cancel.

    bitcoin
    Bitcoin (BTC) $ 105,682.32
    ethereum
    Ethereum (ETH) $ 2,526.44
    tether
    Tether (USDT) $ 1.00
    xrp
    XRP (XRP) $ 2.18
    bnb
    BNB (BNB) $ 651.39
    solana
    Solana (SOL) $ 150.10
    usd-coin
    USDC (USDC) $ 1.00
    dogecoin
    Dogecoin (DOGE) $ 0.185184
    tron
    TRON (TRX) $ 0.286969
    cardano
    Cardano (ADA) $ 0.665216