Close Menu
Karachi Chronicle
  • Home
  • AI
  • Business
  • Entertainment
  • Fashion
  • Politics
  • Sports
  • Tech
  • World

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

What's Hot

Joni Ernst: Iowa Senator Joni Ernst won’t launch a major Senate race in 2026

UFC legend explains why athletes like LeBron James are redefineing longevity

Cryptocurrency Live News & Updates: Vaneck proposes SolanaETF for traditional investors

Facebook X (Twitter) Instagram
  • Home
  • About us
  • Advertise
  • Contact us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
Facebook X (Twitter) Instagram Pinterest Vimeo
Karachi Chronicle
  • Home
  • AI
  • Business
  • Entertainment
  • Fashion
  • Politics
  • Sports
  • Tech
  • World
Karachi Chronicle
You are at:Home » Meta AI introduces Cocomix: a pre-deletion framework that integrates token prediction with continuous concepts
AI

Meta AI introduces Cocomix: a pre-deletion framework that integrates token prediction with continuous concepts

Adnan MaharBy Adnan MaharFebruary 13, 2025No Comments4 Mins Read0 Views
Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Reddit
Share
Facebook Twitter LinkedIn Pinterest WhatsApp Email


The dominant approach to pretraining large-scale language models (LLMS) relies on the following token predictions and has proven effective in capturing language patterns. However, this method has significant limitations. Language tokens often need to convey surface-level information and models need to process huge amounts of data to develop deeper inference capabilities. Furthermore, token-based learning struggles with long-term dependency capture, making tasks that require planning and abstraction more difficult. Researchers have investigated alternative strategies such as knowledge distillation and structured input enhancement, but these approaches do not completely address the limitations of token-based learning. This raises an important question. Can LLM train in a way that combines token-level processing with conceptual understanding? Meta AI introduces continuous concept mixing (CoComix) as a potential solution.

Cocomix: Another Approach to Pretraining

Cocomix integrates token predictions with modeling of continuous concepts derived from hidden states of preprocessed models. This method incorporates the training process by employing a sparse autoencoder (SAE) to extract high-level semantic representations and interior them with token embedding. This design allows the model to increase its ability to recognize and process broader concept structures, while retaining the benefits of token-based learning. By enriching the token-based paradigm with concept-level information, Cocomix aims to improve inference efficiency and model interpretability.

Technical details and benefits

Cocomix works through three main components:

Concept extraction via Sparse Auto Encoder (SAE): Preprocessed SAEs identify potential semantic features from hidden states of models and capture information beyond individual tokens. Concept selection with attribute scoring: Not all extracted concepts contribute equally to prediction. Cocomix employs attribution methods to determine which concepts are most influential and should be retained. Interrade a continuous concept with token representation: The selected concept is compressed into a continuous vector and integrated into a hidden state with the embedding of the token, making it possible for the model to utilize both the token level and conceptual information.

This approach increases sample efficiency and allows for comparable performance when the model has fewer tokens. Additionally, CoComix improves interpretability by allowing extracted concepts to be inspected and adjusted, and gives a clearer view of how the model handles information.

Performance and evaluation

Meta AI has evaluated Cocomix across multiple benchmarks, including OpenWebtext, Lambada, Wikitext-103, Hellaswag, Piqa, Siqa, Arc-Easy, and Winogrande. The findings show that:

Improved sample efficiency: CoComix requires 21.5% less tokens than training tokens, while consistent with performance of the following token predictions. Enhanced generalization: Over a variety of model sizes (69m, 386m, and 1.38b parameters), Cocomix showed consistent improvements in downstream task performance. Effective Knowledge Transfer: Cocomix supports knowledge transfer from smaller models to larger models, surpassing traditional knowledge distillation techniques. Greater interpretability: Continuous concept integration provides greater control and transparency in model decision-making, allowing for a clearer understanding of internal processes.

Conclusion

Cocomix presents an alternative approach to LLM pre-initiation by combining token prediction with concept-based inference. By incorporating structured representations extracted via SAE, CoComix improves efficiency and interpretability without disrupting the underlying next token prediction framework. Experimental results suggest that this method provides a balanced method for improving training of language models, especially in areas that require structured reasoning and transparent decision-making. Future research could focus on improving concept extraction methods and further integration of continuous representations into predeletion workflows.

Please see the paper and the github page. All credits for this study will be sent to researchers in this project. Also, feel free to follow us on Twitter. Don’t forget to join 75K+ ML SubredDit.

Commended open source AI platform recommended: “Intelagent is an open source multi-agent framework for evaluating complex conversational AI systems” (promotion)

Asif Razzaq is CEO of Marktechpost Media Inc.. As a visionary entrepreneur and engineer, ASIF is committed to leveraging the possibilities of artificial intelligence for social benefits. His latest efforts are the launch of MarkTechPost, an artificial intelligence media platform. This is distinguished by its detailed coverage of machine learning and deep learning news, and is easy to understand by a technically sound and wide audience. The platform has over 2 million views each month, indicating its popularity among audiences.

✅ (Recommended) Join the Telegram Channel



Source link

Share. Facebook Twitter Pinterest LinkedIn Reddit WhatsApp Telegram Email
Previous ArticleMany ways America can first hurt global trade
Next Article Build a dynamic, role-based AI agent using Amazon Bedrock inline agents
Adnan Mahar
  • Website

Adnan is a passionate doctor from Pakistan with a keen interest in exploring the world of politics, sports, and international affairs. As an avid reader and lifelong learner, he is deeply committed to sharing insights, perspectives, and thought-provoking ideas. His journey combines a love for knowledge with an analytical approach to current events, aiming to inspire meaningful conversations and broaden understanding across a wide range of topics.

Related Posts

Dig into Google Deepmind CEO “Shout Out” Chip Engineers and Openai CEO Sam Altman, Sundar Pichai responds with emojis

June 1, 2025

Google, Nvidia invests in AI startup Safe Superintelligence, co-founder of Openai Ilya Sutskever

April 14, 2025

This $30 billion AI startup can be very strange by a man who said that neural networks may already be aware of it

February 24, 2025
Leave A Reply Cancel Reply

Top Posts

20 Most Anticipated Sex Movies of 2025

January 22, 2025349 Views

President Trump’s SEC nominee Paul Atkins marries multi-billion dollar roof fortune

December 14, 2024113 Views

How to tell the difference between fake and genuine Adidas Sambas

December 26, 202475 Views

Alice Munro’s Passive Voice | New Yorker

December 23, 202472 Views
Don't Miss
AI June 1, 2025

Dig into Google Deepmind CEO “Shout Out” Chip Engineers and Openai CEO Sam Altman, Sundar Pichai responds with emojis

Demis Hassabis, CEO of Google Deepmind, has expanded public approval to its chip engineers, highlighting…

Google, Nvidia invests in AI startup Safe Superintelligence, co-founder of Openai Ilya Sutskever

This $30 billion AI startup can be very strange by a man who said that neural networks may already be aware of it

As Deepseek and ChatGpt Surge, is Delhi behind?

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

About Us
About Us

Welcome to Karachi Chronicle, your go-to source for the latest and most insightful updates across a range of topics that matter most in today’s fast-paced world. We are dedicated to delivering timely, accurate, and engaging content that covers a variety of subjects including Sports, Politics, World Affairs, Entertainment, and the ever-evolving field of Artificial Intelligence.

Facebook X (Twitter) Pinterest YouTube WhatsApp
Our Picks

Joni Ernst: Iowa Senator Joni Ernst won’t launch a major Senate race in 2026

UFC legend explains why athletes like LeBron James are redefineing longevity

Cryptocurrency Live News & Updates: Vaneck proposes SolanaETF for traditional investors

Most Popular

10 things you should never say to an AI chatbot

November 10, 20040 Views

Character.AI faces lawsuit over child safety concerns

December 12, 20050 Views

Analyst warns Salesforce investors about AI agent optimism

July 1, 20070 Views
© 2025 karachichronicle. Designed by karachichronicle.
  • Home
  • About us
  • Advertise
  • Contact us
  • DMCA
  • Privacy Policy
  • Terms & Conditions

Type above and press Enter to search. Press Esc to cancel.