Friday, May 16, 2025
News PouroverAI
Visit PourOver.AI
No Result
View All Result
  • Home
  • AI Tech
  • Business
  • Blockchain
  • Data Science & ML
  • Cloud & Programming
  • Automation
  • Front-Tech
  • Marketing
  • Home
  • AI Tech
  • Business
  • Blockchain
  • Data Science & ML
  • Cloud & Programming
  • Automation
  • Front-Tech
  • Marketing
News PouroverAI
No Result
View All Result

This AI Research from Cohere AI Introduces the Mixture of Vectors (MoV) and Mixture of LoRA (MoLORA) to Mitigate the Challenges Associated with Scaling Instruction-Tuned LLMs at Scale

December 22, 2023
in Data Science & ML
Reading Time: 3 mins read
0 0
A A
0
Share on FacebookShare on Twitter


With the growing advancements in the field of Artificial Intelligence (AI), researchers are constantly coming up with new transformations and innovations. One such pioneering development is in the domain of Mixture of Experts (MoE) architecture, a well-known neural framework known for its capacity to maximize overall performance at a constant computing cost.

However, when AI models get bigger, traditional MoEs have trouble keeping track of every memory expert. To overcome this, in recent research, a team of Cohere researchers has studied about ways to expand the capabilities of MoE by presenting a very parameter-efficient version that solves these scalability problems. Lightweight experts have been combined with the MoE architecture in order to achieve this.

The suggested MoE architecture is a highly effective approach for parameter-efficient fine-tuning (PEFT) as it surpasses the drawbacks of conventional models. The team has shared that incorporating lightweight experts is the primary innovation enabling the model to surpass conventional PEFT techniques. Even when updating only the lightweight experts, which is less than 1% of a model with 11 billion parameters, the performance demonstrated was comparable to full fine-tuning.

The model’s capacity to generalize to tasks that haven’t been seen before, highlighting its independence from prior task knowledge, is one amazing feature of the research. This suggests that the proposed MoE architecture is not limited to particular domains and can successfully adjust to new tasks.

The results have demonstrated the adaptability of the combination of skilled architects. The suggested MoE variant has shown great performance in spite of strict parameter limits, which emphasizes how flexible and effective MoEs are, especially in difficult situations with constrained resources.

The team has summarized their primary contributions as follows.

The research presents a unique design incorporating lightweight and modular experts to improve the Mixture of Experts (MoEs). This makes it possible to fine-tune dense models with low efficiency of less than 1% parameter updates.

The suggested techniques often beat conventional parameter-efficient techniques in fine-tuning instructions, exhibiting better results on untested tasks. Notable improvements have been achieved by the Mixture of (IA)³ Vectors (MoV), which outperforms the standard (IA)³ at 3B and 11B model sizes by up to 14.57% and 8.39%, respectively. This superiority holds true for a variety of scales, expert variations, model types, and trainable parameter budgets.

The study has shown that, with only a small percentage of the model parameters updated, the suggested MoV architecture can perform comparably to complete fine-tuning at large scales. Results from 8 previously unpublished tasks have shown competitive performance with far lower computational costs, just 0.32% and 0.86% of the parameters in the 3B and 11B models, respectively.

 

In-depth ablation studies have been carried out to systematically assess the effectiveness of several MoE architectures and Parameter-Efficient Fine-Tuning (PEFT) techniques, which highlight how sensitive MoE is to hyperparameter optimization and cover a wide range of model sizes, adapter kinds, expert counts, and routing strategies.

Check out the Paper and Github. All credit for this research goes to the researchers of this project. Also, don’t forget to join our 34k+ ML SubReddit, 41k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the latest AI research news, cool AI projects, and more.

If you like our work, you will love our newsletter.

Tanya Malhotra is a final year undergrad from the University of Petroleum & Energy Studies, Dehradun, pursuing BTech in Computer Science Engineering with a specialization in Artificial Intelligence and Machine Learning.She is a Data Science enthusiast with good analytical and critical thinking, along with an ardent interest in acquiring new skills, leading groups, and managing work in an organized manner.

🚀 Boost your LinkedIn presence with Taplio: AI-driven content creation, easy scheduling, in-depth analytics, and networking with top creators – Try it free now!.



Source link

Tags: challengesCohereInstructionTunedIntroducesLLMsLoRAMitigateMixtureMoLORAMoVResearchscalescalingVectors
Previous Post

Researchers from Apple Unveil DataComp: A Groundbreaking 12.8 Billion Image-Text Pair Dataset for Advanced Machine Learning Model Development and Benchmarking

Next Post

How To Stay Ahead of the Ever-Changing Digital Threats

Related Posts

AI Compared: Which Assistant Is the Best?
Data Science & ML

AI Compared: Which Assistant Is the Best?

June 10, 2024
5 Machine Learning Models Explained in 5 Minutes
Data Science & ML

5 Machine Learning Models Explained in 5 Minutes

June 7, 2024
Cohere Picks Enterprise AI Needs Over ‘Abstract Concepts Like AGI’
Data Science & ML

Cohere Picks Enterprise AI Needs Over ‘Abstract Concepts Like AGI’

June 7, 2024
How to Learn Data Analytics – Dataquest
Data Science & ML

How to Learn Data Analytics – Dataquest

June 6, 2024
Adobe Terms Of Service Update Privacy Concerns
Data Science & ML

Adobe Terms Of Service Update Privacy Concerns

June 6, 2024
Build RAG applications using Jina Embeddings v2 on Amazon SageMaker JumpStart
Data Science & ML

Build RAG applications using Jina Embeddings v2 on Amazon SageMaker JumpStart

June 6, 2024
Next Post
How To Stay Ahead of the Ever-Changing Digital Threats

How To Stay Ahead of the Ever-Changing Digital Threats

Revenera launches monetisation analytics dashboard

Revenera launches monetisation analytics dashboard

Risk Management Strategies for Blockchain Security

Risk Management Strategies for Blockchain Security

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

  • Trending
  • Comments
  • Latest
Is C.AI Down? Here Is What To Do Now

Is C.AI Down? Here Is What To Do Now

January 10, 2024
23 Plagiarism Facts and Statistics to Analyze Latest Trends

23 Plagiarism Facts and Statistics to Analyze Latest Trends

June 4, 2024
Porfo: Revolutionizing the Crypto Wallet Landscape

Porfo: Revolutionizing the Crypto Wallet Landscape

October 9, 2023
A Complete Guide to BERT with Code | by Bradney Smith | May, 2024

A Complete Guide to BERT with Code | by Bradney Smith | May, 2024

May 19, 2024
How To Build A Quiz App With JavaScript for Beginners

How To Build A Quiz App With JavaScript for Beginners

February 22, 2024
Saginaw HMI Enclosures and Suspension Arm Systems from AutomationDirect – Library.Automationdirect.com

Saginaw HMI Enclosures and Suspension Arm Systems from AutomationDirect – Library.Automationdirect.com

December 6, 2023
Can You Guess What Percentage Of Their Wealth The Rich Keep In Cash?

Can You Guess What Percentage Of Their Wealth The Rich Keep In Cash?

June 10, 2024
AI Compared: Which Assistant Is the Best?

AI Compared: Which Assistant Is the Best?

June 10, 2024
How insurance companies can use synthetic data to fight bias

How insurance companies can use synthetic data to fight bias

June 10, 2024
5 SLA metrics you should be monitoring

5 SLA metrics you should be monitoring

June 10, 2024
From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

June 10, 2024
UGRO Capital: Targeting to hit milestone of Rs 20,000 cr loan book in 8-10 quarters: Shachindra Nath

UGRO Capital: Targeting to hit milestone of Rs 20,000 cr loan book in 8-10 quarters: Shachindra Nath

June 10, 2024
Facebook Twitter LinkedIn Pinterest RSS
News PouroverAI

The latest news and updates about the AI Technology and Latest Tech Updates around the world... PouroverAI keeps you in the loop.

CATEGORIES

  • AI Technology
  • Automation
  • Blockchain
  • Business
  • Cloud & Programming
  • Data Science & ML
  • Digital Marketing
  • Front-Tech
  • Uncategorized

SITEMAP

  • Disclaimer
  • Privacy Policy
  • DMCA
  • Cookie Privacy Policy
  • Terms and Conditions
  • Contact us

Copyright © 2023 PouroverAI News.
PouroverAI News

No Result
View All Result
  • Home
  • AI Tech
  • Business
  • Blockchain
  • Data Science & ML
  • Cloud & Programming
  • Automation
  • Front-Tech
  • Marketing

Copyright © 2023 PouroverAI News.
PouroverAI News

Welcome Back!

Login to your account below

Forgotten Password? Sign Up

Create New Account!

Fill the forms bellow to register

All fields are required. Log In

Retrieve your password

Please enter your username or email address to reset your password.

Log In