Friday, May 9, 2025
News PouroverAI
Visit PourOver.AI
No Result
View All Result
  • Home
  • AI Tech
  • Business
  • Blockchain
  • Data Science & ML
  • Cloud & Programming
  • Automation
  • Front-Tech
  • Marketing
  • Home
  • AI Tech
  • Business
  • Blockchain
  • Data Science & ML
  • Cloud & Programming
  • Automation
  • Front-Tech
  • Marketing
News PouroverAI
No Result
View All Result

Amazon Researchers Introduce DistTGL: A Breakthrough in Scalable Memory-Based Temporal Graph Neural Networks for GPU Clusters

October 1, 2023
in AI Technology
Reading Time: 3 mins read
0 0
A A
0
Share on FacebookShare on Twitter


Numerous real-world graphs include crucial temporal domain data.

Both spatial and temporal information are crucial in spatial-temporal applications like traffic and weather forecasting.

Researchers have recently developed Temporal Graph Neural Networks (TGNNs) to take advantage of temporal information in dynamic graphs, building on the success of Graph Neural Networks (GNNs) in learning static graph representation.

TGNNs have shown superior accuracy on a variety of downstream tasks like temporal link prediction and dynamic node classification on a variety of dynamic graphs, including social network graphs, traffic graphs, and knowledge graphs, significantly outperforming static GNNs and other conventional methods.

On dynamic graphs, as time passes, there are more associated events on each node.

When this number is high, TGNNs are unable to fully capture the history using either temporal attention-based aggregation or historical neighbor sampling techniques.

Researchers have created Memory-based Temporal Graph Neural Networks (M-TGNNs) that store node-level memory vectors to summarize independent node history to make up for the lost history.

Despite M-TGNNs’ success, their poor scalability makes it challenging to implement them in large-scale production systems.

Due to the temporal dependencies that the auxiliary node memory generates, training mini-batches must be brief and scheduled in chronological sequence.

Utilizing data parallelism in M-TGNN training is particularly difficult in two ways:

  • Merely raising the batch size results in information loss and the loss of information about the temporal dependency between occurrences.
  • A unified version of the node memory must be accessed and maintained by all trainers, which creates a massive amount of remote traffic in distributed systems.

New research by the University of Southern California and AWS offers DistTGL, a scalable and effective method for M-TGNN training on distributed GPU clusters.

DistTGL enhances the current M-TGNN training systems in three ways:

  • Model: The accuracy and convergence rate of the M-TGNNs’ node memory is improved by introducing more static node memory.
  • Algorithm: To address the issues of accuracy loss and communication overhead in dispersed settings, the team provides a novel training algorithm.
  • System: To reduce the overhead associated with mini-batch generation, they develop an optimized system using prefetching and pipelining techniques.

DistTGL significantly improves on prior approaches in terms of convergence and training throughput.

DistTGL is the first effort that scales M-TGNN training to distributed GPU clusters.

Github has DistTGL publicly available.

They present two innovative parallel training methodologies — epoch parallelism and memory parallelism — based on the distinctive properties of M-TGNN training, which enable M-TGNNs to capture the same number of dependent graph events on several GPUs as on a single GPU.

Based on the dataset and hardware characteristics, they offer heuristic recommendations for selecting the best training setups.

The researchers serialize memory operations on the node memory and effectively execute them by a separate daemon process, eliminating complicated and expensive synchronizations to overlap mini-batch creation and GPU training.

In trials, DistTGL outperforms the state-of-the-art single-machine approach by more than 10 times when scaling to several GPUs in convergence rate.

Check out the Paper.

All Credit For This Research Goes To the Researchers on This Project.

Also, don’t forget to join our 30k+ ML SubReddit, 40k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the latest AI research news, cool AI projects, and more.

If you like our work, you will love our newsletter.

Dhanshree Shenwai is a Computer Science Engineer and has a good experience in FinTech companies covering Financial, Cards & Payments and Banking domain with keen interest in applications of AI.

She is enthusiastic about exploring new technologies and advancements in today’s evolving world making everyone’s life easy.

🚀 The end of project management by humans (Sponsored)



Source link

Tags: AmazonBreakthroughClustersDistTGLGPUgraphIntroduceMemoryBasednetworksNeuralResearchersScalableTemporal
Previous Post

Frontech LED monitor 49.53cm wide screen monitor FT-1993

Next Post

A guide to feature flags in Laravel using Laravel Pennant

Related Posts

How insurance companies can use synthetic data to fight bias
AI Technology

How insurance companies can use synthetic data to fight bias

June 10, 2024
From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset
AI Technology

From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

June 10, 2024
How Game Theory Can Make AI More Reliable
AI Technology

How Game Theory Can Make AI More Reliable

June 9, 2024
Decoding Decoder-Only Transformers: Insights from Google DeepMind’s Paper
AI Technology

Decoding Decoder-Only Transformers: Insights from Google DeepMind’s Paper

June 9, 2024
Buffer of Thoughts (BoT): A Novel Thought-Augmented Reasoning AI Approach for Enhancing Accuracy, Efficiency, and Robustness of LLMs
AI Technology

Buffer of Thoughts (BoT): A Novel Thought-Augmented Reasoning AI Approach for Enhancing Accuracy, Efficiency, and Robustness of LLMs

June 9, 2024
Deciphering Doubt: Navigating Uncertainty in LLM Responses
AI Technology

Deciphering Doubt: Navigating Uncertainty in LLM Responses

June 9, 2024
Next Post
A guide to feature flags in Laravel using Laravel Pennant

A guide to feature flags in Laravel using Laravel Pennant

When Does Optimizing a Proper Loss Yield Calibration?

When Does Optimizing a Proper Loss Yield Calibration?

How Forbes Identifies The Most Promising AI Companies In The World

How Forbes Identifies The Most Promising AI Companies In The World

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

  • Trending
  • Comments
  • Latest
Is C.AI Down? Here Is What To Do Now

Is C.AI Down? Here Is What To Do Now

January 10, 2024
Porfo: Revolutionizing the Crypto Wallet Landscape

Porfo: Revolutionizing the Crypto Wallet Landscape

October 9, 2023
A Complete Guide to BERT with Code | by Bradney Smith | May, 2024

A Complete Guide to BERT with Code | by Bradney Smith | May, 2024

May 19, 2024
A faster, better way to prevent an AI chatbot from giving toxic responses | MIT News

A faster, better way to prevent an AI chatbot from giving toxic responses | MIT News

April 10, 2024
Part 1: ABAP RESTful Application Programming Model (RAP) – Introduction

Part 1: ABAP RESTful Application Programming Model (RAP) – Introduction

November 20, 2023
Saginaw HMI Enclosures and Suspension Arm Systems from AutomationDirect – Library.Automationdirect.com

Saginaw HMI Enclosures and Suspension Arm Systems from AutomationDirect – Library.Automationdirect.com

December 6, 2023
Can You Guess What Percentage Of Their Wealth The Rich Keep In Cash?

Can You Guess What Percentage Of Their Wealth The Rich Keep In Cash?

June 10, 2024
AI Compared: Which Assistant Is the Best?

AI Compared: Which Assistant Is the Best?

June 10, 2024
How insurance companies can use synthetic data to fight bias

How insurance companies can use synthetic data to fight bias

June 10, 2024
5 SLA metrics you should be monitoring

5 SLA metrics you should be monitoring

June 10, 2024
From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

June 10, 2024
UGRO Capital: Targeting to hit milestone of Rs 20,000 cr loan book in 8-10 quarters: Shachindra Nath

UGRO Capital: Targeting to hit milestone of Rs 20,000 cr loan book in 8-10 quarters: Shachindra Nath

June 10, 2024
Facebook Twitter LinkedIn Pinterest RSS
News PouroverAI

The latest news and updates about the AI Technology and Latest Tech Updates around the world... PouroverAI keeps you in the loop.

CATEGORIES

  • AI Technology
  • Automation
  • Blockchain
  • Business
  • Cloud & Programming
  • Data Science & ML
  • Digital Marketing
  • Front-Tech
  • Uncategorized

SITEMAP

  • Disclaimer
  • Privacy Policy
  • DMCA
  • Cookie Privacy Policy
  • Terms and Conditions
  • Contact us

Copyright © 2023 PouroverAI News.
PouroverAI News

No Result
View All Result
  • Home
  • AI Tech
  • Business
  • Blockchain
  • Data Science & ML
  • Cloud & Programming
  • Automation
  • Front-Tech
  • Marketing

Copyright © 2023 PouroverAI News.
PouroverAI News

Welcome Back!

Login to your account below

Forgotten Password? Sign Up

Create New Account!

Fill the forms bellow to register

All fields are required. Log In

Retrieve your password

Please enter your username or email address to reset your password.

Log In