Enhancing Distributed Machine Learning through Data Shuffling: Techniques, Challenges, and Implications

In distributed machine learning, data shuffling is a crucial data preprocessing technique that significantly impacts the efficiency and performance of model training. As distributed machine learning scales across multiple computing nodes, the ability to shuffle data effectively and efficiently has b...

Full description

Saved in:
Bibliographic Details
Main Author: Zhang Zikai
Format: Article
Language:English
Published: EDP Sciences 2025-01-01
Series:ITM Web of Conferences
Online Access:https://www.itm-conferences.org/articles/itmconf/pdf/2025/04/itmconf_iwadi2024_03018.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!