DAC-SGD: A Distributed Stochastic Gradient Descent Algorithm Based on Asynchronous Connection

2017 
In the data mining practice, it happens that the algorithm used in mining tasks needs to deal with the multiple distributed data source, while the required datasets are located in different companies or organizations and reside in different system and technology environments. In traditional mining solutions or algorithms, data located in different source need to be copied and integrated into a homogenous computation environment, and then the mining can be executed, which leads to large data transmission and high storage costs. Even the data mining can be in feasible due to the data ownership problems. In this paper, a distributed asynchronous connection approach for the well-used stochastic gradient descent algorithm (SGD) was presented, and a distributed implementation for it was done to cope with the multiple distributed data source problems. In which, the main process of the algorithm was executed asynchronously in distributed computation node and the model can be trained locally in multiple data sources based on their own computation environment, so as to avoid the data integration and centralized processing. And the feasibility and performance for the proposed algorithm was evaluated based on experimental studies.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    17
    References
    0
    Citations
    NaN
    KQI
    []