Learning Graphical Models from a Distributed Stream
This topic contains 0 replies, has 1 voice, and was last updated by arXiv 1 year, 3 months ago.

Learning Graphical Models from a Distributed Stream
A current challenge for data management systems is to support the construction and maintenance of machine learning models over data that is large, multidimensional, and evolving. While systems that could support these tasks are emerging, the need to scale to distributed, streaming data requires new models and algorithms. In this setting, as well as computational scalability and model accuracy, we also need to minimize the amount of communication between distributed processors, which is the chief component of latency. We study Bayesian networks, the workhorse of graphical models, and present a communicationefficient method for continuously learning and maintaining a Bayesian network model over data that is arriving as a distributed stream partitioned across multiple processors. We show a strategy for maintaining model parameters that leads to an exponential reduction in communication when compared with baseline approaches to maintain the exact MLE (maximum likelihood estimation). Meanwhile, our strategy provides similar prediction errors for the target distribution and for classification tasks.
Learning Graphical Models from a Distributed Stream
by Yu Zhang, Srikanta Tirthapura, Graham Cormode
https://arxiv.org/pdf/1710.02103v1.pdf
You must be logged in to reply to this topic.