Scaling up: Distributed machine learning with cooperation
- NYNEX Science & Technology, White Plains, NY (United States)
- Univ. of Pittsburgh, PA (United States)
Machine-learning methods are becoming increasingly popular for automated data analysis. However, standard methods do not scale up to massive scientific and business data sets without expensive hardware. This paper investigates a practical alternative for scaling up: the use of distributed processing to take advantage of the often dormant PCs and workstations available on local networks. Each workstation runs a common rule-learning program on a subset of the data. We first show that for commonly used rule-evaluation criteria, a simple form of cooperation can guarantee that a rule will look good to the set of cooperating learners if and only if it would look good to a single learner operating with the entire data set. We then show how such a system can further capitalize on different perspectives by sharing learned knowledge for significant reduction in search effort. We demonstrate the power of the method by learning from a massive data set taken from the domain of cellular fraud detection. Finally, we provide an overview of other methods for scaling up machine learning.
- OSTI ID:
- 430637
- Report Number(s):
- CONF-960876-; TRN: 96:006521-0012
- Resource Relation:
- Conference: 13. National conference on artifical intelligence and the 8. Innovative applications of artificial intelligence conference, Portland, OR (United States), 4-8 Aug 1996; Other Information: PBD: 1996; Related Information: Is Part Of Proceedings of the thirteenth national conference on artificial intelligence and the eighth innovative applications of artificial intelligence conference. Volume 1 and 2; PB: 1626 p.
- Country of Publication:
- United States
- Language:
- English
Similar Records
Machine Learning for Big Data: A Study to Understand Limits at Scale
Generalization in quantum machine learning from few training data