Data processing: artificial intelligence – Machine learning
Reexamination Certificate
2001-06-04
2008-08-05
Starks, Jr., Wilbert L (Department: 2129)
Data processing: artificial intelligence
Machine learning
C706S045000
Reexamination Certificate
active
07409371
ABSTRACT:
A model is constructed for an initial subset of the data using a first parameter estimation algorithm. The model may be evaluated, for example, by applying the model to a holdout data set of the data. If the model is not acceptable, additional data is added to the data subset and the first parameter estimation algorithm is repeated for the aggregate data subset. An appropriate subset of the data exists when the first parameter estimation algorithm produces an acceptable model. The appropriate subset of the data may then be employed by a second parameter estimation algorithm, which may be a more accurate version of the first algorithm or a different algorithm altogether, to build a statistical model to characterize the data.
REFERENCES:
patent: 5140530 (1992-08-01), Guha et al.
patent: 6047277 (2000-04-01), Parry et al.
patent: 6272479 (2001-08-01), Farry et al.
patent: 6768982 (2004-07-01), Collins et al.
Microsoft Press Computer Dictionary: Third Edition, The Microsoft Press, 1997, p. 20.
Leung, Y., et al., “A New Method for Mining Regression Classes in Large Data Sets”,IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, No. 1, 2001, p. 5-21.
Ozelkan EC (Reprint), et al.; “Multi-Objective Fuzzy Regression: A General Framework”,Computers&Operations Research, vol. 27, No. 7-8, 2000, p. 635-652.
Yilmaz, A., et al.; “Input Data Analysis Using Neural Networks”,Simulation, vol. 74, No. 3, 2000, p. 128-137.
Meek, C., et al.; “The Learning Curve Method Applied to Clustering”,Technical Report, Jul. 2000, p. 1-4.
Meek, C., et al.; “The Learning Curve Method Applied to Clustering”,Technical Report MSR-TR-01-34, Feb. 2001, p. 1-19.
Peter Cheeseman and John Stutz, Bayesian Classification (AutoClass): Theory and Results, Advances in Knowledge Discovery and Data Mining, 1995, pp. 153-180, AAAI Press, Menlo Park, CA.
Ronald A. Howard, Decision Analysis: Applied Decision Theory, Proceedings of the Fourth International Conference on Operational Research, 1966, pp. 55-71, Wiley-Interscience.
George H. John and Pat Langley, Static Versus Dynamic Sampling for Data Mining, Proceedings of the Second International Conference on Knowledge Discovery and Data Mining, 1996, pp. 367-370, AAAI/MIT Press.
Carl Myers Kadie, Seer: Maximum Likelihood Regression For Learning-Speed Curves, Thesis, 1995, 104 pages, Department of Computer Science, University of Illinois, Urbana, IL.
D. Pearce, Cost Benefit Analysis, 1983, pp. 59-89, St. Martin's Press, New York.
Foster Provost, David Jensen, and Tim Oates, Efficient Progressive Sampling, Conference on Knowledge Discovery in Data, Proceedings of the Fifth ACM SIGKDD International Conference on Knowledge Discovery And Data Mining, 1999, pp. 23-32 pages, ACM, New York.
B. Thiesson, C. Meek, D.M. Chickering, and D. Heckerman, Computationally Efficient Methods for Selecting Among Mixtures of Graphical Models, Bayesian Statistics 6: Proceedings of Sixth Valcenia International Meeting, 1999, pp. 631-656, Clarendon Press, Oxford.
Heckerman David E.
Meek Christopher A.
Thiesson Bo
Amn, Turocy & Calvin, LLP
Microsoft Corporation
Starks, Jr. Wilbert L
LandOfFree
Efficient determination of sample size to facilitate... does not yet have a rating. At this time, there are no reviews or comments for this patent.
If you have personal experience with Efficient determination of sample size to facilitate..., we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Efficient determination of sample size to facilitate... will most certainly appreciate the feedback.
Profile ID: LFUS-PAI-O-3994148