Efficient determination of sample size to facilitate...

Data processing: artificial intelligence – Machine learning

Reexamination Certificate

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

C706S045000

Reexamination Certificate

active

07409371

ABSTRACT:
A model is constructed for an initial subset of the data using a first parameter estimation algorithm. The model may be evaluated, for example, by applying the model to a holdout data set of the data. If the model is not acceptable, additional data is added to the data subset and the first parameter estimation algorithm is repeated for the aggregate data subset. An appropriate subset of the data exists when the first parameter estimation algorithm produces an acceptable model. The appropriate subset of the data may then be employed by a second parameter estimation algorithm, which may be a more accurate version of the first algorithm or a different algorithm altogether, to build a statistical model to characterize the data.

REFERENCES:
patent: 5140530 (1992-08-01), Guha et al.
patent: 6047277 (2000-04-01), Parry et al.
patent: 6272479 (2001-08-01), Farry et al.
patent: 6768982 (2004-07-01), Collins et al.
Microsoft Press Computer Dictionary: Third Edition, The Microsoft Press, 1997, p. 20.
Leung, Y., et al., “A New Method for Mining Regression Classes in Large Data Sets”,IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, No. 1, 2001, p. 5-21.
Ozelkan EC (Reprint), et al.; “Multi-Objective Fuzzy Regression: A General Framework”,Computers&Operations Research, vol. 27, No. 7-8, 2000, p. 635-652.
Yilmaz, A., et al.; “Input Data Analysis Using Neural Networks”,Simulation, vol. 74, No. 3, 2000, p. 128-137.
Meek, C., et al.; “The Learning Curve Method Applied to Clustering”,Technical Report, Jul. 2000, p. 1-4.
Meek, C., et al.; “The Learning Curve Method Applied to Clustering”,Technical Report MSR-TR-01-34, Feb. 2001, p. 1-19.
Peter Cheeseman and John Stutz, Bayesian Classification (AutoClass): Theory and Results, Advances in Knowledge Discovery and Data Mining, 1995, pp. 153-180, AAAI Press, Menlo Park, CA.
Ronald A. Howard, Decision Analysis: Applied Decision Theory, Proceedings of the Fourth International Conference on Operational Research, 1966, pp. 55-71, Wiley-Interscience.
George H. John and Pat Langley, Static Versus Dynamic Sampling for Data Mining, Proceedings of the Second International Conference on Knowledge Discovery and Data Mining, 1996, pp. 367-370, AAAI/MIT Press.
Carl Myers Kadie, Seer: Maximum Likelihood Regression For Learning-Speed Curves, Thesis, 1995, 104 pages, Department of Computer Science, University of Illinois, Urbana, IL.
D. Pearce, Cost Benefit Analysis, 1983, pp. 59-89, St. Martin's Press, New York.
Foster Provost, David Jensen, and Tim Oates, Efficient Progressive Sampling, Conference on Knowledge Discovery in Data, Proceedings of the Fifth ACM SIGKDD International Conference on Knowledge Discovery And Data Mining, 1999, pp. 23-32 pages, ACM, New York.
B. Thiesson, C. Meek, D.M. Chickering, and D. Heckerman, Computationally Efficient Methods for Selecting Among Mixtures of Graphical Models, Bayesian Statistics 6: Proceedings of Sixth Valcenia International Meeting, 1999, pp. 631-656, Clarendon Press, Oxford.

LandOfFree

Say what you really think

Search LandOfFree.com for the USA inventors and patents. Rate them and share your experience with other people.

Rating

Efficient determination of sample size to facilitate... does not yet have a rating. At this time, there are no reviews or comments for this patent.

If you have personal experience with Efficient determination of sample size to facilitate..., we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Efficient determination of sample size to facilitate... will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFUS-PAI-O-3994148

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.