Exponential priors for maximum entropy models

Data processing: measuring – calibrating – or testing – Measurement system – Statistical measurement

Reexamination Certificate

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

C702S179000, C702S180000, C702S182000, C702S183000, C702S184000, C709S206000

Reexamination Certificate

active

10766348

ABSTRACT:
The subject invention provides for systems and methods that facilitate optimizing one or mores sets of training data by utilizing an Exponential distribution as the prior on one or more parameters in connection with a maximum entropy (maxent) model to mitigate overfitting. Maxent is also known as logistic regression. More specifically, the systems and methods can facilitate optimizing probabilities that are assigned to the training data for later use in machine learning processes, for example. In practice, training data can be assigned their respective weights and then a probability distribution can be assigned to those weights.

REFERENCES:
patent: 6125362 (2000-09-01), Elworthy
patent: 6161130 (2000-12-01), Horvitz et al.
patent: 6304841 (2001-10-01), Berger et al.
patent: 6553358 (2003-04-01), Horvitz
patent: 6606620 (2003-08-01), Sundaresan et al.
patent: 6609094 (2003-08-01), Basu et al.
patent: 6697769 (2004-02-01), Goodman et al.
patent: 2003/0105638 (2003-06-01), Taira
patent: 2003/0126102 (2003-07-01), Borthwick
patent: 2004/0260922 (2004-12-01), Goodman et al.
patent: 2005/0096907 (2005-05-01), Bacchiani et al.
T. Jaakkola, M. Meila, and T. Jebara. Maximum entropy discrimination. Advances in Neural Information Processing Systems 12. Cambridge, MA: MIT Press, 2000, pp. 470-477.
T. Pedersen, R. Bruce, and J. Wiebe. Sequential model selection for word sense disambiguation. Proceedings of the 1997 Conference on Applied Natural Language Processing. Washington, D.C., 1997, pp. 388-395.
R. Lau. Adaptive statistical language modelling. M.S. thesis, Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology, Cambridge, MA, 1994. 65 pages.
J. Goodman, Exponential Priors for Maximum Entropy Models, Technical Report, Microsoft Research, Jun. 2003, 14 pages.
J. Goodman, Exponential Priors for Maximum Entropy Models, North American ACL, 2004, 14 pages.
M. Banko and E. Brill. Mitigating the Paucity of Data Problem: Exploring the Effect of Training Corpus Size on Classifier Performance for NLP. In Proc. of the Conference on Human Language Technology, 2001. 5 pages.
A.L. Berger, et al. A Maximum Entropy Approach to Natural Language Processing. Computational Linguistics, 22(1): 39-71, 1996.
S.F. Chen and R. Rosenfeld. A Survey of Smoothing Techniques for ME Models. IEEE Transactions on Speech and Audio Processing, vol. 8 No. 1, Jan. 2000. 14 pages.
S. Della Pietra, et al. Inducing Features of Random Fields. IEEE Transactions on Pattern Analysis and Machine Intelligence, 19(4): 380-393, 1997.
I.J. Good. The Population Frequencies of Species and the Estimation of Population Parameters. Biometrika. vol. 40 No. ¾, pp. 237-264, 1953.
J. Goodman. Classes for Fast Maximum Entropy Training. In ICASSP 2001. 4 pages.
C.M. Kadie, et al. CFW: A Collaborative Filtering System using Posteriors over Weights of Evidence. In Proc. of UAI, pp. 242-250, 2002.
R. Kneser and H. Ney. Improved Backing-off for M-gram Language Modeling. In ICASSP, vol. 1, pp. 181-184, 1995.
W. Newman. An Extension to the Maximum Entropy Method. IEEE Transactions on Information Theory, vol. IT-23, No. 1, Jan. 1997. 5 pages.
J. Darroch and D. Ratcliff. Generalized Iterative Scaling for Log-linear Models. The Annals of Mathematical Statistics, 43: 1470-1480, 1972.
S.F. Chen and J. Goodman. An Empirical Study of Smoothing Techniques for Language Modeling. Computer Speech and Language, 13: 359-394, Oct. 1999.
A. Ratnaparkhi. Maximum Entropy Models for Natural Language Ambiguity Resolution. PhD Thesis, University of Pennsylvania, 1998. 163 pages.
J. Reynar and A. Ratnaparkhi. A Maximum Entropy Approach to Identifying Sentence Boundaries. In ANLP, 1997. 4 pages.
R. Rosenfeld. Adaptive Statistical Language Modeling: A Maximum Entropy Approach. PhD Thesis, Carnegie Mellon University, Apr. 1994. 114 pages.
S. Khudanpur. A Method of Maximum Entropy Estimation with Relaxed Constraints. In 1995 Johns Hopkins University Language Modeling Workshop, 1995. 18 pages.
P.M. Williams. Bayesian Regularization and Pruning using a Laplace Prior. Neural Computation, vol. 7, pp. 117-143, 1995.
Chen and Rosenfeld. A Gaussian Prior for Smoothing Maximum Entropy Models. Feb. 1999.

LandOfFree

Say what you really think

Search LandOfFree.com for the USA inventors and patents. Rate them and share your experience with other people.

Rating

Exponential priors for maximum entropy models does not yet have a rating. At this time, there are no reviews or comments for this patent.

If you have personal experience with Exponential priors for maximum entropy models, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Exponential priors for maximum entropy models will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFUS-PAI-O-3845235

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.