Statistics 598M Final Project

The final project is a paper presentation. Every group is required to read a paper, understand the theory, methodology and examples in the paper, and present it to the class. You also need to hand in a brief summary of the paper with your comments after presentation. A list of the possible topics and papers is give below. You can choose a paper from the list, and email me your choice. If two groups have chosen the same paper, it will be given to the group that has sent me the choice first. You can also choose a topic or a paper outside this list, however, you need send me the paper and get my approval to use it for the project.

The presentation is 25 minutes. Two groups will be scheduled on Friday, May 2, in class, and the rest four groups on Tuesday, May 6, 10:20-12:20pm.

Topics and Papers

Dimension Reduction and Regression

  • Zhenyue Zhang and Hongyuan Zha Principal Manifolds and Nonlinear Dimension Reduction via Local Tangent Space Alignment (2002)
  • Chris Ding, Xiaofeng He, Hongyuan Zha, Horst Simon, Adaptive dimension reduction for clustering high dimensional data
  • Sanjeev Arora, Ravi Kannan Learning mixtures of arbitrary gaussians (2001)
  • Avner Magen Dimensionality Reductions that Preserve Volumes and Distance to Affine Spaces, and their Algorithmic Applications
  • Bura, E. and Cook, R. D. (2001). Estimating the structural dimension of regressions via parametric inverse regression, Journal of the Royal Statistical Society 63, 393-410.
  • Yingcun Xia, Howell Tong, W. K. Li, Li-Xing Zhu An adaptive estimation of dimension reduction space,
  • JRSSB,(2002)
  • Ker-Chau Li Nonlinear Confounding in High-Dimensional RegressionAnnals,(1997)
  • Bayesian Trees

  • Chipman, H., George, E., and McCulloch, R. (1998) ``Bayesian CART Model Search (with discussion)'', Journal of the American Statistical Association, 93, 935--960
  • Hugh A. Chipman, Edward I. George, Robert E. McCulloch Bayesian Treed Models, Machine Learning,2002
  • Spline

  • Charles Kooperberg; Smarajit Bose; Charles J. Stone Polychotomous Regression (in Theory and Methods),Journal of the American Statistical Association, Vol. 92, No. 437. (Mar., 1997), pp. 117-127.
  • Charles J. Stone; Mark H. Hansen; Charles Kooperberg; Young K. Truong Polynomial Splines and their Tensor Products in Extended Linear Modeling. Annals of Statistics, Vol. 25, No. 4. (Aug., 1997), pp. 1371-1425
  • Hansen and Kooperberg Spline Adaptation in Extended Linear Models, Statistical Science, 17, 20-21
  • Boosting

  • Y. Freund. An adaptive version of the boost by majority algorithm. Machine Learning, 43(3):293-318, June 2001
  • G. Ratsch, A. Demiriz, and K. Bennett Sparse regression ensembles in infinite and finite hypothesis spaces. Machine Learning, 48(1-3):193-221, 2002
  • H. Drucker Effect of pruning and early stopping stopping on performance of a boosted ensemble. In Proceedings of the International Meeting on Nonlinear Methods and Data Mining, pages 26-40, Rome, Italy, 2000
  • R.E. Schapire, Y. Freund, P. Bartlett, and W. S. Lee. Boosting the margin: A new explanation for the effectiveness of voting methods. The Annals of Statistics, 26(5):1651-1686, October 1998
  • Support Vector Machine

  • Lee, Y., Lin, Y. and Wahba, G. (2002) Multicategory Support Vector Machines, Theory, and Application to the Classification of Microarray Data and Satellite Radiance Data. Revision invited. Journal of American Statistical Association
  • O. L. Mangasarian and D. R. Musicant. Data discrimination via nonlinear generalized support vector machines. Technical Report 99-03, University of Wisconsin, Computer Sciences Department, Madison, WI, USA, 1999
  • B. Scholkopf, J. Platt, J. Shawe-Taylor, A. J. Smola, and R. C. Williamson. Estimating the support of a high-dimensional distribution. Technical Report 99-87, Microsoft Research, 1999. To appear in Neural Computation, 2001.
  • Y. Lin, G. Wahba, H. Zhang, and Y. Lee. Statistical properties and adaptive tuning of support vector machines. Department of statistics technical report 1022, University of Wisconsin-Madison, 2000.
  • Bernhard Scholkopf, Alex J. Smola, Robert C. Williamson, and Peter L. Bartlett New Support Vector Algorithms Neural Comp. 2000 12: 1207-1245