Principal component and multiple correspondence analysis for handling mixed variables in the smoothed location model

The issue of classifying objects into groups when the measured variables are mixtures of continuous and binary variables has attracted the attention of statisticians. Among the discriminant methods in classification, Smoothed Location Model (SLM) is used to handle data that contains both continuous...

Full description

Saved in:
Bibliographic Details
Main Author: Ngu, Penny Ai Huong
Format: Thesis
Language:eng
eng
Published: 2016
Subjects:
Online Access:https://etd.uum.edu.my/6034/1/s817094_01.pdf
https://etd.uum.edu.my/6034/2/s817094_02.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
id my-uum-etd.6034
record_format uketd_dc
institution Universiti Utara Malaysia
collection UUM ETD
language eng
eng
advisor Hamid, Hashibah
Aziz, Nazrina
topic QA299.6-433 Analysis
spellingShingle QA299.6-433 Analysis
Ngu, Penny Ai Huong
Principal component and multiple correspondence analysis for handling mixed variables in the smoothed location model
description The issue of classifying objects into groups when the measured variables are mixtures of continuous and binary variables has attracted the attention of statisticians. Among the discriminant methods in classification, Smoothed Location Model (SLM) is used to handle data that contains both continuous and binary variables simultaneously. However, this model is infeasible if the data is having a large number of binary variables. The presence of huge binary variables will create numerous multinomial cells that will later cause the occurrence of large number of empty cells. Past studies have shown that the occurrence of many empty cells affected the performance of the constructed smoothed location model. In order to overcome the problem of many empty cells due to large number of measured variables (mainly binary), this study proposes four new SLMs by combining the existing SLM with Principal Component Analysis (PCA) and four types of Multiple Correspondence Analysis (MCA). PCA is used to handle large continuous variables whereas MCA is used to deal with huge binary variables. The performance of the four proposed models, SLM+PCA+Indicator MCA, SLM+PCA+Burt MCA, SLM+PCA+Joint Correspondence Analysis (JCA), and SLM+PCA+Adjusted MCA are compared based on the misclassification rate. Results of a simulation study show that SLM+PCA+JCA model performs the best in all tested conditions since it successfully extracted the smallest amount of binary components and executed with the shortest computational time. Investigations on a real data set of full breast cancer also showed that this model produces the lowest misclassification rate. The next lowest misclassification rate is obtained by SLM+PCA+Adjusted MCA followed by SLM+PCA+Burt MCA and SLM+PCA+Indicator MCA models. Although SLM+PCA+Indicator MCA model gives the poorest performance but it is still better than a few existing classification methods. Overall, the developed smoothed location models can be considered as alternative methods for classification tasks in handling large number of mixed variables, mainly the binary.
format Thesis
qualification_name masters
qualification_level Master's degree
author Ngu, Penny Ai Huong
author_facet Ngu, Penny Ai Huong
author_sort Ngu, Penny Ai Huong
title Principal component and multiple correspondence analysis for handling mixed variables in the smoothed location model
title_short Principal component and multiple correspondence analysis for handling mixed variables in the smoothed location model
title_full Principal component and multiple correspondence analysis for handling mixed variables in the smoothed location model
title_fullStr Principal component and multiple correspondence analysis for handling mixed variables in the smoothed location model
title_full_unstemmed Principal component and multiple correspondence analysis for handling mixed variables in the smoothed location model
title_sort principal component and multiple correspondence analysis for handling mixed variables in the smoothed location model
granting_institution Universiti Utara Malaysia
granting_department Awang Had Salleh Graduate School of Arts & Sciences
publishDate 2016
url https://etd.uum.edu.my/6034/1/s817094_01.pdf
https://etd.uum.edu.my/6034/2/s817094_02.pdf
_version_ 1747828011314446336
spelling my-uum-etd.60342021-04-19T02:43:12Z Principal component and multiple correspondence analysis for handling mixed variables in the smoothed location model 2016 Ngu, Penny Ai Huong Hamid, Hashibah Aziz, Nazrina Awang Had Salleh Graduate School of Arts & Sciences Awang Had Salleh Graduate School of Arts and Sciences QA299.6-433 Analysis The issue of classifying objects into groups when the measured variables are mixtures of continuous and binary variables has attracted the attention of statisticians. Among the discriminant methods in classification, Smoothed Location Model (SLM) is used to handle data that contains both continuous and binary variables simultaneously. However, this model is infeasible if the data is having a large number of binary variables. The presence of huge binary variables will create numerous multinomial cells that will later cause the occurrence of large number of empty cells. Past studies have shown that the occurrence of many empty cells affected the performance of the constructed smoothed location model. In order to overcome the problem of many empty cells due to large number of measured variables (mainly binary), this study proposes four new SLMs by combining the existing SLM with Principal Component Analysis (PCA) and four types of Multiple Correspondence Analysis (MCA). PCA is used to handle large continuous variables whereas MCA is used to deal with huge binary variables. The performance of the four proposed models, SLM+PCA+Indicator MCA, SLM+PCA+Burt MCA, SLM+PCA+Joint Correspondence Analysis (JCA), and SLM+PCA+Adjusted MCA are compared based on the misclassification rate. Results of a simulation study show that SLM+PCA+JCA model performs the best in all tested conditions since it successfully extracted the smallest amount of binary components and executed with the shortest computational time. Investigations on a real data set of full breast cancer also showed that this model produces the lowest misclassification rate. The next lowest misclassification rate is obtained by SLM+PCA+Adjusted MCA followed by SLM+PCA+Burt MCA and SLM+PCA+Indicator MCA models. Although SLM+PCA+Indicator MCA model gives the poorest performance but it is still better than a few existing classification methods. Overall, the developed smoothed location models can be considered as alternative methods for classification tasks in handling large number of mixed variables, mainly the binary. 2016 Thesis https://etd.uum.edu.my/6034/ https://etd.uum.edu.my/6034/1/s817094_01.pdf text eng public https://etd.uum.edu.my/6034/2/s817094_02.pdf text eng public masters masters Universiti Utara Malaysia Abdi, H., & Valentin, D. (2007). Multiple Correspondence Analysis. Encyclopedia of Measurement and Statistics: USA: Sage. Adler, N. & Golany, B. (2002). Including Principal Component Weight to Improve Discrimination in Data Envelopment Analysis. Journal of the Operational Research Society, 53, 985–991. Afifi, A. A., & Elashoff, R. M. (1969). Multivariate Two Sample Tests with Dichotomous and Continuous Variables: The Location Model. The Annals of Mathematical Statistics, 40(1), 290–298. Akturk, D., Gun, S. & Kumuk, T. (2007). Multiple Correspondence Analysis Technique Used in Analyzing the Categorical Data in Social Sciences. Journal of Applied Sciences, 7(4), 585–588. Anderson, T. W. (1963). Asymptotic Theory for Principal Component Analysis. The Annals of Mathematical Statistics, 34(1), 122–148. Asparoukhov, O. & Krazanowski, W. J. (2000). Non-parametric Smoothing of the Location Model in Mixed Variables Discrimination. Statistics and Computing, 10(4), 289-297. Ayele, D., Zewotir, T. & Mwambi, H. (2013). Multiple Correspondence Analysis as a Tool for Analysis of Large Health Surveys in African Settings. African Health Sciences, 14(4), 1036–1045. Bakar, Z. A., Mohemad, R., Ahmad, A., & Deris, M. M. (2006). A Comparative Study for Outlier Detection Techniques in Data Mining. IEEE Conference on Cybernetics and Intelligent Systems, 1–6. Bar-Hen, A. (2002). ). Generalized Principal Component Analysis of Continuous and Discrete Variables. Interstat Statistics, 8(6), 1–26. Baxter, M. J. (1995). Standardization and Transformation in Principal Component Analysis, With Applications to Archaemotry. Journal of the Royal Statistical Society. Series C, 44(4), 513–527. Belhumeur, P. N., Hespanha, J. P. & Kriegman, D. J. (1997). Eigenfaces vs. Fisherfaces: Recognition using Class Specific Linear Projection. IEEE Transaction on Pattern Analysis and Machine Intelligence, 19(7), 711–720. Bellman, R. (1961). Adaptive Control Process: A Guided Tour. Princeton: Princeton University Press. Benzecri, J. P. (1992). Correspondence Analysis Handbook. New York, NY: Marcel Dekker, Inc. Bishop, C. M. (1995). Neural Network for Pattern Recognition. New York, NY: Oxford University Press. Bittencourt, H. R. & Clarke, R. T. (2003). Logistic Discriminant between Classes with Nearly Equal Spectral Response in High Dimensionality. In Proceedings of the IEEE International of Geoscience and Remote Sensing Symposiums, 6, 21-25 July 2003, Toulouse, France (pp.3748-3750). Piscataway, NJ: IEEE Operations Center. Breiman, L. (1996). Bagging Predictors. Machine Learning, 24, 123–140. Buhlmann. P. & Geer, S. (2011). Statistics for High Dimensional Data: Methods, Theory and Applications. Berlin, Heidelberg: Springer-Verlag. Buttrey, S. (1998). Nearest-neighbour classification with categorical variables. Computational Statistics and Data Analysis, 28, 157–169. Camiz, S. & Gomes, G. C. (2013). Joint Correspondence Analysis versus Multiple Correspondence Analysis: A Solution to an Undetected Problem. In A. Giusti, G. Ritter & M. Vichi (Eds), Classification and Data Mining: Studies in Classification, Data Analysis and Knowledge Organization. (pp. 11-19). Berlin, Heidelberg: Springer-Verlag. Cateni, S., Vannucci, M. & Colla, V. (2013). Multivariate Analysis in Management, Engineering and the Sciences. Classification and Ordination Methods as a Tool for Analyzing of Plant Communities. Retrieved from http://www.intechopen.com/books/multivariate-analysis-in-managementengineering-and-the- sciences/classification-and-ordination-methods-as-a-toolfor-analyzing-of-plant-communities#SEC6 Chanda, B. & Murthy, C. A. (2008). Advance in Intelligent Information Processing. Singapore: World Science Publishing. Chang, P. C. & Afifi, A. A. (1974). Classification Based on Dichotomous and Continuous Variables. Journal of the American Statistical Association, 69(346), 336–339. Choi, S. C. (1986). Discrimination and Classificaton: Overview. Computers and Mathematics with Applications, 12A(2), 173–177. Chou, Y., & Wang, W. (2010). Checking Dimensionality in Item Response Models With Principal Component Analysis on Standardized Residuals. Educational and Psychological Measurement, 70(5), 717–731. Costa, P. S., Santos, N. C., Cunha, P., Cotter, J., & Sousa, N. (2013). The use of multiple correspondence analysis to explore associations between categories of qualitative variables in healthy ageing. Journal of Aging Research, 1–12. Retrieved from http://dx.doi.org/10.1155/2013/302163 Cox, D. R. (1966). Some procedures associated with the logistic qualitative response curve. In Research Papers in Statistics: Festschrift for J. Neyman, Ed. F. N. David. New York, NY: Wiley. Cox, T. F., & Pearce, K. F. (1997). A robust logistic discrimination model. Statistics and Computing, 7(3), 155–161. D’Enza, A. I. & Greenacre, M. J. (2012). Multiple Correspondence Analysis for the Quantification and Visualization of Large Categorical Data Sets. In A. Di Ciaccio, M. Coli & J. M. A. Ibanez (Eds.), Advanced Statistical Methods for the Analysis of Large Data-Sets: Studies in Theoretical and Applied Statistics (pp. 453-463). Berlin, Heideberg: Springer-Verlag. Das, K. (2007). Feature Extraction and Classification for Large-scale Data Analysis (Unpublished doctoral dissertation). Department of Electrical and Computer Engineering, University of California. Das, K., Meyer, J. & Nenadic, Z (2006). Analysis of Large-Scale Brain Data for Brain-Computer Interfaces. In Proceeding of the 28th IEEE Annual International Conference of Engineering in Medicine and Biology Society, 30 August-3 September 2006, New York (pp. 5731-5734). Retrieved from http://cbmspc.eng.uci.edu/PUBLICATIONS/zn:06b.pdf. Daudin, J. J. (1986). Selection of Variables in Mixed-Variable Discriminant Analysis. Biometrics, 42(3), 473–481. Day, N. E. & Kerridge, D. F. (1967). A General Maximum Likelihood Discriminant. Biometrics, 23, 313–323. de Leeuw, J. (2006). Here’s Looking at Multivariable. In J. Blasius & M. J. Greenacre (Eds.), Visualization of Categorical Data (pp. 1-11). San Diego: Academic Press. Deng, H., Jin, L., Zhen, L., & Huang, J. (2005). A New Facial Expression Recognition Method on Local Gabor Filter Bank and PCA plus LDA. Journal of Information Technology, 11(11), 86–96. Desikachar, P. & Viswanathan, B. (2011). Patterns of Labour Market Insecurity in Rural India: A Multidimensional and Multivariate Analysis (No. Working Paper No. 62/2011). Madras School of Economics. Doey, L., & Kurta, J. (2011). Correspondence analysis applied to psychological research. Tutorials in Quantitative Methods for Psychology, 7(1), 5–14. Efron, B. (1983). Estimating the Error Rate of a Prediction Rule: Improvement on Cross Validation. Journal of the American Statistical Association, 78, 316–331. Fisher, R. A. (1936). The Use of Multiple Measurements in Taxonomic Problems. Annals of Eugenics, 7(1), 179–188. Fix, E. & Hodges, J. L. (1951). Discriminatory Analysis-Nonparametric Discrimination: Consistency Properties (No. Report No. 51-4). Retrieved from http://www.dtic.mil/dtic/tr/fulltext/u2/a800276.pdf Garcia, T., & Grande, I. (2003). A model for the valuation of farmland in Spain: The case for the use of multivariate analysis. Journal of Property Investment & Finance, 21(2), 136–153. Geisser, S. (1975). The Predictive Sample Reuse Method with Applications. Journal of American Statistical Association, 70, 320–328. Gervini, D. & Rousson, V. (2004). Criteria for Evaluating Dimension-Reducing Components for Multivariate Data. The American Statistician, 58(1), 72–76. Ghosh, A. & Barman, S. (2013). Prediction of Prostate Cancer Cells Based on Principal Component Analysis Techinque. Procedia Technology, 10, 37–44. Ghosh, A. (2011). Forecasting BSE Sensex under Optimal Conditions: An Investigation Post Factor Analysis. Journal of Business Studies Quarterly, 3(2), 52–73. Giri, N. C. (2004). Multivariate Statistical Analysis. New York, NY: Marcel Dekker Inc. Glynn, D. (2012). Correspondence Analysis: Exploring Data and Identifying Patterns. In. D. Glynn & J. Robinson (Eds), Polysemy and Synonymy: Corpus Methods and Applications in Cognitive Linguistics (pp. 133-179). Amsterdam: John Benjamins. Gnanadesikan, R., Roger, K., Breiman, L., Dunn, O. J., Friedman, J. H., Fu, K. S., Hartigan, J. A., Kettenring, J. R., Lachenbruch, P. A., Olshen, R. A. & Rohlf, F. J. (1989). Discriminant analysis and clustering. Panel on Discriminant Analysis, Classification and Clustering. Statistical Science, 4(1), 34–69. Gnanadesikan, R. (1977). Methods for Statistical Data Analysis of Multivariate Observations. New York, NY: John Wiley & Sons, Inc. Gottumukkal, R. & Asari, K. (2004). An Improved Face Recognition Technique Based on Modular PCA Approach. Patern Recognit Lett, 25, 429–436. Green, P. E., Krieger, A. M. & Carroll, J. D. (1987). Multidimensional Scaling: S Complementary Approach. Journal of Advertisement Research, 21–27. Greenacre, M. J. & Blasius, J. (2006). Multiple Correspondence Analysis and Related Methods. London: Taylor and Francis Group. Greenacre, M. (1984). Theory and Applications of Correspondence Analysis. London: Academic Press. Greenacre, M. (2007). Correspondence Analysis in Practice (2nd ed.). Boco Raton: Chapman & Hall. Greenacre, M. J. (2006). Typing Up the Loose Ends in Simple, Multiple and Joint Correspondence Analysis. In A. Rizzi & M. Vichi (Eds.),. Berlin, Heidelberg: Physica-Verlag. Griebel, m. & Hullmann, A. (2013). Dimensionality Reduction of High Dimensional Data with a Non-Linear Principal Component Aligned Generative Topographic Mapping. Institute for Numerical Simulation. Guttmann, L. (1941). The Quantification of a Class of Attributes: A Theory and Method of Scale Construction. In P. Horst, P. Wallin & L. Gutmann (Eds.). New York, NY: Social Science Research Council. Hamid, H. & Mahat, N. I. (2013). Using Principal Component Analysis to extract mixed variables for smoothed location model. Far East Journal of Mathematical Sciences (FJMS), 80(1), 33–54. Hamid, H. (2014). Integrated Smoothed Location Model and Data Reduction Approaches for Multi Variables Classification. Doctor of Philosophy. Universiti Utara Malaysia. Hand, D. J. (1981). Discrimination and Classification. Chichester: John Wiley & Sons. Hauser, R. P., & Booth, D. (2011). Predicting Bankruptcy with Robust Logistic Regression. Journal of Data Science, 9, 565–584. Hinneburg, A., & Keim, D. A. (1999). Optimal grid-clustering: Towards breaking the curse of dimensionality in high-dimensional clustering. International Conference on Very Large Database, 1, 506–517. Hirschfeld, H. O. (1935). A Connection Between Correlation and Contingency. Proceedings of the Cambridge Philosophical Society, 31, 520–524. Hirsh, O., Bosner, S., Hullermeier, E., Senge, R., Dembczynski, K. & D.-, & Banzhoff, N. (2011). Multivariate Modeling to Identify Patterns in Clinical Data: The Example of Chest Pain. BMC Medical Research Methodology, 11, 155–164. Hoffman, D. L., & Batra, R. (1991). Viewer Response to Programs: Dimensionality and Concurrent Behavior. Journal of Advertising Research, 23, 45–46. Hoffman, D. L., & Franke, G. R. (1986). Correspondence Analysis: Graphical Representation of Categorical Data in Marketing Research. Journal of Marketing Research, 23, 213–227. Hoffman, E. (1999). Standard Statistical Classification: Basic Principles. Statistical Commission, 8, 1–30. Holden, J. E., Finch, W. H., & Kelley, K. (2011). A comparison of two-group classification methods. Educational and Psychological Measurement, 71(5), 870–901. Hotelling, H. (1933). Analysis of a Complex of Statistical Variables into Principal Components. Journal of Educational Psychology, 24, 417–441. Hunter, E. J. (2009). Classification Made Simple: An Introduction to Knowledge Organisation and Informative Retrieval (3rd ed). England: Ashgate Publishing Company. Hyvarinen, A. & Oja, E. (2000). ndependent Component Analysis: Algorithms and Applications. Neural Network, 13(4-5), 411–430. Jackson, D. A. (1993). Stopping Rules in Principal Components Analysis : A Comparison of Heuristical and Statistical Approaches. Ecology, 74(8), 2204– 2214. Jackson, J. E. (1991). Stopping Rules in Principal Components Analysis: A Camparison Heuristical and Statistical Approach. Ecology, 74(8), 2204–2214. Jeffers, J. N. R. (1967). Two Case Studies in the Application of Principal Component Analysis. Journal of the Royal Statistical Society, 16(3), 225–236. Jolliffe, I. T. (1986). Principal Component Analysis. New York, NY: Springer-Verlag. Jolliffe, I. T. (2002). Principal Component Analysis (2nd ed.). New York, NY: Springer-Verlag. Jombart, T., Devillard, S., & Balloux, F. (2010). Discriminant analysis of principal components: a new method for the analysis of genetically structured populations. BMC Genetics, 11(94), 1–15. Kaiser, H. F. (1961). A note on Gutman’s Lower Bound for the Number of Common Factors. British Journal of Mathematical and Statistical Psycology, 14, 1–2. Kant, I. (1968). The Critique of Pure Reason. New York, NY: Hackett Publishing Co. Kemsley, E. K. (1996). Discriminant Analysis of High-Dimensional Data: A Comparison of Principal Component Analysis and Partial Least Squares Data Reduction Methods. Chemometrics and Intelligent Systems, 33, 47–61. Knoke, J. D. (1982). Discriminant Analysis with Discrete and Continuous Variables. Biometrics, 38(1), 191–200. Kolenikov, S., & Angeles, G. (2009). Socioeconomic Status Measurement With Discrete Proxy Variables: Is Principal Component Analysis a Reliable Answer? Review of Income and Wealth, 55(1), 128–165. Krzanowski, W. J. (1975). Discrimination and Classification Using both Binary and Continuous Variables. Journal of American Statistical Association, 70, 782–790. Krzanowski, W. J. (1977). The Performance of Fisher’s Linear Discriminant Function under Non-optimal Conditions. Technometrics, 19, 191–200. Krzanowski, W. J. (1979). Some Linear Transformation for Mixtures of Binary and Continuous Variables with Particular Reference to Linear Discriminant Analysis. Biometrika, 66(1), 33–39. Krzanowski, W. J. (1980). Mixtures of Continuous and Categorical Variables in Discriminant Analysis. Biometrics, 36, 493-499. Krzanowski, W. J. (1982). Mixtures of Continuous and Categorical Variables in Discriminant Analysis: A Hypothesis- Testing Approach. Biometric, 38, 991–1002. Krzanowski, W. J. (1983a). Stepwise Location Model Choice in Mixed Variables in Discriminant Analysis. Applied Statistics, 32(3), 260–266. Krzanowski, W. J. (1993). The Location Model for Mixtures of Categorical and Continuous Variables. Journal of Classification, 10, 25–49. Krzanowski, W. J. (1995). Selection of Variables, and Assessment of Their Performance, in Mixed Variable Discriminant Analysis. Computational Statistics and Data Analysis, 19, 419-431. Kuo, C.F, Syu, S. S, Lin, C. H. & Peng, K. C. (2012). Application of Principal Component Analysis and Gray Relational Method in the Optimization of the Met Spinning Process Using the Cooling Air System. Textile Research Journal, 83(4), 371–380. Lachenbruch, P. A., & Mickey, M. R. (1968). Estimation of Error Rates in Discriminant Analysis. Technometrics, 10, 1–11. Lachenbruch, P. A., Sneeringer, C., & Revo, L. T. (1975). Robustness of the Linear and Quadratic Discrimination Function to Certain Types of Nonnormality. Communications in Statistics, 1, 39–56. Lachenbruch, P. A. (1975). Discriminant Analysis. New York, NY: Hafner. Lee, S., Zou, F., & Wright, F. A. (2010). Convergence and Prediction of Principal Component Scores in High-Dimensional Settings. Annals of Statistics, 38(6), 3605–3629. Leon, A. R., Soo, A. & Williamson, T. (2011). Classification with Discrete and Continuous Variables via General Mixed-Data Models. Journal of Applied Statistics, 38(5), 1021–1032. Li, X. & Xu, R. (2009). High Dimensional Data Analysis in Cancer Research. New York, NY: Springer-Verlag. Li, Q. (2006). An Integrated Framework of Feature Selection and Extraction for Appearance-based Recognition (Unpublished doctoral dissertation). University of Delaware Newark, DE, USA. Loslever, P. (2009). Using Multiple Correspondence Analysis with Membership Values when the System Study Yields Miscellaneous Datasets. Cybernetics and Systems, 40(7), 633–652. Mahat, N. I., Krzanowski, W. J., & Hernandez, A. (2007). Variable selection in discriminant analysis based on the location model for mixed variables. Advances in Data Analysis and Classification, 1, 105–122. Mahat, N. I., Krzanowski, W. J., & Hernandez, A. (2009). Strategies for Non- Parametric Smoothing of the Location Model in Mixed-Variable Discriminant Analysis. Modern Applied Science, 3(1), 151-163. Massey, W. F. (1965). Principal Components Regression in Exploratory Statistical Research. Journal of American Statistical Association, 60, 234–246. McCulloch, W. S. & Pitts, W. (1943). A Logical Calculus of the Ideas Immanent in Nervous Activity. Bulletin of Mathematical Biophysics, 5(4), 115–133. McLachlan, G. J. (1992). Discriminant Analysis and Statistical Pattern Recognition. New York, NY: John Wiley & Sons, Inc. Messaoud R. B., Boussaid, O. & Rabaseda, S. L. (2007). A Multiple Correspondence Analysis to Organize Data Cubes. In O. Vasilecas, J. Eder & A. Caplinskas (Eds.). In Databases and Information System IV: Frontier in Artificial Intelligence and Applications (pp. 133–146). Amsterdam: IOS Press. Meulman, J.J., van Der Kooji, A. J. & Heiser, W. J. (2004). Principal Component Analysis with Nonlinear Optimal Scaling Transformation for Ordinal and Nominal data. In D. Kaplan (Eds.), The SAGE Handbook of Quantitative Methodology for the Social Sciences. Thousand Oaks: Sage. Mitchell, T. M. (1997). Machine Learning. New York, NY: McGraw-Hill. Nenadic, O. & Greenacre, M. (2007). Correspondence Analysis in R, with Two and Three Dimensional Graphics: The ca Package. Journal of Statistical Software, 20(3), 1–13. Olkin, I. & Tate, R. F. (1961). Multivariate Correlation Models with Discrete and Continuous Variables. The Annals of Mathematical Statistics, 32, 448–465. Panea, B., Casasús, I., Blanco, M., & Joy, M. (2009). The use of correspondence analysis in the study of beef quality: a case study. Spanish Journal of Agricultural Research, 7(4), 876–885. Pearson, K. (1901). On Lines and Planes of Closest Fit to Systems of Points in Space. Philosophical Magazine, 6, 559–572. Peter, G. M., Joop, T. & Charles, O. (1997). Multiple Correspondence Analysis as A Tool for Quantification or Classification of Career Data. Journal of Educational and Behavioral Statistics, 22(4), 447–477. Pyryt, M. C. (2004). Pegnato Revisited : Using Discriminant Analysis to Identify Gifted Children. Psychology Science, 46(3), 342–347. Quinn, G. P. & Keough, M. J. (2002). Experimental Design and Data Analysis for Biologists. New York, NY: Cambridge University Press. Rao, C. R. (1964). The Use and Interpretation of Principal Component Analysis in Applied Research. Sankhya A, 26, 329–358. Reise, S. P., Waller, N. G. & Comrey, A. L. (2000). Factor Analysis and Scale Revision. Psychological Assessment, 12, 287-297. Rencher, A. C. (2002). Methods of Multivariate Analysis: Wiley Series in Probability and Statistics (2nd ed.). New York, NY: John Wiley & Sons, Inc. Saporta, G., & Niang, N. (2006). ). Correspondence Analysis and Classification. In: Michael Greenacre & Jorg Blasius (eds.) Multiple Correspondence Analysis and Related Methods. Boca Raton: Chapman & Hall/CRC. Schürks, M., Buring, J. E., & Kurth, T. (2011). Migraine features, associated symptoms and triggers: a principal component analysis in the Women’s Health Study. International Headache Society, 31(7), 861–869. Simon, R., Radmacher, M. D., Dobbin, K. & McShane, L. M. (2003). Pitfalls in the Use of DNA Microarray Data for Diagnostic and Prognostic Classification. Journal of National Cancer Institute, 95(1), 14–18. Siswadi, Muslim, A. & Bakhtiar, T. (2012). Variable Selection Using Principal Component and Procrustes Analyses and its Application in Educational Data. Journal of Asian Scientific Research, 2(12), 856–865. Smith, C. A. B. (1947). Some Examples of Discrimination. Annals of Eugenics1, 18, 272–283. Soni, S. & Shrivastave, S. (2010). Classification of Indian Stock Market Data Using Machine Learning Algorithms. International Journal on Computer Science and Engineering, 2(9), 2942–2946. Stern, H. S. (1996). Neural Networks in Applied Statistics. Technometrics, 38, 217–225. Taniguchi, M., Hirukawa, J. & Tamaki, K. (2008). Optimal Statistical Inference in Financial Engineering. Chapman: Boca Raton. Tenenhaus, M. & Young, F. W. (1985). An Analysis and Synthesis of Multiple Correspondence Analysis, Optimal Scaling, Dual Scaling, Homogeneity Analysis and Other Methods for Quantifying Categorical Multivariate Data. Psychometrika, 50(1), 91–119. Tian, Y., Guo, P. & Lyu, M. R. (2005). Comparative Studies on Feature Extraction Methods for Multispectral Remote Sensing Image Classification. In In Proceedings of the International Conference on Systems, Man and Cybernetics (pp. 1275–1279). Berlin, Heidelberg: Springer-Verlag. Turk, M. A. & Pentland, A. P. (1991). Eigenfaces for Recognition. Journal of Cognitive Neuroscience, 3(1), 71–86. Vapnik, V. (1995). The Nature of Statistical Learning Theory. New York, NY: Springer-Verlag. Veer, L. J., Dai, H., van de Vijver, M. J., He, Y. D., Hart, A. A. M., Mao, M., … Friend, S. H. (2002). Gene expression profiling predicts clinical outcome of breast cancer. Nature, 415(6871), 530–536. Vlachonikolis, I. G., & Marriott, F. H. C. (1982). Discrimination with mixed binary and continuous data. Journal of the Royal Statistical Society, Series C (Applied Statistics), 31(1), 23–31. Wahl, P. W. & Kronmal, R. A. (1977). Discriminant Functions when Covariances are Unequal and Sample Sizes are Moderate. Biometrics, 33, 479–484. Wakaki, H. (1990). Comparison of Linear and Quadratic Discriminant Funcitons. Biometrika, 77, 227–229. Wang, X. & Tang, X. (2004). Experimental study on multiple LDA classifier combination for high dimensional data classification. In Proceedings of the 5th International Workshop on Multiple Classifier Systems (Vol. 3077, pp. 344–353). Cagliari, Italy: Springer-Verlag. William, B. K. (1983). Some Observations on the Use of Discrimnant Analysis in Ecology. Ecology Society of America, 64(8), 1283–1291. Wold, H. (1966). Nonlinear Estimation by Iterative Least Squares Procedures. In F. N. David (Eds.). In Research Papers in Statistics (pp. 411–444). New York, NY: John Wiley & Sons, Inc. Wold, S., Esbensen, K., & Geladi, P. (1987). Principal component analysis. Chemometrics and Intelligent Laboratory Systems, 2(1-3), 37–52. Yang, C., Peng, H. & Wang, J. (2008). A New Feature Extraction Approach Based on Sentences Element Analysis. In Proceedings of the International Conference on Computational Intelligence and Security, 1, 13-17 December 2008 (pp. 90-95). Washington, DC: IEEE Computer Society Press. Yanqin, T., & Ping, G. (2005). Comparative Studies on Dimension Reduction Methods for Multispectral Remote Sensing Image. In 5 IEEE International Conference on Systems, Man and Cybernetics (pp. 1275–1279). Berlin, Heidelberg: Springer-Verlag. Young, P. D. (2009). Dimension reduction and missing data in statistical discrimination. (Doctoral dissertation, USA Baylor University). Retrieved from http://beardocs.baylor.edu/xmlui/bitstream/handle/ 2104/5543/philip_young_phd.pdf. Yu, L., & Liu, H. (2003). Feature Selection for High-Dimensional Data: A Fast Correlation-Based Filter Solution. In International Conference on Machine Learning (ICML) (pp. 856–863). Washington, DC: Arizona State University. Zhou, X. S. & Huang, T. S. (2001). Small sample learning during multimedia retrieval using BiasMap. Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. CVPR 2001, 1, 11–17. Zhu, M. (2001). Feature extraction and dimension reduction with applications to classification and the analysis of co-occurrence data. (Doctoral dissertation, Stanford University). Retrieved from http://search.proquest.com.ezproxy.cul.columbia. edu/docview/304728120/abstract/B59F30759A1641ABPQ/ 1?accountid=10226