This book provides a practical and fairly comprehensive review of Data Science through the lens of dimensionality reduction, as well as hands-on techniques to tackle problems with data collected in the real world. State-of-the-art results and solutions from statistics, computer science and mathematics are explained from the point of view of a practitioner in any domain science, such as biology, cyber security, chemistry, sports science and many others. Quantitative and qualitative assessment methods are described to implement and validate the solutions back in the real world where the problems originated.
The ability to generate, gather and store volumes of data in the order of tera- and exo bytes daily has far outpaced our ability to derive useful information with available computational resources for many domains.
This book focuses on data science and problem definition, data cleansing, feature selection and extraction, statistical, geometric, information-theoretic, biomolecular and machine learning methods for dimensionality reduction of big datasets and problem solving, as well as a comparative assessment of solutions in a real-world setting.
This book targets professionals working within related fields with an undergraduate degree in any science area, particularly quantitative. Readers should be able to follow examples in this book that introduce each method or technique. These motivating examples are followed by precise definitions of the technical concepts required and presentation of the results in general situations. These concepts require a degree of abstraction that can be followed by re-interpreting concepts like in the original example(s). Finally, each section closes with solutions to the original problem(s) afforded by these techniques, perhaps in various ways to compare and contrast dis/advantages to other solutions.
2.1.1 Linear Multiple Regression Model: Continuous Response
2.1.2 Logistic Regression: Categorical Response
2.1.3 Variable Selection and Model Building
2.1.4 Generalized Linear Model (GLM)
2.1.5 Decision Trees
2.1.6 Bayesian Learning
2.2 Machine Learning Solutions: Supervised
2.2.1 k-Nearest Neighbors (kNN)
2.2.2 Ensemble Methods
2.2.3 Support Vector Machines (SVMs)
2.2.4 Neural Networks (NNs)
2.3 Machine Learning Solutions: Unsupervised
2.3.1 Hard Clustering
2.3.2 Soft Clustering
2.4 Controls, Evaluation and Assessment
2.4.1 Evaluation Methods
2.4.2 Metrics for Assessment
3. What is Dimensionality Reduction (DR)?
3.1 Dimensionality Reduction
3.2 Major Approaches to Dimensionality Reduction
3.2.1 Conventional Statistical Approaches
3.2.2 Geometric Approaches
3.2.3 Information-theoretic Approaches
3.2.4 Molecular Computing Approaches
3.3 The Blessings of Dimensionality
4. Conventional Statistical Approaches
4.1 Principal Component Analysis (PCA)
4.1.1 Obtaining the Principal Components
4.1.2 Singular value decomposition (SVD)
4.2 Nonlinear PCA
4.2.1 Kernel PCA
4.2.2 Independent component analysis (ICA)
4.3 Nonnegative Matrix Factorization (NMF)
4.3.1 Approximate Solutions
4.3.2 Clustering and Other Applications
4.4 Discriminant Analysis
4.4.1 Linear discriminant analysis (LDA)
4.4.2 Quadratic discriminant analysis (QDA)
4.5 Sliced Inverse Regression (SIR)
5. Geometric Approaches
5.1 Introduction to Manifolds
5.2 Manifold Learning Methods
5.2.1 Multi-Dimensional Scaling (MDS)
5.2.2 Isometric Mapping (ISOMAP)
5.2.3 t-Stochastic Neighbor Embedding ( t-SNE )
5.3 Exploiting Randomness (RND)
6. Information-theoretic Approaches
6.1 Shannon Entropy (H)
6.2 Reduction by Conditional Entropy
6.3 Reduction by Iterated Conditional Entropy
6.4 Reduction by Conditional Entropy on Targets
6.5 Other Variations
7. Molecular Computing Approaches
7.1 Encoding Abiotic Data into DNA
7.2 Deep Structure of DNA Spaces
7.2.1 Structural Properties of DNA Spaces
7.2.2 Noncrosshybridizing (nxh) Bases
7.3 Reduction by Genomic Signatures
7.3.1 Background
7.3.2 Genomic Signatures
7.4 Reduction by Pmeric Signatures
8. Statistical Learning Approaches
8.1 Reduction by Multiple Regression
8.2 Reduction by Ridge Regression
8.3 Reduction by Lasso Regression
8.4 Selection versus Shrinkage
8.5 Further refinements
9. Machine Learning Approaches
9.1 Autoassociative Feature Encoders
9.1.1 Undercomplete Autoencoders
9.1.2 Sparse Autoencoders
9.1.3 Variational Autoencoders
9.1.4 Dimensionality Reduction in MNIST Images
9.2 Neural Feature Selection
9.2.1 Facial Features, Expressions and Displays
9.2.2 The Cohn-Kanade Dataset
9.2.3 Primary and Derived Features
9.3 Other Methods
10. Metaheuristics of DR Methods
10.1 Exploiting Feature Grouping
10.2 Exploiting Domain Knowledge
10.2.1 What is Domain Knowledge?
10.2.2 Domain Knowledge for Dimensionality Reduction
10.3 Heuristic Rules for Feature Selection, Extraction and Number
10.4 About Explainability of Solutions
10.4.1 What is Explainability?
10.4.2 Explainability in Dimensionality Reduction
10.5 Choosing Wisely
10.6 About the Curse of Dimensionality
10.7 About the No-Free-Lunch Theorem (NFL)
11. Appendices
11.1 Statistics and Probability Background
11.1.1 Commonly Used Discrete Distributions
11.1.2 Commonly Used Continuous Distributions
11.1.3 Major Results In Probability and Statistics
11.2 Linear Algebra Background
11.2.1 Fields, Vector Spaces and Subspaces
11.2.2 Linear independence, Bases and Dimension
11.2.3 Linear Transformations and Matrices
11.2.4 Eigenvalues and Spectral Decomposition
11.3 Computer Science Background
11.3.1 Computational Science and Complexity
11.3.2 Machine Learning
11.4 Typical Data Science Problems
11.5 A Sample of Common and Big Datasets
11.6 Computing Platforms
11.6.1 The Environment R
11.6.2 Python environments
References
Max H. Garzon is professor of computer science and bioinformatics at the U of Memphis. He has (co-)authored about 200 books, book chapters, journal or refereed conference publications. The main theme of his research is biomolecule-based computing and applications to areas such as bioinformatics, nanotechnology, self-assembly, machine learning and foundations of data science. He has served on the editorial board and as guest editor of several journals and as mentor of about 90 MS and PhD students in these areas. He has also served as TPC member and organizer of many scientific conferences and professional meetings and has been a visiting professor and guest scientist at several research institutions around the world.
Ching-Chi Yang is an assistant professor of mathematical sciences at the University of Memphis. He received a doctoral degree in statistics from The Pennsylvania State University in 2019. His primary interests focus on statistical learning, dimensional analysis, industrial and engineering statistics, response surface methodology. His related research projects vary from response surface methodology, tropical cyclone predictions, to stock price predictions. He has received awards including the American Society for Quality 2018 Fall Technical Conference Student Scholarship, and Jack and Eleanor Pettit Scholarship in Science from Penn State University.
Deepak Venugopal is an associate professor in the department of Computer Science at University of Memphis. His research interests lie in the fields of Machine Learning and Artificial Intelligence. In particular, he has made research contributions to statistical relational learning, Neuro-Symbolic AI, explainable AI and AI-based educational technologies. Dr. Venugopal regularly teaches Machine learning and AI courses both at the graduate and undergraduate levels.
Nirman Kumar is an assistant professor of computer science at the University of Memphis since 2016. His research area is approximation algorithms and Computational geometry. Nirman holds a Phd and Masters degree in Computer Science from the University of Illinois, and a Bachelors degree in Computer Science and Enginnering from the Indian Institute of Technology, Kanpur.
Kalidas Jana is post-doctoral research fellow in Economics and Data Science at the University of Memphis. He received Ph.D. in Economics from North Carolina State University in 2005. His research interests are in Econometrics and Data Science.
Lih-Yuan Deng received the B.S. and M.S. degree in Mathematics from National Taiwan University, Taiwan in 1975 and 1977, respectively. He also received M.S. and Ph.D. degrees in Computer Science and Statistics from University of Wisconsin-Madison, USA in 1982 and 1984, respectively. He is currently professor in the department of Mathematical Sciences, University of Memphis, USA. His active research work is mainly in the area of “design of random number generators” for computer simulation and computer security applications.
This book provides a practical and fairly comprehensive review of Data Science through the lens of dimensionality reduction, as well as hands-on techniques to tackle problems with data collected in the real world. State-of-the-art results and solutions from statistics, computer science and mathematics are explained from the point of view of a practitioner in any domain science, such as biology, cyber security, chemistry, sports science and many others. Quantitative and qualitative assessment methods are described to implement and validate the solutions back in the real world where the problems originated.
The ability to generate, gather and store volumes of data in the order of tera- and exo bytes daily has far outpaced our ability to derive useful information with available computational resources for many domains.
This book focuses on data science and problem definition, data cleansing, feature selection and extraction, statistical, geometric, information-theoretic, biomolecular and machine learning methods for dimensionality reduction of big datasets and problem solving, as well as a comparative assessment of solutions in a real-world setting.
This book targets professionals working within related fields with an undergraduate degree in any science area, particularly quantitative. Readers should be able to follow examples in this book that introduce each method or technique. These motivating examples are followed by precise definitions of the technical concepts required and presentation of the results in general situations. These concepts require a degree of abstraction that can be followed by re-interpreting concepts like in the original example(s). Finally, each section closes with solutions to the original problem(s) afforded by these techniques, perhaps in various ways to compare and contrast dis/advantages to other solutions.