AD ALTA
JOURNAL OF INTERDISCIPLINARY RESEARCH
will allow them to learn more about the behavior of each
pupil, but also of the group of students as a whole.
It enables the possibility to expand our understanding of the
process of virtual learning. It is essential that eLearning
professionals get to know how students learn and acquire
knowledge. Big Data gives us the opportunity to gain a
deeper understanding of the process of eLearning and how
students are responding to the eLearning courses. This
information can be used to design new learning methods.
4 Conclusion
Feature selection has been usually used as a preprocessing step
that condenses the extents of a problematic and advances
classification precision. The need for this kind of technique has
improved intensely in recent years in order to cope with
situations categorized by both a high number of input features
and/or of models. In other words, the big data bang now has the
added problem of big dimensionality. This research work
assessed the main need for feature selection and momentarily
revised the most popular feature selection methods and some
typical applications that are used for virtual learning in higher
education. While feature selection may well be one of the
improved preprocessing methods, it is vital not to overlook the
factors affecting feature selection choices. For illustration, it is
important to choose a satisfactory discretization method, given
that some feature selection approaches especially those from the
information theory field were developed to work with separate
data. Certainly, it has been established that the choice of method
affects the results of the feature selection process in virtual
learning.
Literature:
1.
A.J. Miller, Selection of subsets of regression variables, J.
Roy. Stat. Soc. Ser. A (Gener.) (1984) 389–425.
2.
A.L. Blum, P. Langley, Selection of relevant features and
examples in machine learning, Artif. Intell. 97 (1) (1997) 245–
271.
3.
B. Bonev, Feature Selection based on Information Theory,
Universidad de Alicante, 2010.
4.
C. Boutsidis, P. Drineas, M.W. Mahoney, Unsupervised
feature selection for the k-means clustering problem, Adv.
Neural Inform. Process. Syst (2009) 153–161.
5.
D.L. Donoho et al., High-dimensional data analysis: the
curses and blessings of dimensionality, AMS Math. Challenges
Lect. (2000) 1–32.
6.
D. Paul, E. Bair, T. Hastie, R. Tibshirani,
‘‘Preconditioning’’ for feature selection and regression in high-
dimensional problems, Ann. Stat. (2008) 1595–1618.
7.
Guyon, Feature Extraction: Foundations and Applications,
vol. 207, Springer, 2006.
8.
Guyon, A. Elisseeff, An introduction to variable and feature
selection, J. Mach. Learn. Res. 3 (2003) 1157–1182.
9.
G. Hughes, On the mean accuracy of statistical pattern
recognizers, IEEE Trans. Inform. Theory 14 (1) (1968) 55–63.
10.
H. Liu, H. Motoda, Computational Methods of Feature
Selection, CRC Press, 2007.
11.
H. Liu, R. Setiono, Chi2: feature selection and
discretization of numeric attributes, in: Tools with Artificial
Intelligence, 1995. Proceedings., Seventh International
Conference on, IEEE, 1995, pp. 388–391.
12.
J.R. Quinlan, Induction of decision trees, Mach. Learn. 1 (1)
(1986) 81–106.
13.
Kononenko, Estimating attributes: analysis and extensions
of relief, in: Machine Learning: ECML-94, Springer, 1994, pp.
171–182.
14.
K. Weinberger, A. Dasgupta, J. Langford, A. Smola, J.
Attenberg, Feature hashing for large scale multitask learning,
in: Proceedings of the 26th Annual International Conference on
Machine Learning, ACM, 2009, pp. 1113–1120.
15.
L. Yu, H. Liu, Efficient feature selection via analysis of
relevance and redundancy, J. Mach. Learn. Res. 5 (2004) 1205–
1224.
16.
M. Pal, G.M. Foody, Feature selection for classification of
hyperspectral data by svm, IEEE Trans. Geosci. Remote Sens.
48 (5) (2010) 2297–2307.
17.
M. Tan, I.W. Tsang, L. Wang, Towards ultrahigh
dimensional feature selection for big data, J. Mach. Learn. Res.
15 (1) (2014) 1371–1429.
18.
M. Dash, H. Liu, Feature selection for classification, Intell.
Data Anal. 1 (3) (1997) 131–156.
19.
P.P. Ohanian, R.C. Dubes, Performance evaluation for four
classes of textural features, Pattern Recognit. 25 (8) (1992) 819–833.
20.
R. Leardi, A. Lupiáñez González, Genetic algorithms
applied to feature selection in pls regression: how and when to
use them, Chemomet. Intell. Lab. Syst. 41 (2) (1998) 195–207.
21.
R.O. Duda, P.E. Hart, D.G. Stork, Pattern Classification,
John Wiley & Sons, 1999.
22.
R. Bellman, Dynamic Programming, vol. 18, Princeton UP,
Princeton, NJ, 1957.
23.
R. Kohavi, G.H. John, Wrappers for feature subset
selection, Artif. Intell. 97 (1) (1997) 273–324.
24.
V. Roth, T. Lange, Feature selection in clustering problems,
Adv. Neural Inform. Process. Syst. (2003).
25.
V. Bolón-Canedo, N. Sánchez-Maroño, A. Alonso-
Betanzos, J.M. Benítez, F. Herrera, A review of microarray
datasets and applied feature selection methods, Inform. Sci. 282
(2014) 111–135.
26.
Y. Zhai, Y. Ong, I. Tsang, The emerging ‘big dimensionality’,
Comput. Intell. Mag., IEEE 9 (3) (2014) 14–26.
27.
Z.A. Zhao, H. Liu, Spectral Feature Selection for Data
Mining, Chapman & Hall/CRC, 2011.
Primary Paper Section: A
Secondary Paper Section: AO, AM
- 66 -
vol. 8
issue 1