Time: | Mon. 10-12 ct and Thu. 10-11 ct |
Location: | B26 and C213 |
Start: | Mon., October 22nd |
Tutorial: | Thu. 11-12 |
Machine Learning (also known as Data Mining, Pattern Recognition, Data Analysis, and Classification) is a research area at the intersection of computer science, artificial intelligence, mathematics and statistics, that addresses these problems. It covers general methods and techniques that then can be applied to a vast set of applications such as predicting customer behavior, steering a robot, detect spam, and predict the folding of a protein, to name just a few.
In this lecture we will study the most basic learning problems, starting with regression and classification problems (supervised learning). Here, we will look at different models such as linear classifiers, decision trees, neural networks, support vector machines, and simple types of Bayesian networks. Later on we also will look into clustering and dimensionality reduction (unsupervised learning).
The Machine Learning course is prerequisite for most courses and seminars of ISMLL. It is recommended for students from third term onwards.
Textbooks:
- Richard O. Duda, Peter E. Hart, David G. Stork (2001): Pattern Classification, 2nd edition.
- Christopher M. Bishop (2006): Pattern Recognition and Machine Learning.
- Trevor Hastie, Robert Tibshirani, Jerome Friedman (2001): The Elements of Statistical Learning.
- Brian D. Ripley (1996): Pattern Recognition and Neural Networks.
- Tom Mitchell (1997): Machine Learning.
- Ian Witten, Eibe Frank (2005): Data Mining: Practical Machine Learning Tools and Techniques, Second Edition.