machine learning lab at diku
TRANSCRIPT
24/12/2014 DIKU Machine Learning Lab
http://image.diku.dk/MLLab/ 1/6
Machine Learning and Data Mining Research at DIKU
The amount and complexity of available data is steadily increasing. To make use of this wealth ofinformation, computing systems are needed that turn the data into knowledge. Machine learning isabout developing the required software that automatically analyses data for making predictions,categorizations, and recommendations. Machine learning algorithms are already an integral part oftoday's computing systems for example in search engines, recommender systems, or biometricalapplications and have reached superhuman performance in some domains. DIKU's researchpushes the boundaries and aims at more robust, more efficient, and more widely applicablemachine learning techniques.
Stateoftheart machine learningMachine learning is a branch of computer science and applied statistics covering software thatimproves its performance at a given task based on sample data or experience. The machinelearning research at DIKU, the Department of Computer Science at the University of Copenhagen,is concerned with the design and analysis of adaptive systems for pattern recognition andbehaviour generation.
We develop machine learning algorithms for making new discoveries in science[image from SkyML project].
Our fields of expertise are
classification, regression, and density estimation techniques for data mining and modelling,pattern recognition, and time series prediction; andcomputational intelligence methods for nonlinear optimisation including vector optimisation andmulticriteria decision making.
Successful realworld applications include the design of biometric and medical image processing
24/12/2014 DIKU Machine Learning Lab
http://image.diku.dk/MLLab/ 2/6
Medical image analysis is a major application area[taken from Prasoon et al., 2012, (top) and
Winter et al., 2008 (bottom)].
We apply machine learning algorithms forhydroacoustic signal classification tosupport the verification of the
Comprehensive NuclearTestBan Treaty[Tuma et al., 2012].
systems, chemical processes and plants, advanceddriver assistance systems, robot controllers, timeseries predictors for physical processes, systems forsports analytics, acoustic signal classification systems,automatic quality control for production lines, andsequence analysis in bioinformatics.
To build efficient and autonomous machine learningsystems we draw inspiration from optimisation andcomputing theory as well as biological informationprocessing. We analyse our algorithms theoretically andcritically evaluate them on realworld problems.Increasing the robustness and improving scalability ofselfadaptive, learning computer systems are crosscutting issues in our work. The following sectionshighlight some of our research activities.
Efficient autonomous machinelearningWe strive for computer systems that can dealautonomously and flexibly with our needs. They mustwork in scenarios that have not been fully specified and must be able to cope with unpredictedsituations. Incomplete descriptions of application scenarios are inevitable because we needalgorithms for domains where the designer's knowledge is not perfect, the solutions to particularproblems are simply unknown, and/or the sheer complexity and variability of the task and theenvironment precludes a sufficiently accurate domain description. Although such systems are ingeneral too complex to be designed manually, large amounts of data describing the task and theenvironment are often available or can be automatically obtained. To take proper advantage of thisavailable information, we need to develop systems that selfadapt and automatically improvebased on sample data – systems that learn.
Machine learning algorithms are already an integral part oftoday's computing systems, for example in internet searchengines, recommender systems, or biometrical applications.Highly specialised technical solutions for restricted taskdomains exist that have reached superhuman performance.Despite these successes, there are fundamental challengesthat must be met if we are to develop more general learningsystems.
First, present adaptive systems often lack autonomy androbustness. For example, they usually require a human expertto select the training examples, the learning method and itsparameters, and an appropriate representation or structure forthe learning system. This dependence on expert supervision isretarding the ubiquitous deployment of adaptive softwaresystems. We therefore work on algorithms that can handle large multimodal data sets, thatactively select training patterns, and that autonomously build appropriate internal representationsbased on data from different sources. These representations should foster learning, generalisation,and communication. Second, current adaptive systems succumb to scalability problems.
24/12/2014 DIKU Machine Learning Lab
http://image.diku.dk/MLLab/ 3/6
Covariance matrix adaptationevolution strategy (CMAES).
On the one hand, the ever growing amounts of data require highly efficient largescale learningalgorithms. On the other hand, learning and generalisation from very few examples is also achallenging problem. This scenario often occurs in manmachine interaction, for example insoftware personalisation or when generalisation from few database queries is required. We addressthe scaling problems by using taskspecific architectures incorporating both new concepts inspiredby natural adaptive systems as well as recent methods from algorithmic engineering andmathematical programming.
Selected methodsWe address all major learning paradigms, unsupervised, supervised, and reinforcement learning.These are closely connected. For instance, unsupervised learning can be used to find appropriaterepresentations for supervised learning and reliable supervised learning techniques are theprerequisite for successful reinforcement learning. Over the years, we used, analysed, and refineda broad spectrum of machine learning techniques. Currently our methodological research focuseson the following methods.
Supervised learning
Schema of multiclass support vector machine classification[taken from Dogan et al., 2011].
Support vector machines (SVMs) and other kernelbased algorithms are stateoftheart in patternrecognition. They perform well in many applications, especially in classification tasks. The kerneltrick allows for an easy handling of nonstandard data (e.g., biological sequences, multimodal data)and permits a better mathematical analysis of the adaptive system because of the convenientstructure of the hypothesis space. Developing and analysing kernelbased methods, in particularincreasing autonomy and improving scalability of SVMs, is currently one of the most activebranches of our research.
Reinforcement learningThe feedback in today's most challenging applications for adaptivesystems is sparse, unspecific, and/or delayed, for instance inautonomous robotics or in manmachine interaction. Supervised learningcannot be used directly in such a case, but the task can be cast into areinforcement learning (RL) problem. Reinforcement learning is learningfrom the consequences of interactions with an environment without beingexplicitly taught. Because the performance of standard RL techniques isfalling short of expectations, we are developing new RL algorithmsrelying on gradientbased and evolutionary direct policy search.
24/12/2014 DIKU Machine Learning Lab
http://image.diku.dk/MLLab/ 4/6
Markov random fieldfor rerepresenting data.
Contributing hypervolume ofcandidate solutions in
multiobjective optimization[Suttorp et al, 2006].
Direct policy search for adaptation in intelligent driver assistance systems[taken from Pellecchia et al., 2005].
Unsupervised and deep learningWe employ probabilistic generative models to learn and to describe probabilitydistributions. Our research focuses on Markov random fields, in which theconditional independence structure between random variables is described byan undirected graph. We are particularly interested in models that allow forlearning hierarchical representations of data in an unsupervised manner.
Nonlinear optimisationLearning is closely linked to optimisation. Thus, we are alsoworking on general gradientbased and direct search andoptimisation algorithms. This includes randomised methods,especially evolutionary algorithms (EAs), which are inspiredby neoDarwinian evolution theory. Efficient evolutionaryoptimisation can be achieved by an automatic adjustment ofthe search strategy. We are developing EAs with this ability,especially realvalued EAs that learn the metric underlying theproblem at hand (e.g., dependencies between variables).Currently, we are working on variablemetric EAs for RL andfor efficient vector (multiobjective) optimisation. The latterwill become increasingly relevant for industrial and scientificapplications in the future, because many problems areinherently multiobjective.
Team
Pengfei DiaoFabian GiesekeOswin KrauseChristian IgelMichiel KallenbergJan KremerDídac Rodríguez ArbonèsYevgeny SeldinKristoffer StensboSmidtLauge Sørensen
24/12/2014 DIKU Machine Learning Lab
http://image.diku.dk/MLLab/ 5/6
Matthias Tuma
Selected PublicationsPlease click here for a full list of Christian's papers and here for a full list of Yevgeny's papers.
Fabian Gieseke, Justin Heinermann, Cosmin Oancea, and Christian Igel. Buffer kd Trees:Processing Massive Nearest Neighbor Queries on GPUs. JMLR W&CP 32 (ICML) pp. 172180,2014
Yevgeny Seldin, Peter L. Bartlett, Koby Crammer, and Yasin AbbasiYadkori. Prediction with limitedadvice and multiarmed bandits with paid observations. In JMLR W&CP, 32 (ICML), 2014
Yevgeny Seldin and Aleksandrs Slivkins. One practical algorithm for both stochastic and adversarialbandits. In JMLR W&CP, 32 (ICML), 2014
Kai Brügge, Asja Fischer, and Christian Igel. The flipthestate transition operator for restrictedBoltzmann machines. Machine Learning 13, pp. 5369, 2013
Fabian Gieseke, Christian Igel, and Tapio Pahikkala. Polynomial runtime bounds for fixedrankunsupervised leastsquares classification. JMLR W&CP 29 (ACML), pp. 6271, 2013
Oswin Krause, Asja Fischer, Tobias Glasmachers, and Christian Igel. Approximation properties ofDBNs with binary hidden units and realvalued visible units. JMLR W&CP 28 (ICML), pp. 419–426, 2013
Ilya Tolstikhin and Yevgeny Seldin. PACBayesEmpiricalBernstein Inequality. In Advances inNeural Information Processing Systems (NIPS), 2013
Kim Steenstrup Pedersen, Kristoffer StensboSmidt, Andrew Zirm, and Christian Igel. Shape IndexDescriptors Applied to TextureBased Galaxy Analysis. International Conference on ComputerVision (ICCV), pp 24402447, IEEE Press, 2013
Yevgeny Seldin, François Laviolette, Nicolò CesaBianchi, John ShaweTaylor, and Peter Auer. PACBayesian inequalities for martingales. IEEE Transactions on Information Theory, 58(12), pp.70867093, 2012
Asja Fischer and Christian Igel. Bounding the Bias of Contrastive Divergence Learning. NeuralComputation 23, pp. 664673, 2011
Yevgeny Seldin, Peter Auer, François Laviolette, John ShaweTaylor, and Ronald Ortner. PACBayesian analysis of contextual bandits. In Advances in Neural Information ProcessingSystems (NIPS), 2011
Tobias Glasmachers and Christian Igel. Maximum Likelihood Model Selection for 1Norm SoftMargin SVMs with Multiple Parameters. IEEE Transactions on Pattern Analysis and MachineIntelligence 32(8), pp. 15221528, 2010 source code
Yevgeny Seldin and Naftali Tishby. PACBayesian analysis of coclustering and beyond. Journal ofMachine Learning Research 11, pp. 3595−3646, 2010
Thorsten Suttorp, Nikolaus Hansen, and Christian Igel. Efficient Covariance Matrix Update forVariable Metric Evolution Strategies. Machine Learning 75, pp. 167197, 2009 source code
Verena HeidrichMeisner and Christian Igel. Hoeffding and Bernstein Races for Selecting Policies inEvolutionary Direct Policy Search. In L. Bottou and M. Littman, eds.: Proceedings of the
24/12/2014 DIKU Machine Learning Lab
http://image.diku.dk/MLLab/ 6/6
Department of Computer ScienceUniversity of CopenhagenUniversitetsparken 1 2100 København Ø
International Conference on Machine Learning (ICML 2009), pp. 401408, 2009
Christian Igel, Verena HeidrichMeisner, and Tobias Glasmachers. Shark. Journal of MachineLearning Research 9, pp. 993996, 2008 source code
Tobias Glasmachers and Christian Igel. MaximumGain Working Set Selection for SVMs. Journal ofMachine Learning Research 7, pp. 14371466, 2006 source code
ContactChristian Igel, Professor mso, Dr. habil.
University of CopenhagenUniversitetsparken 52100 København Ø
Email: [email protected]:HCØ Building E, Office 4.0.2Phone:(+45) 21849673
Contact:The Image Group / Machine Learning Lab