Wulfram GerstnerWulfram Gerstner is Director of the Laboratory of Computational Neuroscience LCN at the EPFL. His research in computational neuroscience concentrates on models of spiking neurons and spike-timing dependent plasticity, on the problem of neuronal coding in single neurons and populations, as well as on the link between biologically plausible learning rules and behavioral manifestations of learning. He teaches courses for Physicists, Computer Scientists, Mathematicians, and Life Scientists at the EPFL. After studies of Physics in Tübingen and at the Ludwig-Maximilians-University Munich (Master 1989), Wulfram Gerstner spent a year as a visiting researcher in Berkeley. He received his PhD in theoretical physics from the Technical University Munich in 1993 with a thesis on associative memory and dynamics in networks of spiking neurons. After short postdoctoral stays at Brandeis University and the Technical University of Munich, he joined the EPFL in 1996 as assistant professor. Promoted to Associate Professor with tenure in February 2001, he is since August 2006 a full professor with double appointment in the School of Computer and Communication Sciences and the School of Life Sciences. Wulfram Gerstner has been invited speaker at numerous international conferences and workshops. He has served on the editorial board of the Journal of Neuroscience, Network: Computation in Neural Systems',
Journal of Computational Neuroscience', and `Science'.
Michel BierlaireBorn in 1967, Michel Bierlaire holds a PhD in Mathematical Sciences from the Facultés Universitaires Notre-Dame de la Paix, Namur, Belgium (University of Namur). Between 1995 and 1998, he was research associate and project manager at the Intelligent Transportation Systems Program of the Massachusetts Institute of Technology (Cambridge, Ma, USA). Between 1998 and 2006, he was a junior faculty in the Operations Research group ROSO within the Institute of Mathematics at EPFL. In 2006, he was appointed associate professor in the School of Architecture, Civil and Environmental Engineering at EPFL, where he became the director of the Transport and Mobility laboratory. Since 2009, he is the director of TraCE, the Transportation Center. From 2009 to 2017, he was the director of Doctoral Program in Civil and Environmental Engineering at EPFL. In 2012, he was appointed full professor at EPFL. Since September 2017, he is the head of the Civil Engineering Institute at EPFL. His main expertise is in the design, development and applications of models and algorithms for the design, analysis and management of transportation systems. Namely, he has been active in demand modeling (discrete choice models, estimation of origin-destination matrices), operations research (scheduling, assignment, etc.) and Dynamic Traffic Management Systems. As of August 2021, he has published 136 papers in international journals, 4 books, 41 book chapters, 193 articles in conference proceedings, 182 technical reports, and has given 195 scientific seminars. His Google Scholar h-index is 68. He is the founder, organizer and lecturer of the EPFL Advanced Continuing Education Course "Discrete Choice Analysis: Predicting Demand and Market Shares". He is the founder of hEART: the European Association for Research in Transportation. He was the founding Editor-in-Chief of the EURO Journal on Transportation and Logistics, from 2011 to 2019. He is an Associate Editor of Operations Research. He is the editor of two special issues for the journal Transportation Research Part C. He has been member of the Editorial Advisory Board (EAB) of Transportation Research Part B since 1995, of Transportation Research Part C since January 1, 2006.
Anthony Christopher DavisonAnthony Davison has published on a wide range of topics in statistical theory and methods, and on environmental, biological and financial applications. His main research interests are statistics of extremes, likelihood asymptotics, bootstrap and other resampling methods, and statistical modelling, with a particular focus on the first currently. Statistics of extremes concerns rare events such as storms, high winds and tides, extreme pollution episodes, sporting records, and the like. The subject has a long history, but under the impact of engineering and environmental problems has been an area of intense development in the past 20 years. Davison''s PhD work was in this area, in a project joint between the Departments of Mathematics and Mechanical Engineering at Imperial College, with the aim of modelling potential high exposures to radioactivity due to releases from nuclear installations. The key tools developed, joint with Richard Smith, were regression models for exceedances over high thresholds, which generalized earlier work by hydrologists, and formed the basis of some important later developments. This has led to an ongoing interest in extremes, and in particular their application to environmental and financial data. A major current interest is the development of suitable methods for modelling rare spatio-temporal events, particularly but not only in the context of climate change. Likelihood asymptotics too have undergone very substantial development since 1980. Key tools here have been saddlepoint and related approximations, which can give remarkably accurate approximate distribution and density functions even for very small sample sizes. These approximations can be used for wide classes of parametric models, but also for certain bootstrap and resampling problems. The literature on these methods can seem arcane, but they are potentially widely applicable, and Davison wrote a book joint with Nancy Reid and Alessandra Brazzale intended to promote their use in applications. Bootstrap methods are now used in many areas of application, where they can provide a researcher with accurate inferences tailor-made to the data available, rather than relying on large-sample or other approximations of doubtful validity. The key idea is to replace analytical calculations of biases, variances, confidence and prediction intervals, and other measures of uncertainty with computer simulation from a suitable statistical model. In a nonparametric situation this model consists of the data themselves, and the simulation simply involves resampling from the existing data, while in a parametric case it involves simulation from a suitable parametric model. There is a wide range of possibilities between these extremes, and the book by Davison and Hinkley explores these for many data examples, with the aim of showing how and when resampling methods succeed and why they can fail. He was Editor of Biometrika (2008-2017), Joint Editor of Journal of the Royal Statistical Society, series B (2000-2003), editor of the IMS Lecture Notes Monograph Series (2007), Associate Editor of Biometrika (1987-1999), and Associate Editor of the Brazilian Journal of Probability and Statistics (1987 2006). Currently he on the editorial board of Annual Reviews of Statistics and its Applications. He has served on committees of Royal Statistical Society and of the Institute of Mathematical Statistics. He is an elected Fellow of the American Statistical Assocation and of the Institute of Mathematical Statistics, an elected member of the International Statistical Institute, and a Chartered Statistician. In 2009 he was awarded a laurea honoris causa in Statistical Science by the University of Padova, in 2011 he held a Francqui Chair at Hasselt University, and in 2012 he was Mitchell Lecturer at the University of Glasgow. In 2015 he received the Guy Medal in Silver of the Royal Statistical Society and in 2018 was a Medallion Lecturer of the Institute of Mathematical Statistics.