Publication

What to Choose Next? A Paradigm for Testing Human Sequential Decision Making

Related publications (33)

End-to-End Learning for Stochastic Optimization: A Bayesian Perspective

Daniel Kuhn, Yves Rychener, Tobias Sutter

We develop a principled approach to end-to-end learning in stochastic optimization. First, we show that the standard end-to-end learning algorithm admits a Bayesian interpretation and trains a posterior Bayes action map. Building on the insights of this an ...
2023

Learning continuous-time working memory tasks with on-policy neural reinforcement learning

Davide Zambrano

An animals' ability to learn how to make decisions based on sensory evidence is often well described by Reinforcement Learning (RL) frameworks. These frameworks, however, typically apply to event-based representations and lack the explicit and fine-grained ...
2021

Decision Learning and Adaptation Over Multi-Task Networks

Ali H. Sayed

This paper studies the operation of multi-agent networks engaged in multi-task decision problems under the paradigm of simultaneous learning and adaptation. Two scenarios are considered:one in which a decision must be taken among multiple states of nature ...
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC2021

Optimal Adversarial Policies in the Multiplicative Learning System With a Malicious Expert

Negar Kiyavash, Seyed Jalal Etesami

We consider a learning system based on the conventional multiplicative weight ( MW) rule that combines experts' advice to predict a sequence of true outcomes. It is assumed that one of the experts is malicious and aims to impose the maximum loss on the sys ...
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC2021

Surprise-based model estimation in reinforcement learning: algorithms and brain signatures

Vasiliki Liakoni

Learning how to act and adapting to unexpected changes are remarkable capabilities of humans and other animals. In the absence of a direct recipe to follow in life, behaviour is often guided by rewarding and by surprising events. A positive or a negative o ...
EPFL2021

Ageing and reinforcement learning

Michael Herzog, Wei-Hsiang Lin

Learning to achieve one’s goal in a complex environment is a complicated task. In reinforcement learning (RL) tasks, an agent interacts with the environment to learn optimal actions. In humans, striatal areas are strongly involved in these tasks. During ag ...
2021

Signal generation for haptic using Reinforcement Learning

Marc Michel Charles Favier

The goal of this report is to present you my semester project on signal generation for haptic interfaces using Reinforcement Learning algorithm. The aim of this project is to improve the signal generated by state of the art methods. The vibration are gener ...
2021

Deep Reinforcement Learning for room temperature control: a black-box pipeline from data to policies

Colin Neil Jones, Bratislav Svetozarevic, Loris Di Natale

Deep Reinforcement Learning (DRL) recently emerged as a possibility to control complex systems without the need to model them. However, since weeks long experiments are needed to assess the performance of a building controller, people still have to rely on ...
2021

Neural Correlates of Reinforcement Learning: eligibility trace, reward prediction error, novelty and surprise

He Xu

In reinforcement learning (RL), an agent makes sequential decisions to maximise the reward it can obtain from an environment. During learning, the actual and expected outcomes are compared to tell whether a decision was good or bad. The difference between ...
EPFL2020

A Reinforcement Learning Approach to Train Routing and Scheduling

Zora Oswald

Good train scheduling for a big network with many trains is very hard to achieve. As the trains are competing for the tracks with one another, the number of constraints grows rapidly. Trying to take advantage of emerging technologies in the areas of optimi ...
2020

Graph Chatbot

Chat with Graph Search

Ask any question about EPFL courses, lectures, exercises, research, news, etc. or try the example questions below.

DISCLAIMER: The Graph Chatbot is not programmed to provide explicit or categorical answers to your questions. Rather, it transforms your questions into API requests that are distributed across the various IT services officially administered by EPFL. Its purpose is solely to collect and recommend relevant references to content that you can explore to help you answer your questions.