NEURAL REPRESENTATIONS IN MULTI-TASK LEARNING GUIDED BY TASK-DEPENDENT CONTEXTS

Abstract

The ability to switch between tasks effectively in response to external stimuli is a hallmark of cognitive control. Our brain is able to filter and integrate external information to accomplish goal-directed behavior. Task switching occurs rapidly and efficiently, allowing us to perform multiple tasks with ease. In a similar way, artificial neural networks can be tailored to exhibit multi-task capabilities and achieve high performance across domains. In terms of explainability, understanding how neural networks make predictions is crucial in many real-world applications, for instance, in guiding clinical decisions. In this study, we delve into neural representations learned by multi-tasking architectures. Concretely, we compare individual and parallel networks with taskswitching networks. Task-switching networks leverage task-dependent contexts to learn disentangled representations without hurting the overall task accuracy. We show that task-switching networks operate in an intermediate regime between individual and parallel networks. In addition, we show that shared representations are produced by the emergence neurons encoding multiple tasks. Furthermore, we study the role of contexts across network processing and show its role at aligning the task with the relevant features. Finally, we investigate how the magnitude of contexts affects the performance in task-switching networks.

1. INTRODUCTION

Living involves constantly gauging and selecting the optimal task to perform. This decision is the result of the interaction of different elements, such as our current goals, external circumstances, or the stimulus context (Monsell, 2003) . In the brain, task switching and cognitive control has been associated principally with the prefrontal cortex (PFC), which provides top-down regulation to other cortical brain areas Johnston et al. (2007) . In this sense, PFC controls the activation of multiple neural pathways that are activated or deactivated to ultimately result in task execution. Neural pathways are formed by collections of neurons that cooperate to yield a specific effect. Although single neurons have been historically regarded as centre of processing in the brain, now we are moving towards a framework in which neural networks conform the functional processing unit of the nervous system Yuste (2015). In the visual cortex, where subpopulations of neurons are responsible of the processing of different visual field features for correct object recognition DiCarlo et al. (2012) . According to the gating theory (Miller et al., 2001) , context-dependent top-down inputs from the PFC regulate the activation of neural pathways. The encoding of relationships between stimulus and contexts can be facilitated by tuning neural activity to multiple tasks. In the PFC of monkeys, it has been observed that single neurons exhibit nonlinear responses to multiple stimuli (Rigotti et al., 2013) . This behavior, or mixed selectivity, favors high-dimensional representation of neural activation, which allows linear readouts to generate a vast number of responses, in opposition to low-dimensional representations (Fusi et al., 2016) . Recently, artificial neural networks have been revisited as models of neural computation, many findings suggesting their practicality for assessing brain theories Richards et al. (2019) . For example, the internal representations learned by neural networks have been associated with representations in the brain in multi-tasking settings (Ito et al., 2022; Ito & Murray, 2021; Flesch et al., 2022a; b) . In this paper, we investigate the neural representations learned by feedforward multi-tasking architectures. Neural networks have been designed to be capable of processing multiple tasks in parallel, 1

