<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>News on Thomas Bouvier</title><link>https://thomas-bouvier.io/en/news/</link><description>Recent content in News on Thomas Bouvier</description><generator>Hugo</generator><language>en-US</language><lastBuildDate>Mon, 29 Dec 2025 00:00:00 +0000</lastBuildDate><atom:link href="https://thomas-bouvier.io/en/news/index.xml" rel="self" type="application/rss+xml"/><item><title>I started a new position at CEA (Maison de la Simulation)</title><link>https://thomas-bouvier.io/en/news/cea-position/</link><pubDate>Mon, 29 Dec 2025 00:00:00 +0000</pubDate><guid>https://thomas-bouvier.io/en/news/cea-position/</guid><description>&lt;p&gt;I started working at CEA (Maison de la Simulation) as a research engineer. My role is part of the NumPEx project (&lt;a href="http://www.numpex.fr"&gt;http://www.numpex.fr&lt;/a&gt;) which aims to build a software stack for &lt;a href="https://en.wikipedia.org/wiki/Exascale_computing"&gt;Exascale supercomputers&lt;/a&gt; including &lt;a href="https://www.franceinfo.fr/replay-radio/nouveau-monde/bienvenue-a-alice-recoque-le-premier-supercalculateur-exascale-francais_7617590.html"&gt;Alice Recoque&lt;/a&gt;, scheduled for deployment in 2027. This machine will be among the most powerful in the world (&lt;a href="https://top500.org"&gt;Top500 ranking&lt;/a&gt;), used for both traditional scientific applications and artificial intelligence workloads.&lt;/p&gt;
&lt;p&gt;My responsibilities include:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Help designing and implementing the packaging and continuous integration strategy for the NumPEx project.&lt;/li&gt;
&lt;li&gt;Participating in the deployment and testing of the infrastructure.&lt;/li&gt;
&lt;li&gt;Providing user support and training on packaging, deployment, and testing tools and workflows.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;The project’s packaging strategy is based on cutting-edge open-source tools such as:&lt;/p&gt;</description></item><item><title>We are launching the OpenFresque digital commons</title><link>https://thomas-bouvier.io/en/news/openfresque-launch/</link><pubDate>Tue, 17 Jun 2025 00:00:00 +0000</pubDate><guid>https://thomas-bouvier.io/en/news/openfresque-launch/</guid><description>&lt;p&gt;The Fresk-type awareness workshops enable the engagement of various sectors of activity (industries, transportation, food, buildings, etc.) in their necessary transitions, both in France and around the world.&lt;/p&gt;
&lt;p&gt;Currently, tools to manage these events (ticketing, opportunity sharing, pathways, resources) are provided by workshop developers, but this is done in a heterogeneous and non-shared manner.&lt;/p&gt;
&lt;p&gt;This poses several problems:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;
&lt;p&gt;Workshop developers invest in parallel to develop or finance external solutions without pooling these resources.&lt;/p&gt;</description></item><item><title>I will be defending my PhD on Monday, November 4th, 2024</title><link>https://thomas-bouvier.io/en/news/phd-defense/</link><pubDate>Thu, 31 Oct 2024 00:00:00 +0000</pubDate><guid>https://thomas-bouvier.io/en/news/phd-defense/</guid><description>&lt;p&gt;My PhD defense is scheduled for &lt;strong&gt;Monday, November 4th, 2024, at 1:30 pm&lt;/strong&gt;. The event will take place at IRISA/Inria Rennes (Markov room), 263 Avenue du Général Leclerc, 35042 Rennes (&lt;a href="https://www.openstreetmap.org/way/81586498"&gt;maps link&lt;/a&gt;). I can&amp;rsquo;t wait to share this moment with my jury, family, friends and colleagues.&lt;/p&gt;
&lt;p&gt;It will also be broadcast live, please get in touch to get the link!&lt;/p&gt;
&lt;p&gt;The reviewed dissertation of my thesis can be found here: &lt;a href="https://thomas-bouvier.io/papers/phd24.pdf"&gt;link to dissertation&lt;/a&gt;. It is entitled: “&lt;em&gt;Distributed Rehearsal Buffers for Continual Learning at Scale&lt;/em&gt;”. The abstract can be found at the bottom of this page. The presentation will be in English.&lt;/p&gt;</description></item><item><title>I will give a talk at JLESC 16 @ Kobe</title><link>https://thomas-bouvier.io/en/news/jlesc-16/</link><pubDate>Tue, 23 Apr 2024 00:00:00 +0000</pubDate><guid>https://thomas-bouvier.io/en/news/jlesc-16/</guid><description>&lt;p&gt;I will present our ongoing work entitled &amp;ldquo;Efficient Distributed Continual Learning for Steering Experiments in Real-Time&amp;rdquo; at &lt;a href="https://sites.google.com/view/jlesc16"&gt;JLESC 16&lt;/a&gt;. This presentation is an update on the progress of the JLESC project entitled &lt;a href="https://jlesc.github.io/projects/continual_learning_project/"&gt;&amp;ldquo;Towards Continual Learning at Scale&amp;rdquo;&lt;/a&gt;, which has been running since 2022. You can find the full program &lt;a href="https://docs.google.com/spreadsheets/d/1ohehnazz5gbpNjA-52BljhpQTzp4QUVHYPNkww5hJd4/edit?gid=871164259#gid=871164259"&gt;here&lt;/a&gt;.&lt;/p&gt;
&lt;h2 id="efficient-distributed-continual-learning-for-steering-experiments-in-real-time---project-update"&gt;Efficient Distributed Continual Learning for Steering Experiments in Real-Time - Project Update&lt;/h2&gt;
&lt;p&gt;Deep learning has emerged as a powerful method for extracting valuable information from large volumes of data. However, when new training data arrives continuously (i.e., is not fully available from the beginning), incremental training suffers from catastrophic forgetting (i.e., new patterns are reinforced at the expense of previously acquired knowledge). Training from scratch each time new training data becomes available would result in extremely long training times and massive data accumulation. Rehearsal-based continual learning has shown promise for addressing the catastrophic forgetting challenge, but research to date has not addressed performance and scalability. To fill this gap, we propose an approach based on a distributed rehearsal buffer that efficiently complements data-parallel training on multiple GPUs to achieve high accuracy, short runtime, and scalability. It leverages a set of buffers (local to each GPU) and uses several asynchronous techniques for updating these local buffers in an embarrassingly parallel fashion, all while handling the communication overheads necessary to augment input mini-batches (groups of training samples fed to the model) using unbiased, global sampling. After evaluating our approach on classification problems, we further propose a generalization of rehearsal buffers to support generative learning tasks, as well as more advanced rehearsal strategies (notably dark experience replay, leveraging knowledge distillation). We illustrate these extensions with a real-life HPC streaming application from the domain of ptychographic image reconstruction, in which experiments need to be steered in real-time.&lt;/p&gt;</description></item><item><title>One paper has been accepted at CCGrid 2024</title><link>https://thomas-bouvier.io/en/news/ccgrid24-paper/</link><pubDate>Mon, 12 Feb 2024 00:00:00 +0000</pubDate><guid>https://thomas-bouvier.io/en/news/ccgrid24-paper/</guid><description>&lt;p&gt;Our paper entitled &amp;ldquo;Efficient Data-Parallel Continual Learning with Asynchronous Distributed Rehearsal Buffers&amp;rdquo; has been accepted at &lt;a href="https://2024.ccgrid-conference.org/"&gt;CCGrid 2024&lt;/a&gt;. I will be presenting this work in the &amp;ldquo;ML for Systems and Systems for ML&amp;rdquo; track on May 7 at 2:30 PM in Philadelphia. You can find the full program &lt;a href="https://2024.ccgrid-conference.org/program/"&gt;here&lt;/a&gt;&lt;/p&gt;
&lt;h2 id="efficient-data-parallel-continual-learning-with-asynchronous-distributed-rehearsal-buffers"&gt;Efficient Data-Parallel Continual Learning with Asynchronous Distributed Rehearsal Buffers&lt;/h2&gt;
&lt;p&gt;Deep learning has emerged as a powerful method for extracting valuable information from large volumes of data. However, when new training data arrives continuously (i.e., is not fully available from the beginning), incremental training suffers from catastrophic forgetting (i.e., new patterns are reinforced at the expense of previously acquired knowledge). Training from scratch each time new training data becomes available would result in extremely long training times and massive data accumulation. Rehearsal-based continual learning has shown promise for addressing the catastrophic forgetting challenge, but research to date has not addressed performance and scalability. To fill this gap, we propose an approach based on a distributed rehearsal buffer that efficiently complements data-parallel training on multiple GPUs, allowing us to achieve short runtime and scalability while retaining high accuracy. It leverages a set of buffers (local to each GPU) and uses several asynchronous techniques for updating these local buffers in an embarrassingly parallel fashion, all while handling the communication overheads necessary to augment input mini-batches (groups of training samples fed to the model) using unbiased, global sampling. In this paper we explore the benefits of this approach for classification models. We run extensive experiments on up to 128 GPUs of the ThetaGPU supercomputer to compare our approach with baselines representative of training-from-scratch (the upper bound in terms of accuracy) and incremental training (the lower bound). Results show that rehearsal-based continual learning achieves a top-5 classification accuracy close to the upper bound, while simultaneously exhibiting a runtime close to the lower bound.&lt;/p&gt;</description></item><item><title>I will give a talk at JLESC 15 @ Bordeaux</title><link>https://thomas-bouvier.io/en/news/jlesc-15/</link><pubDate>Sun, 12 Feb 2023 00:00:00 +0000</pubDate><guid>https://thomas-bouvier.io/en/news/jlesc-15/</guid><description>&lt;p&gt;I will present our ongoing work entitled &amp;ldquo;Leveraging Rehearsal Buffers to Enable Efficient Data-Parallel Continual Learning&amp;rdquo; at &lt;a href="https://events.hifis.net/event/617/"&gt;JLESC 15&lt;/a&gt;. This presentation is an update on the progress of the JLESC project entitled &lt;a href="https://jlesc.github.io/projects/continual_learning_project/"&gt;&amp;ldquo;Towards Continual Learning at Scale&amp;rdquo;&lt;/a&gt;, which has been running since 2022. You can find the full program &lt;a href="https://events.hifis.net/event/617/timetable/"&gt;here&lt;/a&gt;.&lt;/p&gt;
&lt;h2 id="leveraging-rehearsal-buffers-to-enable-efficient-data-parallel-continual-learning---project-update"&gt;Leveraging Rehearsal Buffers to Enable Efficient Data-Parallel Continual Learning - Project Update&lt;/h2&gt;
&lt;p&gt;Deep Learning (DL) emerged as a way to extract valuable information from ever-growing volumes of data. However, when trained on sequential tasks ie. without full access to the dataset at the beginning of the training, typical Deep Neural Networks (DNNs) suffer from catastrophic forgetting, a phenomenon causing them to reinforce new patterns at the expense of previously acquired knowledge. This limitation prevents updating models incrementally, which is problematic in many real-life scenarios where the aforementioned datasets are replaced by data streams generated over time by distributed devices. It is unfeasible to train models from scratch every time new samples are being ingested either, as this would lead to prohibitive time and/or resource constraints.&lt;/p&gt;</description></item><item><title>I will give a talk at JLESC 14 @ Urbana-Champaign</title><link>https://thomas-bouvier.io/en/news/jlesc-14/</link><pubDate>Wed, 21 Sep 2022 00:00:00 +0000</pubDate><guid>https://thomas-bouvier.io/en/news/jlesc-14/</guid><description>&lt;p&gt;I will present our new JLESC project entitled &lt;a href="https://jlesc.github.io/projects/continual_learning_project/"&gt;&amp;ldquo;Towards Continual Learning at Scale&amp;rdquo;&lt;/a&gt;, as well as preliminary results, at &lt;a href="https://publish.illinois.edu/14th-jlesc-workshop/"&gt;JLESC 14&lt;/a&gt;. You can find the full program &lt;a href="https://publish.illinois.edu/14th-jlesc-workshop/agenda/"&gt;here&lt;/a&gt;. I&amp;rsquo;m really excited to be giving my first public talk in the US, which will take place at the University of Illinois Urbana-Champaign. 😊&lt;/p&gt;
&lt;h2 id="towards-continual-learning-at-scale---project-kick-off"&gt;Towards Continual Learning at Scale - Project Kick-off&lt;/h2&gt;
&lt;p&gt;During the past decade, Deep learning (DL) supported the shift from rule-based systems towards statistical models. Deep Neural Networks (DNNs) are achieving high accuracy on various benchmarks by extracting patterns from complex datasets. Although presenting promising results, most existing supervised learning algorithms operate under the assumptions that the data is (i) i.i.d.; (ii) static; and (iii) available before the training process. These constraints limit their use in real-life scenarios where the aforementioned datasets are replaced by high volume, high velocity data streams generated over time by distributed devices. It is unfeasible to keep training models in an offline fashion from scratch every time new data arrives, as this would lead to prohibitive time and/or resource constraints. At the same time, it is not possible to train learning models incrementally either, due to catastrophic forgetting, a phenomenon causing typical DNNs to reinforce new patterns at the expense of previously acquired knowledge i.e. inducing biases.&lt;/p&gt;</description></item><item><title>I will be spending 4 months working at Argonne National Lab @ Chicago</title><link>https://thomas-bouvier.io/en/news/anl-appointment/</link><pubDate>Wed, 13 Apr 2022 00:00:00 +0000</pubDate><guid>https://thomas-bouvier.io/en/news/anl-appointment/</guid><description>&lt;p&gt;This summer, I will be working as a visiting PhD student at Argonne National Lab, in the context of the &lt;a href="https://team.inria.fr/unify/"&gt;UNIFY Associate Team&lt;/a&gt;. I will be mentored by &lt;a href="http://bnicolae.net/"&gt;Bogdan Nicolae&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;My work is part of the ongoing &lt;a href="https://jlesc.github.io/projects/continual_learning_project/"&gt;&amp;ldquo;Towards Continual Learning at Scale&amp;rdquo; JLESC project&lt;/a&gt;. The project aims to achieve two main objectives: the (1) design and implementation of a distributed replay buffer leveraging distributed systems effectively and the (2) study of trade-offs introduced by large-scale CL in terms of training time, accuracy and memory usage. All experiments will be conducted on the &lt;a href="https://www.alcf.anl.gov/polaris"&gt;ALCF Polaris&lt;/a&gt; supercomputer.&lt;/p&gt;</description></item></channel></rss>