235 results on '"Van De Weijer, Joost"'
Search Results
2. Context effects on duration, fundamental frequency, and intonation inhuman-directed domestic cat meows
- Author
-
Schötz, Susanne, van de Weijer, Joost, Eklund, Robert, Schötz, Susanne, van de Weijer, Joost, and Eklund, Robert
- Abstract
In this study, we investigated the prosody of domestic cat meows produced in different contexts. Prosodic cues –i.e., variation in intonation, duration, voice quality and fundamental frequency – in humans as well as innonhuman animals carry information about idiosyncratic traits of the signaller, including sex, age, and physicaland mental state. The duration, fundamental frequency (F0) and intonation in a sample of 969 meows recordedin seven different contexts (i.e., cuddle, door, food, greeting, lifting, play, cat carrier) were analysed using linearmixed effects regression and generalized additive models. In this, we controlled for cat age and sex, as meowsproduced by old cats had lower mean F0 than those produced by young cats, and female cats produced meowswith higher mean F0 than male cats. We found significant effects of context on duration and mean F0, but not onF0 range. Furthermore, the results showed that the intonation of meows produced by cats in a cat carrier displayeda falling pattern, while that of meows produced in cuddle and door contexts was relatively level, and thatof meows produced in the other contexts consisted of combinations of rising and falling. The average slope ofmeows produced in cat carrier and play contexts was negative, while that of meows produced in the othercontexts was positive. We argue that this prosodic variation reflects the cats’ mental or emotional state, becauseof valence and arousal differences associated with the various contexts that were included in the study. Furtherstudies will need to confirm this. In addition, we also plan additional analyses of spectral and voice qualityparameters in meows and other cat vocalisation types., Funding: Marcus and Amalia Wallenberg Foundation, Sweden [MAW 2015.0054], Melody In Human–Cat Communication (Meowsic)
- Published
- 2024
- Full Text
- View/download PDF
3. Accelerated Inference and Reduced Forgetting: The Dual Benefits of Early-Exit Networks in Continual Learning
- Author
-
Szatkowski, Filip, Yang, Fei, Twardowski, Bartłomiej, Trzciński, Tomasz, van de Weijer, Joost, Szatkowski, Filip, Yang, Fei, Twardowski, Bartłomiej, Trzciński, Tomasz, and van de Weijer, Joost
- Abstract
Driven by the demand for energy-efficient employment of deep neural networks, early-exit methods have experienced a notable increase in research attention. These strategies allow for swift predictions by making decisions early in the network, thereby conserving computation time and resources. However, so far the early-exit networks have only been developed for stationary data distributions, which restricts their application in real-world scenarios with continuous non-stationary data. This study aims to explore the continual learning of the early-exit networks. We adapt existing continual learning methods to fit with early-exit architectures and investigate their behavior in the continual setting. We notice that early network layers exhibit reduced forgetting and can outperform standard networks even when using significantly fewer resources. Furthermore, we analyze the impact of task-recency bias on early-exit inference and propose Task-wise Logits Correction (TLC), a simple method that equalizes this bias and improves the network performance for every given compute budget in the class-incremental setting. We assess the accuracy and computational cost of various continual learning techniques enhanced with early-exits and TLC across standard class-incremental learning benchmarks such as 10 split CIFAR100 and ImageNetSubset and show that TLC can achieve the accuracy of the standard methods using less than 70\% of their computations. Moreover, at full computational budget, our method outperforms the accuracy of the standard counterparts by up to 15 percentage points. Our research underscores the inherent synergy between early-exit networks and continual learning, emphasizing their practical utility in resource-constrained environments.
- Published
- 2024
4. Get What You Want, Not What You Don't: Image Content Suppression for Text-to-Image Diffusion Models
- Author
-
Li, Senmao, van de Weijer, Joost, Hu, Taihang, Khan, Fahad Shahbaz, Hou, Qibin, Wang, Yaxing, Yang, Jian, Li, Senmao, van de Weijer, Joost, Hu, Taihang, Khan, Fahad Shahbaz, Hou, Qibin, Wang, Yaxing, and Yang, Jian
- Abstract
The success of recent text-to-image diffusion models is largely due to their capacity to be guided by a complex text prompt, which enables users to precisely describe the desired content. However, these models struggle to effectively suppress the generation of undesired content, which is explicitly requested to be omitted from the generated image in the prompt. In this paper, we analyze how to manipulate the text embeddings and remove unwanted content from them. We introduce two contributions, which we refer to as $\textit{soft-weighted regularization}$ and $\textit{inference-time text embedding optimization}$. The first regularizes the text embedding matrix and effectively suppresses the undesired content. The second method aims to further suppress the unwanted content generation of the prompt, and encourages the generation of desired content. We evaluate our method quantitatively and qualitatively on extensive experiments, validating its effectiveness. Furthermore, our method is generalizability to both the pixel-space diffusion models (i.e. DeepFloyd-IF) and the latent-space diffusion models (i.e. Stable Diffusion)., Comment: ICLR 2024. Our code is available in https://github.com/sen-mao/SuppressEOT
- Published
- 2024
5. Elastic Feature Consolidation for Cold Start Exemplar-Free Incremental Learning
- Author
-
Magistri, Simone, Trinci, Tomaso, Soutif-Cormerais, Albin, van de Weijer, Joost, Bagdanov, Andrew D., Magistri, Simone, Trinci, Tomaso, Soutif-Cormerais, Albin, van de Weijer, Joost, and Bagdanov, Andrew D.
- Abstract
Exemplar-Free Class Incremental Learning (EFCIL) aims to learn from a sequence of tasks without having access to previous task data. In this paper, we consider the challenging Cold Start scenario in which insufficient data is available in the first task to learn a high-quality backbone. This is especially challenging for EFCIL since it requires high plasticity, which results in feature drift which is difficult to compensate for in the exemplar-free setting. To address this problem, we propose a simple and effective approach that consolidates feature representations by regularizing drift in directions highly relevant to previous tasks and employs prototypes to reduce task-recency bias. Our method, called Elastic Feature Consolidation (EFC), exploits a tractable second-order approximation of feature drift based on an Empirical Feature Matrix (EFM). The EFM induces a pseudo-metric in feature space which we use to regularize feature drift in important directions and to update Gaussian prototypes used in a novel asymmetric cross entropy loss which effectively balances prototype rehearsal with data from new tasks. Experimental results on CIFAR-100, Tiny-ImageNet, ImageNet-Subset and ImageNet-1K demonstrate that Elastic Feature Consolidation is better able to learn new tasks by maintaining model plasticity and significantly outperform the state-of-the-art., Comment: Accepted at Twelfth International Conference on Learning Representations (ICLR 2024)
- Published
- 2024
6. Resurrecting Old Classes with New Data for Exemplar-Free Continual Learning
- Author
-
Goswami, Dipam, Soutif--Cormerais, Albin, Liu, Yuyang, Kamath, Sandesh, Twardowski, Bartłomiej, van de Weijer, Joost, Goswami, Dipam, Soutif--Cormerais, Albin, Liu, Yuyang, Kamath, Sandesh, Twardowski, Bartłomiej, and van de Weijer, Joost
- Abstract
Continual learning methods are known to suffer from catastrophic forgetting, a phenomenon that is particularly hard to counter for methods that do not store exemplars of previous tasks. Therefore, to reduce potential drift in the feature extractor, existing exemplar-free methods are typically evaluated in settings where the first task is significantly larger than subsequent tasks. Their performance drops drastically in more challenging settings starting with a smaller first task. To address this problem of feature drift estimation for exemplar-free methods, we propose to adversarially perturb the current samples such that their embeddings are close to the old class prototypes in the old model embedding space. We then estimate the drift in the embedding space from the old to the new model using the perturbed images and compensate the prototypes accordingly. We exploit the fact that adversarial samples are transferable from the old to the new feature space in a continual learning setting. The generation of these images is simple and computationally cheap. We demonstrate in our experiments that the proposed approach better tracks the movement of prototypes in embedding space and outperforms existing methods on several standard continual learning benchmarks as well as on fine-grained datasets. Code is available at https://github.com/dipamgoswami/ADC., Comment: Accepted at CVPR 2024
- Published
- 2024
7. An Empirical Analysis of Forgetting in Pre-trained Models with Incremental Low-Rank Updates
- Author
-
Soutif--Cormerais, Albin, Magistri, Simone, van de Weijer, Joost, Bagdanov, Andew D., Soutif--Cormerais, Albin, Magistri, Simone, van de Weijer, Joost, and Bagdanov, Andew D.
- Abstract
Broad, open source availability of large pretrained foundation models on the internet through platforms such as HuggingFace has taken the world of practical deep learning by storm. A classical pipeline for neural network training now typically consists of finetuning these pretrained network on a small target dataset instead of training from scratch. In the case of large models this can be done even on modest hardware using a low rank training technique known as Low-Rank Adaptation (LoRA). While Low Rank training has already been studied in the continual learning setting, existing works often consider storing the learned adapter along with the existing model but rarely attempt to modify the weights of the pretrained model by merging the LoRA with the existing weights after finishing the training of each task. In this article we investigate this setting and study the impact of LoRA rank on the forgetting of the pretraining foundation task and on the plasticity and forgetting of subsequent ones. We observe that this rank has an important impact on forgetting of both the pretraining and downstream tasks. We also observe that vision transformers finetuned in that way exhibit a sort of ``contextual'' forgetting, a behaviour that we do not observe for residual networks and that we believe has not been observed yet in previous continual learning works.
- Published
- 2024
8. Calibrating Higher-Order Statistics for Few-Shot Class-Incremental Learning with Pre-trained Vision Transformers
- Author
-
Goswami, Dipam, Twardowski, Bartłomiej, van de Weijer, Joost, Goswami, Dipam, Twardowski, Bartłomiej, and van de Weijer, Joost
- Abstract
Few-shot class-incremental learning (FSCIL) aims to adapt the model to new classes from very few data (5 samples) without forgetting the previously learned classes. Recent works in many-shot CIL (MSCIL) (using all available training data) exploited pre-trained models to reduce forgetting and achieve better plasticity. In a similar fashion, we use ViT models pre-trained on large-scale datasets for few-shot settings, which face the critical issue of low plasticity. FSCIL methods start with a many-shot first task to learn a very good feature extractor and then move to the few-shot setting from the second task onwards. While the focus of most recent studies is on how to learn the many-shot first task so that the model generalizes to all future few-shot tasks, we explore in this work how to better model the few-shot data using pre-trained models, irrespective of how the first task is trained. Inspired by recent works in MSCIL, we explore how using higher-order feature statistics can influence the classification of few-shot classes. We identify the main challenge of obtaining a good covariance matrix from few-shot data and propose to calibrate the covariance matrix for new classes based on semantic similarity to the many-shot base classes. Using the calibrated feature statistics in combination with existing methods significantly improves few-shot continual classification on several FSCIL benchmarks. Code is available at https://github.com/dipamgoswami/FSCIL-Calibration., Comment: Accepted at CLVision workshop (CVPR 2024)
- Published
- 2024
9. LocInv: Localization-aware Inversion for Text-Guided Image Editing
- Author
-
Tang, Chuanming, Wang, Kai, Yang, Fei, van de Weijer, Joost, Tang, Chuanming, Wang, Kai, Yang, Fei, and van de Weijer, Joost
- Abstract
Large-scale Text-to-Image (T2I) diffusion models demonstrate significant generation capabilities based on textual prompts. Based on the T2I diffusion models, text-guided image editing research aims to empower users to manipulate generated images by altering the text prompts. However, existing image editing techniques are prone to editing over unintentional regions that are beyond the intended target area, primarily due to inaccuracies in cross-attention maps. To address this problem, we propose Localization-aware Inversion (LocInv), which exploits segmentation maps or bounding boxes as extra localization priors to refine the cross-attention maps in the denoising phases of the diffusion process. Through the dynamic updating of tokens corresponding to noun words in the textual input, we are compelling the cross-attention maps to closely align with the correct noun and adjective words in the text prompt. Based on this technique, we achieve fine-grained image editing over particular objects while preventing undesired changes to other regions. Our method LocInv, based on the publicly available Stable Diffusion, is extensively evaluated on a subset of the COCO dataset, and consistently obtains superior results both quantitatively and qualitatively.The code will be released at https://github.com/wangkai930418/DPL, Comment: Accepted by CVPR 2024 Workshop AI4CC
- Published
- 2024
10. The male bias can be attenuated in reading: on the resolution of anaphoric expressions following gender-fair forms in French
- Author
-
Tibblin, Julia, Tibblin, Julia, Granfeldt, Jonas, van de Weijer, Joost, Gygax, Pascal, Tibblin, Julia, Tibblin, Julia, Granfeldt, Jonas, van de Weijer, Joost, and Gygax, Pascal
- Abstract
Despite the increased use of different types of gender-fair forms in French, studies investigating how they are interpreted when presented in a sentence remain few. To fill this gap, we conducted a pre-registered study using a timed sentence evaluation task to examine the possibility of speakers’ establishing an anaphoric relationship between a gendered anaphoric expression (femmes ‘women’ or hommes ‘men’) and non-stereotyped role nouns as antecedents. The antecedents were presented in their masculine form or in one out of three different gender-fair forms (complete double forms: les voisines et voisins ‘the neighbours.FEM and neighbours.MASC’, contracted double forms: les voisin·es ‘the neighbours.MASC·FEM’, or gender-neutral forms: le voisinage ‘the neighbourhood’). In line with previous findings, the masculine form led to a male bias in the participants’ mental representations of gender. All three examined gender-fair forms resolved this bias, but comparisons of the different forms to each other revealed differences between them. The results show that complete double forms lead to equally balanced mental representations of gender while contracted double forms slightly favour representation of women. Finally, gender-neutral forms result in a small male bias, although significantly smaller than the one produced by the masculine form. The results are discussed in relation to the mental models theory and provide new and important insights on how gender-fair forms in French are interpreted. 
- Published
- 2023
11. 3D-Aware Multi-Class Image-to-Image Translation with NeRFs
- Author
-
Li, Senmao, van de Weijer, Joost, Wang, Yaxing, Khan, Fahad, Liu, Meiqin, Yang, Jian, Li, Senmao, van de Weijer, Joost, Wang, Yaxing, Khan, Fahad, Liu, Meiqin, and Yang, Jian
- Abstract
Recent advances in 3D-aware generative models (3D-aware GANs) combined with Neural Radiance Fields (NeRF) have achieved impressive results. However no prior works investigate 3D-aware GANs for 3D consistent multi-class image-to-image (3D-aware I2I) translation. Naively using 2D-I2I translation methods suffers from unrealistic shape/identity change. To perform 3D-aware multi-class I2I translation, we decouple this learning process into a multi-class 3D-aware GAN step and a 3D-aware I2I translation step. In the first step, we propose two novel techniques: a new conditional architecture and an effective training strategy. In the second step, based on the well-trained multi-class 3D-aware GAN architecture, that preserves view-consistency, we construct a 3D-aware I2I translation system. To further reduce the view-consistency problems, we propose several new techniques, including a U-net-like adaptor network design, a hierarchical representation constrain and a relative regularization loss. In extensive experiments on two datasets, quantitative and qualitative results demonstrate that we successfully perform 3D-aware I2I translation with multi-view consistency. Code is available in 3DI2I., Funding Agencies|Key Laboratory of Advanced Information Science and Network Technology of Beijing [XDXX2202]; Youth Foundation [62202243]; Spanish Government [PID2019-104174GB-I00, TED2021-132513B-I00]
- Published
- 2023
- Full Text
- View/download PDF
12. MineGAN plus plus : Mining Generative Models for Efficient Knowledge Transfer to Limited Data Domains
- Author
-
Wang, Yaxing, Gonzalez-Garcia, Abel, Wu, Chenshen, Herranz, Luis, Khan, Fahad, Jui, Shangling, Yang, Jian, van de Weijer, Joost, Wang, Yaxing, Gonzalez-Garcia, Abel, Wu, Chenshen, Herranz, Luis, Khan, Fahad, Jui, Shangling, Yang, Jian, and van de Weijer, Joost
- Abstract
Given the often enormous effort required to train GANs, both computationally as well as in dataset collection, the re-use of pretrained GANs largely increases the potential impact of generative models. Therefore, we propose a novel knowledge transfer method for generative models based on mining the knowledge that is most beneficial to a specific target domain, either from a single or multiple pretrained GANs. This is done using a miner network that identifies which part of the generative distribution of each pretrained GAN outputs samples closest to the target domain. Mining effectively steers GAN sampling towards suitable regions of the latent space, which facilitates the posterior finetuning and avoids pathologies of other methods, such as mode collapse and lack of flexibility. Furthermore, to prevent overfitting on small target domains, we introduce sparse subnetwork selection, that restricts the set of trainable neurons to those that are relevant for the target dataset. We perform comprehensive experiments on several challenging datasets using various GAN architectures (BigGAN, Progressive GAN, and StyleGAN) and show that the proposed method, called MineGAN, effectively transfers knowledge to domains with few target images, outperforming existing methods. In addition, MineGAN can successfully transfer knowledge from multiple pretrained GANs. MineGAN., Funding Agencies|Huawei Kirin Solution; MCIN/AEI [PID2019-104174GB-I00, PID2021-128178OB-I00]; ERDF A way of making Europe; Ramon y Cajal fellowship - MCIN/AEI [RYC2019-027020-I]; CERCA Programme of Generalitat de Catalunya; Youth Foundation [62202243]
- Published
- 2023
- Full Text
- View/download PDF
13. ICICLE: Interpretable Class Incremental Continual Learning
- Author
-
Rymarczyk, Dawid, van de Weijer, Joost, Zieliński, Bartosz, Twardowski, Bartłomiej, Rymarczyk, Dawid, van de Weijer, Joost, Zieliński, Bartosz, and Twardowski, Bartłomiej
- Abstract
Continual learning enables incremental learning of new tasks without forgetting those previously learned, resulting in positive knowledge transfer that can enhance performance on both new and old tasks. However, continual learning poses new challenges for interpretability, as the rationale behind model predictions may change over time, leading to interpretability concept drift. We address this problem by proposing Interpretable Class-InCremental LEarning (ICICLE), an exemplar-free approach that adopts a prototypical part-based approach. It consists of three crucial novelties: interpretability regularization that distills previously learned concepts while preserving user-friendly positive reasoning; proximity-based prototype initialization strategy dedicated to the fine-grained setting; and task-recency bias compensation devoted to prototypical parts. Our experimental results demonstrate that ICICLE reduces the interpretability concept drift and outperforms the existing exemplar-free methods of common class-incremental learning when applied to concept-based models., Comment: Accepted to ICCV 2023
- Published
- 2023
14. Towards Label-Efficient Incremental Learning: A Survey
- Author
-
Kilickaya, Mert, van de Weijer, Joost, Asano, Yuki M., Kilickaya, Mert, van de Weijer, Joost, and Asano, Yuki M.
- Abstract
The current dominant paradigm when building a machine learning model is to iterate over a dataset over and over until convergence. Such an approach is non-incremental, as it assumes access to all images of all categories at once. However, for many applications, non-incremental learning is unrealistic. To that end, researchers study incremental learning, where a learner is required to adapt to an incoming stream of data with a varying distribution while preventing forgetting of past knowledge. Significant progress has been made, however, the vast majority of works focus on the fully supervised setting, making these algorithms label-hungry thus limiting their real-life deployment. To that end, in this paper, we make the first attempt to survey recently growing interest in label-efficient incremental learning. We identify three subdivisions, namely semi-, few-shot- and self-supervised learning to reduce labeling efforts. Finally, we identify novel directions that can further enhance label-efficiency and improve incremental learning scalability. Project website: https://github.com/kilickaya/label-efficient-il.
- Published
- 2023
15. Density Map Distillation for Incremental Object Counting
- Author
-
Wu, Chenshen, van de Weijer, Joost, Wu, Chenshen, and van de Weijer, Joost
- Abstract
We investigate the problem of incremental learning for object counting, where a method must learn to count a variety of object classes from a sequence of datasets. A na\"ive approach to incremental object counting would suffer from catastrophic forgetting, where it would suffer from a dramatic performance drop on previous tasks. In this paper, we propose a new exemplar-free functional regularization method, called Density Map Distillation (DMD). During training, we introduce a new counter head for each task and introduce a distillation loss to prevent forgetting of previous tasks. Additionally, we introduce a cross-task adaptor that projects the features of the current backbone to the previous backbone. This projector allows for the learning of new features while the backbone retains the relevant features for previous tasks. Finally, we set up experiments of incremental learning for counting new objects. Results confirm that our method greatly reduces catastrophic forgetting and outperforms existing methods., Comment: Accepted at CVPR2023: Workshop on Continual Learning in Computer Vision
- Published
- 2023
16. Projected Latent Distillation for Data-Agnostic Consolidation in Distributed Continual Learning
- Author
-
Carta, Antonio, Cossu, Andrea, Lomonaco, Vincenzo, Bacciu, Davide, van de Weijer, Joost, Carta, Antonio, Cossu, Andrea, Lomonaco, Vincenzo, Bacciu, Davide, and van de Weijer, Joost
- Abstract
Distributed learning on the edge often comprises self-centered devices (SCD) which learn local tasks independently and are unwilling to contribute to the performance of other SDCs. How do we achieve forward transfer at zero cost for the single SCDs? We formalize this problem as a Distributed Continual Learning scenario, where SCD adapt to local tasks and a CL model consolidates the knowledge from the resulting stream of models without looking at the SCD's private data. Unfortunately, current CL methods are not directly applicable to this scenario. We propose Data-Agnostic Consolidation (DAC), a novel double knowledge distillation method that consolidates the stream of SC models without using the original data. DAC performs distillation in the latent space via a novel Projected Latent Distillation loss. Experimental results show that DAC enables forward transfer between SCDs and reaches state-of-the-art accuracy on Split CIFAR100, CORe50 and Split TinyImageNet, both in reharsal-free and distributed CL scenarios. Somewhat surprisingly, even a single out-of-distribution image is sufficient as the only source of data during consolidation.
- Published
- 2023
17. 3D-Aware Multi-Class Image-to-Image Translation with NeRFs
- Author
-
Li, Senmao, van de Weijer, Joost, Wang, Yaxing, Khan, Fahad Shahbaz, Liu, Meiqin, Yang, Jian, Li, Senmao, van de Weijer, Joost, Wang, Yaxing, Khan, Fahad Shahbaz, Liu, Meiqin, and Yang, Jian
- Abstract
Recent advances in 3D-aware generative models (3D-aware GANs) combined with Neural Radiance Fields (NeRF) have achieved impressive results. However no prior works investigate 3D-aware GANs for 3D consistent multi-class image-to-image (3D-aware I2I) translation. Naively using 2D-I2I translation methods suffers from unrealistic shape/identity change. To perform 3D-aware multi-class I2I translation, we decouple this learning process into a multi-class 3D-aware GAN step and a 3D-aware I2I translation step. In the first step, we propose two novel techniques: a new conditional architecture and an effective training strategy. In the second step, based on the well-trained multi-class 3D-aware GAN architecture, that preserves view-consistency, we construct a 3D-aware I2I translation system. To further reduce the view-consistency problems, we propose several new techniques, including a U-net-like adaptor network design, a hierarchical representation constrain and a relative regularization loss. In extensive experiments on two datasets, quantitative and qualitative results demonstrate that we successfully perform 3D-aware I2I translation with multi-view consistency., Comment: Accepted by CVPR2023
- Published
- 2023
18. StyleDiffusion: Prompt-Embedding Inversion for Text-Based Editing
- Author
-
Li, Senmao, van de Weijer, Joost, Hu, Taihang, Khan, Fahad Shahbaz, Hou, Qibin, Wang, Yaxing, Yang, Jian, Li, Senmao, van de Weijer, Joost, Hu, Taihang, Khan, Fahad Shahbaz, Hou, Qibin, Wang, Yaxing, and Yang, Jian
- Abstract
A significant research effort is focused on exploiting the amazing capacities of pretrained diffusion models for the editing of images. They either finetune the model, or invert the image in the latent space of the pretrained model. However, they suffer from two problems: (1) Unsatisfying results for selected regions, and unexpected changes in nonselected regions. (2) They require careful text prompt editing where the prompt should include all visual objects in the input image. To address this, we propose two improvements: (1) Only optimizing the input of the value linear network in the cross-attention layers, is sufficiently powerful to reconstruct a real image. (2) We propose attention regularization to preserve the object-like attention maps after editing, enabling us to obtain accurate style editing without invoking significant structural changes. We further improve the editing technique which is used for the unconditional branch of classifier-free guidance, as well as the conditional one as used by P2P. Extensive experimental prompt-editing results on a variety of images, demonstrate qualitatively and quantitatively that our method has superior editing capabilities than existing and concurrent works.
- Published
- 2023
19. Brain responses to negated and affirmative meanings in the auditory modality
- Author
-
Farshchi, Sara, Andersson, Annika, van de Weijer, Joost, Paradis, Carita, Farshchi, Sara, Andersson, Annika, van de Weijer, Joost, and Paradis, Carita
- Abstract
Negation is frequently used in natural language, yet relatively little is known about its processing. More importantly, what is known regarding the neurophysiological processing of negation is mostly based on results of studies using written stimuli (the word-by-word paradigm). While the results of these studies have suggested processing costs in connection to negation (increased negativities in brain responses), it is difficult to know how this translates into processing of spoken language. We therefore developed an auditory paradigm based on a previous visual study investigating processing of affirmatives, sentential negation (not), and prefixal negation (un-). The findings of processing costs were replicated but differed in the details. Importantly, the pattern of ERP effects suggested less effortful processing for auditorily presented negated forms (restricted to increased anterior and posterior positivities) in comparison to visually presented negated forms. We suggest that the natural flow of spoken language reduces variability in processing and therefore results in clearer ERP patterns.
- Published
- 2023
- Full Text
- View/download PDF
20. There are more women in joggeur·euses than in joggeurs : On the effects of gender-fair forms on perceived gender ratios in French role nouns
- Author
-
Tibblin, Julia, van de Weijer, Joost, Granfeldt, Jonas, Gygax, Pascal, Tibblin, Julia, van de Weijer, Joost, Granfeldt, Jonas, and Gygax, Pascal
- Abstract
The present paper reports findings from a controlled large-scale (N = 1018) experimental study investigating how four different gender-fair forms influenced native French speakers’ estimated percentage of women compared to the masculine form (interpretable as generic) in 22 non-stereotyped French role nouns. The findings show that the masculine form generated lower perceived percentages of women compared to all other tested forms. In addition, gender-neutral and double forms were found equally efficient in resolving the male bias induced by the masculine form. Since the role nouns were non-stereotyped in terms of gender, these results suggest that the actual form of a role noun has indeed a strong influence on how the gender ratio of that role noun will be perceived. Moreover, the direction of the questionnaire’s response scale had a significant effect on the results, which entails methodological implications for future research. Finally, the provided ratios can be used for future studies investigating French role nouns in different gender-fair forms. In sum, our study suggests that gender-fair forms in French are an efficient tool for increasing the visibility of women, at least in nouns representing non-stereotypical activities.
- Published
- 2023
21. The First Visual Object Tracking Segmentation VOTS2023 Challenge Results
- Author
-
Kristan, Matej, Matas, Jiri, Danelljan, Martin, Felsberg, Michael, Chang, Hyung Jin, Zajc, Luka Cehovin, Lukezic, Alan, Drbohlav, Ondrej, Zhang, Zhongqun, Tran, Khanh-Tung, Vu, Xuan-Son, Bjorklund, Johanna, Mayer, Christoph, Zhang, Yushan, Ke, Lei, Zhao, Jie, Fernandez, Gustavo, Al-Shakarji, Noor, An, Dong, Arens, Michael, Becker, Stefan, Bhat, Goutam, Bullinger, Sebastian, Chan, Antoni B., Chang, Shijie, Chen, Hanyuan, Chen, Xin, Chen, Yan, Chen, Zhenyu, Cheng, Yangming, Cui, Yutao, Deng, Chunyuan, Dong, Jiahua, Dunnhofer, Matteo, Feng, Wei, Fu, Jianlong, Gao, Jie, Han, Ruize, Hao, Zeqi, He, Jun-Yan, He, Keji, He, Zhenyu, Hu, Xiantao, Huang, Kaer, Huang, Yuqing, Jiang, Yi, Kang, Ben, Lan, Jin-Peng, Lee, Hyungjun, Li, Chenyang, Li, Jiahao, Li, Ning, Li, Wangkai, Li, Xiaodi, Li, Xin, Liu, Pengyu, Liu, Yue, Lu, Huchuan, Luo, Bin, Luo, Ping, Ma, Yinchao, Miao, Deshui, Micheloni, Christian, Palaniappan, Kannappan, Park, Hancheol, Paul, Matthieu, Peng, HouWen, Qian, Zekun, Rahmon, Gani, Scherer-Negenborn, Norbert, Shao, Pengcheng, Shin, Wooksu, Kazemi, Elham Soltani, Song, Tianhui, Stiefelhagen, Rainer, Sun, Rui, Tang, Chuanming, Tang, Zhangyong, Toubal, Imad Eddine, Valmadre, Jack, van de Weijer, Joost, Van Gool, Luc, Vira, Jash, Vujasinovic, Stephane, Wan, Cheng, Wan, Jia, Wang, Dong, Wang, Fei, Wang, Feifan, Wang, He, Wang, Limin, Wang, Song, Wang, Yaowei, Wang, Zhepeng, Wu, Gangshan, Wu, Jiannan, Wu, Qiangqiang, Wu, Xiaojun, Xiao, Anqi, Xie, Jinxia, Xu, Chenlong, Xu, Min, Xu, Tianyang, Xu, Yuanyou, Yan, Bin, Yang, Dawei, Yang, Ming-Hsuan, Yang, Tianyu, Yang, Yi, Yang, Zongxin, Yin, Xuanwu, Yu, Fisher, Yu, Hongyuan, Yu, Qianjin, Yu, Weichen, Yuan, YongSheng, Yuan, Zehuan, Zhang, Jianlin, Zhang, Lu, Zhang, Tianzhu, Zhao, Guodongfang, Zhao, Shaochuan, Zheng, Yaozong, Zhong, Bineng, Zhu, Jiawen, Zhu, Xuefeng, Zhuang, Yueting, Zong, ChengAo, Zuo, Kunlong, Kristan, Matej, Matas, Jiri, Danelljan, Martin, Felsberg, Michael, Chang, Hyung Jin, Zajc, Luka Cehovin, Lukezic, Alan, Drbohlav, Ondrej, Zhang, Zhongqun, Tran, Khanh-Tung, Vu, Xuan-Son, Bjorklund, Johanna, Mayer, Christoph, Zhang, Yushan, Ke, Lei, Zhao, Jie, Fernandez, Gustavo, Al-Shakarji, Noor, An, Dong, Arens, Michael, Becker, Stefan, Bhat, Goutam, Bullinger, Sebastian, Chan, Antoni B., Chang, Shijie, Chen, Hanyuan, Chen, Xin, Chen, Yan, Chen, Zhenyu, Cheng, Yangming, Cui, Yutao, Deng, Chunyuan, Dong, Jiahua, Dunnhofer, Matteo, Feng, Wei, Fu, Jianlong, Gao, Jie, Han, Ruize, Hao, Zeqi, He, Jun-Yan, He, Keji, He, Zhenyu, Hu, Xiantao, Huang, Kaer, Huang, Yuqing, Jiang, Yi, Kang, Ben, Lan, Jin-Peng, Lee, Hyungjun, Li, Chenyang, Li, Jiahao, Li, Ning, Li, Wangkai, Li, Xiaodi, Li, Xin, Liu, Pengyu, Liu, Yue, Lu, Huchuan, Luo, Bin, Luo, Ping, Ma, Yinchao, Miao, Deshui, Micheloni, Christian, Palaniappan, Kannappan, Park, Hancheol, Paul, Matthieu, Peng, HouWen, Qian, Zekun, Rahmon, Gani, Scherer-Negenborn, Norbert, Shao, Pengcheng, Shin, Wooksu, Kazemi, Elham Soltani, Song, Tianhui, Stiefelhagen, Rainer, Sun, Rui, Tang, Chuanming, Tang, Zhangyong, Toubal, Imad Eddine, Valmadre, Jack, van de Weijer, Joost, Van Gool, Luc, Vira, Jash, Vujasinovic, Stephane, Wan, Cheng, Wan, Jia, Wang, Dong, Wang, Fei, Wang, Feifan, Wang, He, Wang, Limin, Wang, Song, Wang, Yaowei, Wang, Zhepeng, Wu, Gangshan, Wu, Jiannan, Wu, Qiangqiang, Wu, Xiaojun, Xiao, Anqi, Xie, Jinxia, Xu, Chenlong, Xu, Min, Xu, Tianyang, Xu, Yuanyou, Yan, Bin, Yang, Dawei, Yang, Ming-Hsuan, Yang, Tianyu, Yang, Yi, Yang, Zongxin, Yin, Xuanwu, Yu, Fisher, Yu, Hongyuan, Yu, Qianjin, Yu, Weichen, Yuan, YongSheng, Yuan, Zehuan, Zhang, Jianlin, Zhang, Lu, Zhang, Tianzhu, Zhao, Guodongfang, Zhao, Shaochuan, Zheng, Yaozong, Zhong, Bineng, Zhu, Jiawen, Zhu, Xuefeng, Zhuang, Yueting, Zong, ChengAo, and Zuo, Kunlong
- Abstract
The Visual Object Tracking Segmentation VOTS2023 challenge is the eleventh annual tracker benchmarking activity of the VOT initiative. This challenge is the first to merge short-term and long-term as well as single-target and multiple-target tracking with segmentation masks as the only target location specification. A new dataset was created; the ground truth has been withheld to prevent overfitting. New performance measures and evaluation protocols have been created along with a new toolkit and an evaluation server. Results of the presented 47 trackers indicate that modern tracking frameworks are well-suited to deal with convergence of short-term and long-term tracking and that multiple and single target tracking can be considered a single problem. A leaderboard, with participating trackers details, the source code, the datasets, and the evaluation kit are publicly available at the challenge website(1).
- Published
- 2023
- Full Text
- View/download PDF
22. Generative Multi-Label Zero-Shot Learning
- Author
-
Gupta, Akshita, Narayan, Sanath, Khan, Salman, Khan, Fahad, Shao, Ling, van de Weijer, Joost, Gupta, Akshita, Narayan, Sanath, Khan, Salman, Khan, Fahad, Shao, Ling, and van de Weijer, Joost
- Abstract
Multi-label zero-shot learning strives to classify images into multiple unseen categories for which no data is available during training. The test samples can additionally contain seen categories in the generalized variant. Existing approaches rely on learning either shared or label-specific attention from the seen classes. Nevertheless, computing reliable attention maps for unseen classes during inference in a multi-label setting is still a challenge. In contrast, state-of-the-art single-label generative adversarial network (GAN) based approaches learn to directly synthesize the class-specific visual features from the corresponding class attribute embeddings. However, synthesizing multi-label features from GANs is still unexplored in the context of zero-shot setting. When multiple objects occur jointly in a single image, a critical question is how to effectively fuse multi-class information. In this work, we introduce different fusion approaches at the attribute-level, feature-level and cross-level (across attribute and feature-levels) for synthesizing multi-label features from their corresponding multi-label class embeddings. To the best of our knowledge, our work is the first to tackle the problem of multi-label feature synthesis in the (generalized) zero-shot setting. Our cross-level fusion-based generative approach outperforms the state-of-the-art on three zero-shot benchmarks: NUS-WIDE, Open Images and MS COCO. Furthermore, we show the generalization capabilities of our fusion approach in the zero-shot detection task on MS COCO, achieving favorable performance against existing methods., Funding Agencies| [PID2021-128178OB-I00]
- Published
- 2023
- Full Text
- View/download PDF
23. Matching on action:Effects of action speed and viewpoint on perceived continuity across match-action film edits
- Author
-
Langkjær, Birger, Gregersen, Andreas Lindegaard, Rédei, Anna Cabak, van de Weijer, Joost, Innes-Ker, Åse, Langkjær, Birger, Gregersen, Andreas Lindegaard, Rédei, Anna Cabak, van de Weijer, Joost, and Innes-Ker, Åse
- Abstract
A match-action cut in feature films connects two shots of a single continuous movement. This type of editing often goes unnoticed and is arguably the most effective form of continuity editing. However, the literature offers little agreement on editing best practice and, by implication, on how our perceptual system deals with disjointed moving images. Studies have suggested that frames should overlap across the cut for the viewer to experience continuity, but also that leaving out frames is preferable, and even that viewers are unable to discriminate such detail. We conducted an experiment to investigate viewer preferences for match-action cuts, using type of cut as well as velocity of movement as predictors and number of overlapping/elliptical frames as the outcome variable. Thirty-nine participants determined the smoothest cut in eight film-clips in a within-subjects design. Surprisingly, we found that average viewer preferences were less than a single frame from a straight cut for all cut-types. We also found that velocity had a small but statistically significant effect on editing preferences. The preference for straight cuts found in the present study runs counter to the idea that perceived continuity across match-action cuts requires objective dis-continuity and suggests that the straight cut provides a simple rule of thumb for film editors. In addition, we interpret the conflicting results from previous studies together with our own findings based on a discrimination task of finding the optimal cut as indicating that human visual perception allows for a window of acceptable continuity cuts centered around the straight cut.
- Published
- 2023
24. Enhancing Perceptual Quality in Video Super-Resolution through Temporally-Consistent Detail Synthesis using Diffusion Models
- Author
-
Rota, Claudio, Buzzelli, Marco, van de Weijer, Joost, Rota, Claudio, Buzzelli, Marco, and van de Weijer, Joost
- Abstract
In this paper, we address the problem of video super-resolution (VSR) using Diffusion Models (DM), and present StableVSR. Our method significantly enhances the perceptual quality of upscaled videos by synthesizing realistic and temporally-consistent details. We turn a pre-trained DM for single image super-resolution into a VSR method by introducing the Temporal Conditioning Module (TCM). TCM uses Temporal Texture Guidance, which provides spatially-aligned and detail-rich texture information synthesized in adjacent frames. This guides the generative process of the current frame toward high-quality and temporally-consistent results. We introduce a Frame-wise Bidirectional Sampling strategy to encourage the use of information from past to future and vice-versa. This strategy improves the perceptual quality of the results and the temporal consistency across frames. We demonstrate the effectiveness of StableVSR in enhancing the perceptual quality of upscaled videos compared to existing state-of-the-art methods for VSR. The code is available at https://github.com/claudiom4sir/StableVSR.
- Published
- 2023
25. Continual Learning: Applications and the Road Forward
- Author
-
Verwimp, Eli, Aljundi, Rahaf, Ben-David, Shai, Bethge, Matthias, Cossu, Andrea, Gepperth, Alexander, Hayes, Tyler L., Hüllermeier, Eyke, Kanan, Christopher, Kudithipudi, Dhireesha, Lampert, Christoph H., Mundt, Martin, Pascanu, Razvan, Popescu, Adrian, Tolias, Andreas S., van de Weijer, Joost, Liu, Bing, Lomonaco, Vincenzo, Tuytelaars, Tinne, van de Ven, Gido M., Verwimp, Eli, Aljundi, Rahaf, Ben-David, Shai, Bethge, Matthias, Cossu, Andrea, Gepperth, Alexander, Hayes, Tyler L., Hüllermeier, Eyke, Kanan, Christopher, Kudithipudi, Dhireesha, Lampert, Christoph H., Mundt, Martin, Pascanu, Razvan, Popescu, Adrian, Tolias, Andreas S., van de Weijer, Joost, Liu, Bing, Lomonaco, Vincenzo, Tuytelaars, Tinne, and van de Ven, Gido M.
- Abstract
Continual learning is a subfield of machine learning, which aims to allow machine learning models to continuously learn on new data, by accumulating knowledge without forgetting what was learned in the past. In this work, we take a step back, and ask: "Why should one care about continual learning in the first place?". We set the stage by examining recent continual learning papers published at four major machine learning conferences, and show that memory-constrained settings dominate the field. Then, we discuss five open problems in machine learning, and even though they might seem unrelated to continual learning at first sight, we show that continual learning will inevitably be part of their solution. These problems are model editing, personalization and specialization, on-device learning, faster (re-)training and reinforcement learning. Finally, by comparing the desiderata from these unsolved problems and the current assumptions in continual learning, we highlight and discuss four future directions for continual learning research. We hope that this work offers an interesting perspective on the future of continual learning, while displaying its potential value and the paths we have to pursue in order to make it successful. This work is the result of the many discussions the authors had at the Dagstuhl seminar on Deep Continual Learning, in March 2023.
- Published
- 2023
26. IterInv: Iterative Inversion for Pixel-Level T2I Models
- Author
-
Tang, Chuanming, Wang, Kai, van de Weijer, Joost, Tang, Chuanming, Wang, Kai, and van de Weijer, Joost
- Abstract
Large-scale text-to-image diffusion models have been a ground-breaking development in generating convincing images following an input text prompt. The goal of image editing research is to give users control over the generated images by modifying the text prompt. Current image editing techniques predominantly hinge on DDIM inversion as a prevalent practice rooted in Latent Diffusion Models (LDM). However, the large pretrained T2I models working on the latent space suffer from losing details due to the first compression stage with an autoencoder mechanism. Instead, other mainstream T2I pipeline working on the pixel level, such as Imagen and DeepFloyd-IF, circumvents the above problem. They are commonly composed of multiple stages, typically starting with a text-to-image stage and followed by several super-resolution stages. In this pipeline, the DDIM inversion fails to find the initial noise and generate the original image given that the super-resolution diffusion models are not compatible with the DDIM technique. According to our experimental findings, iteratively concatenating the noisy image as the condition is the root of this problem. Based on this observation, we develop an iterative inversion (IterInv) technique for this category of T2I models and verify IterInv with the open-source DeepFloyd-IF model.Specifically, IterInv employ NTI as the inversion and reconstruction of low-resolution image generation. In stages 2 and 3, we update the latent variance at each timestep to find the deterministic inversion trace and promote the reconstruction process. By combining our method with a popular image editing method, we prove the application prospects of IterInv. The code will be released upon acceptance. The code is available at \url{https://github.com/Tchuanm/IterInv.git}., Comment: Accepted paper at ICME 2024
- Published
- 2023
27. Exploiting Image-Related Inductive Biases in Single-Branch Visual Tracking
- Author
-
Tang, Chuanming, Wang, Kai, van de Weijer, Joost, Zhang, Jianlin, Huang, Yongmei, Tang, Chuanming, Wang, Kai, van de Weijer, Joost, Zhang, Jianlin, and Huang, Yongmei
- Abstract
Despite achieving state-of-the-art performance in visual tracking, recent single-branch trackers tend to overlook the weak prior assumptions associated with the Vision Transformer (ViT) encoder and inference pipeline. Moreover, the effectiveness of discriminative trackers remains constrained due to the adoption of the dual-branch pipeline. To tackle the inferior effectiveness of the vanilla ViT, we propose an Adaptive ViT Model Prediction tracker (AViTMP) to bridge the gap between single-branch network and discriminative models. Specifically, in the proposed encoder AViT-Enc, we introduce an adaptor module and joint target state embedding to enrich the dense embedding paradigm based on ViT. Then, we combine AViT-Enc with a dense-fusion decoder and a discriminative target model to predict accurate location. Further, to mitigate the limitations of conventional inference practice, we present a novel inference pipeline called CycleTrack, which bolsters the tracking robustness in the presence of distractors via bidirectional cycle tracking verification. Lastly, we propose a dual-frame update inference strategy that adeptively handles significant challenges in long-term scenarios. In the experiments, we evaluate AViTMP on ten tracking benchmarks for a comprehensive assessment, including LaSOT, LaSOTExtSub, AVisT, etc. The experimental results unequivocally establish that AViTMP attains state-of-the-art performance, especially on long-time tracking and robustness., Comment: 13 pages, 8 figures, under review
- Published
- 2023
28. Technical Report for ICCV 2023 Visual Continual Learning Challenge: Continuous Test-time Adaptation for Semantic Segmentation
- Author
-
Sójka, Damian, Liu, Yuyang, Goswami, Dipam, Cygert, Sebastian, Twardowski, Bartłomiej, van de Weijer, Joost, Sójka, Damian, Liu, Yuyang, Goswami, Dipam, Cygert, Sebastian, Twardowski, Bartłomiej, and van de Weijer, Joost
- Abstract
The goal of the challenge is to develop a test-time adaptation (TTA) method, which could adapt the model to gradually changing domains in video sequences for semantic segmentation task. It is based on a synthetic driving video dataset - SHIFT. The source model is trained on images taken during daytime in clear weather. Domain changes at test-time are mainly caused by varying weather conditions and times of day. The TTA methods are evaluated in each image sequence (video) separately, meaning the model is reset to the source model state before the next sequence. Images come one by one and a prediction has to be made at the arrival of each frame. Each sequence is composed of 401 images and starts with the source domain, then gradually drifts to a different one (changing weather or time of day) until the middle of the sequence. In the second half of the sequence, the domain gradually shifts back to the source one. Ground truth data is available only for the validation split of the SHIFT dataset, in which there are only six sequences that start and end with the source domain. We conduct an analysis specifically on those sequences. Ground truth data for test split, on which the developed TTA methods are evaluated for leader board ranking, are not publicly available. The proposed solution secured a 3rd place in a challenge and received an innovation award. Contrary to the solutions that scored better, we did not use any external pretrained models or specialized data augmentations, to keep the solutions as general as possible. We have focused on analyzing the distributional shift and developing a method that could adapt to changing data dynamics and generalize across different scenarios.
- Published
- 2023
29. Dynamic Prompt Learning: Addressing Cross-Attention Leakage for Text-Based Image Editing
- Author
-
Wang, Kai, Yang, Fei, Yang, Shiqi, Butt, Muhammad Atif, van de Weijer, Joost, Wang, Kai, Yang, Fei, Yang, Shiqi, Butt, Muhammad Atif, and van de Weijer, Joost
- Abstract
Large-scale text-to-image generative models have been a ground-breaking development in generative AI, with diffusion models showing their astounding ability to synthesize convincing images following an input text prompt. The goal of image editing research is to give users control over the generated images by modifying the text prompt. Current image editing techniques are susceptible to unintended modifications of regions outside the targeted area, such as on the background or on distractor objects which have some semantic or visual relationship with the targeted object. According to our experimental findings, inaccurate cross-attention maps are at the root of this problem. Based on this observation, we propose Dynamic Prompt Learning (DPL) to force cross-attention maps to focus on correct noun words in the text prompt. By updating the dynamic tokens for nouns in the textual input with the proposed leakage repairment losses, we achieve fine-grained image editing over particular objects while preventing undesired changes to other image regions. Our method DPL, based on the publicly available Stable Diffusion, is extensively evaluated on a wide range of images, and consistently obtains superior results both quantitatively (CLIP score, Structure-Dist) and qualitatively (on user-evaluation). We show improved prompt editing results for Word-Swap, Prompt Refinement, and Attention Re-weighting, especially for complex multi-object scenes., Comment: Neurips 2023. The code page: https://github.com/wangkai930418/DPL
- Published
- 2023
30. FeCAM: Exploiting the Heterogeneity of Class Distributions in Exemplar-Free Continual Learning
- Author
-
Goswami, Dipam, Liu, Yuyang, Twardowski, Bartłomiej, van de Weijer, Joost, Goswami, Dipam, Liu, Yuyang, Twardowski, Bartłomiej, and van de Weijer, Joost
- Abstract
Exemplar-free class-incremental learning (CIL) poses several challenges since it prohibits the rehearsal of data from previous tasks and thus suffers from catastrophic forgetting. Recent approaches to incrementally learning the classifier by freezing the feature extractor after the first task have gained much attention. In this paper, we explore prototypical networks for CIL, which generate new class prototypes using the frozen feature extractor and classify the features based on the Euclidean distance to the prototypes. In an analysis of the feature distributions of classes, we show that classification based on Euclidean metrics is successful for jointly trained features. However, when learning from non-stationary data, we observe that the Euclidean metric is suboptimal and that feature distributions are heterogeneous. To address this challenge, we revisit the anisotropic Mahalanobis distance for CIL. In addition, we empirically show that modeling the feature covariance relations is better than previous attempts at sampling features from normal distributions and training a linear classifier. Unlike existing methods, our approach generalizes to both many- and few-shot CIL settings, as well as to domain-incremental settings. Interestingly, without updating the backbone network, our method obtains state-of-the-art results on several standard continual learning benchmarks. Code is available at https://github.com/dipamgoswami/FeCAM., Comment: Accepted at NeurIPS 2023
- Published
- 2023
31. Plasticity-Optimized Complementary Networks for Unsupervised Continual Learning
- Author
-
Gomez-Villa, Alex, Twardowski, Bartlomiej, Wang, Kai, van de Weijer, Joost, Gomez-Villa, Alex, Twardowski, Bartlomiej, Wang, Kai, and van de Weijer, Joost
- Abstract
Continuous unsupervised representation learning (CURL) research has greatly benefited from improvements in self-supervised learning (SSL) techniques. As a result, existing CURL methods using SSL can learn high-quality representations without any labels, but with a notable performance drop when learning on a many-tasks data stream. We hypothesize that this is caused by the regularization losses that are imposed to prevent forgetting, leading to a suboptimal plasticity-stability trade-off: they either do not adapt fully to the incoming data (low plasticity), or incur significant forgetting when allowed to fully adapt to a new SSL pretext-task (low stability). In this work, we propose to train an expert network that is relieved of the duty of keeping the previous knowledge and can focus on performing optimally on the new tasks (optimizing plasticity). In the second phase, we combine this new knowledge with the previous network in an adaptation-retrospection phase to avoid forgetting and initialize a new expert with the knowledge of the old network. We perform several experiments showing that our proposed approach outperforms other CURL exemplar-free methods in few- and many-task split settings. Furthermore, we show how to adapt our approach to semi-supervised continual learning (Semi-SCL) and show that we surpass the accuracy of other exemplar-free Semi-SCL methods and reach the results of some others that use exemplars., Comment: Accepted at WACV2024
- Published
- 2023
32. Continual Evidential Deep Learning for Out-of-Distribution Detection
- Author
-
Aguilar, Eduardo, Raducanu, Bogdan, Radeva, Petia, Van de Weijer, Joost, Aguilar, Eduardo, Raducanu, Bogdan, Radeva, Petia, and Van de Weijer, Joost
- Abstract
Uncertainty-based deep learning models have attracted a great deal of interest for their ability to provide accurate and reliable predictions. Evidential deep learning stands out achieving remarkable performance in detecting out-of-distribution (OOD) data with a single deterministic neural network. Motivated by this fact, in this paper we propose the integration of an evidential deep learning method into a continual learning framework in order to perform simultaneously incremental object classification and OOD detection. Moreover, we analyze the ability of vacuity and dissonance to differentiate between in-distribution data belonging to old classes and OOD data. The proposed method, called CEDL, is evaluated on CIFAR-100 considering two settings consisting of 5 and 10 tasks, respectively. From the obtained results, we could appreciate that the proposed method, in addition to provide comparable results in object classification with respect to the baseline, largely outperforms OOD detection compared to several posthoc methods on three evaluation metrics: AUROC, AUPR and FPR95., Comment: Accepted at Visual Continual Learning workshop (ICCV2023)
- Published
- 2023
33. Trust your Good Friends: Source-free Domain Adaptation by Reciprocal Neighborhood Clustering
- Author
-
Yang, Shiqi, Wang, Yaxing, van de Weijer, Joost, Herranz, Luis, Jui, Shangling, Yang, Jian, Yang, Shiqi, Wang, Yaxing, van de Weijer, Joost, Herranz, Luis, Jui, Shangling, and Yang, Jian
- Abstract
Domain adaptation (DA) aims to alleviate the domain shift between source domain and target domain. Most DA methods require access to the source data, but often that is not possible (e.g. due to data privacy or intellectual property). In this paper, we address the challenging source-free domain adaptation (SFDA) problem, where the source pretrained model is adapted to the target domain in the absence of source data. Our method is based on the observation that target data, which might not align with the source domain classifier, still forms clear clusters. We capture this intrinsic structure by defining local affinity of the target data, and encourage label consistency among data with high local affinity. We observe that higher affinity should be assigned to reciprocal neighbors. To aggregate information with more context, we consider expanded neighborhoods with small affinity values. Furthermore, we consider the density around each target sample, which can alleviate the negative impact of potential outliers. In the experimental results we verify that the inherent structure of the target features is an important source of information for domain adaptation. We demonstrate that this local structure can be efficiently captured by considering the local neighbors, the reciprocal neighbors, and the expanded neighborhood. Finally, we achieve state-of-the-art performance on several 2D image and 3D point cloud recognition datasets., Comment: Accepted by IEEE TPAMI, extended version of conference paper arXiv:2110.04202
- Published
- 2023
34. ScrollNet: Dynamic Weight Importance for Continual Learning
- Author
-
Yang, Fei, Wang, Kai, van de Weijer, Joost, Yang, Fei, Wang, Kai, and van de Weijer, Joost
- Abstract
The principle underlying most existing continual learning (CL) methods is to prioritize stability by penalizing changes in parameters crucial to old tasks, while allowing for plasticity in other parameters. The importance of weights for each task can be determined either explicitly through learning a task-specific mask during training (e.g., parameter isolation-based approaches) or implicitly by introducing a regularization term (e.g., regularization-based approaches). However, all these methods assume that the importance of weights for each task is unknown prior to data exposure. In this paper, we propose ScrollNet as a scrolling neural network for continual learning. ScrollNet can be seen as a dynamic network that assigns the ranking of weight importance for each task before data exposure, thus achieving a more favorable stability-plasticity tradeoff during sequential task learning by reassigning this ranking for different tasks. Additionally, we demonstrate that ScrollNet can be combined with various CL methods, including regularization-based and replay-based approaches. Experimental results on CIFAR100 and TinyImagenet datasets show the effectiveness of our proposed method. We release our code at https://github.com/FireFYF/ScrollNet.git., Comment: Accepted at Visual Continual Learning workshop (ICCV2023)
- Published
- 2023
35. A Comprehensive Empirical Evaluation on Online Continual Learning
- Author
-
Soutif--Cormerais, Albin, Carta, Antonio, Cossu, Andrea, Hurtado, Julio, Hemati, Hamed, Lomonaco, Vincenzo, Van de Weijer, Joost, Soutif--Cormerais, Albin, Carta, Antonio, Cossu, Andrea, Hurtado, Julio, Hemati, Hamed, Lomonaco, Vincenzo, and Van de Weijer, Joost
- Abstract
Online continual learning aims to get closer to a live learning experience by learning directly on a stream of data with temporally shifting distribution and by storing a minimum amount of data from that stream. In this empirical evaluation, we evaluate various methods from the literature that tackle online continual learning. More specifically, we focus on the class-incremental setting in the context of image classification, where the learner must learn new classes incrementally from a stream of data. We compare these methods on the Split-CIFAR100 and Split-TinyImagenet benchmarks, and measure their average accuracy, forgetting, stability, and quality of the representations, to evaluate various aspects of the algorithm at the end but also during the whole training period. We find that most methods suffer from stability and underfitting issues. However, the learned representations are comparable to i.i.d. training under the same computational budget. No clear winner emerges from the results and basic experience replay, when properly tuned and implemented, is a very strong baseline. We release our modular and extensible codebase at https://github.com/AlbinSou/ocl_survey based on the avalanche framework to reproduce our results and encourage future research., Comment: ICCV Visual Continual Learning Workshop 2023 accepted paper
- Published
- 2023
36. Augmented Box Replay: Overcoming Foreground Shift for Incremental Object Detection
- Author
-
Yuyang, Liu, Yang, Cong, Dipam, Goswami, Xialei, Liu, van de Weijer, Joost, Yuyang, Liu, Yang, Cong, Dipam, Goswami, Xialei, Liu, and van de Weijer, Joost
- Abstract
In incremental learning, replaying stored samples from previous tasks together with current task samples is one of the most efficient approaches to address catastrophic forgetting. However, unlike incremental classification, image replay has not been successfully applied to incremental object detection (IOD). In this paper, we identify the overlooked problem of foreground shift as the main reason for this. Foreground shift only occurs when replaying images of previous tasks and refers to the fact that their background might contain foreground objects of the current task. To overcome this problem, a novel and efficient Augmented Box Replay (ABR) method is developed that only stores and replays foreground objects and thereby circumvents the foreground shift problem. In addition, we propose an innovative Attentive RoI Distillation loss that uses spatial attention from region-of-interest (RoI) features to constrain current model to focus on the most important information from old model. ABR significantly reduces forgetting of previous classes while maintaining high plasticity in current classes. Moreover, it considerably reduces the storage requirements when compared to standard image replay. Comprehensive experiments on Pascal-VOC and COCO datasets support the state-of-the-art performance of our model.
- Published
- 2023
37. Improving Online Continual Learning Performance and Stability with Temporal Ensembles
- Author
-
Soutif--Cormerais, Albin, Carta, Antonio, Van de Weijer, Joost, Soutif--Cormerais, Albin, Carta, Antonio, and Van de Weijer, Joost
- Abstract
Neural networks are very effective when trained on large datasets for a large number of iterations. However, when they are trained on non-stationary streams of data and in an online fashion, their performance is reduced (1) by the online setup, which limits the availability of data, (2) due to catastrophic forgetting because of the non-stationary nature of the data. Furthermore, several recent works (Caccia et al., 2022; Lange et al., 2023) arXiv:2205.13452 showed that replay methods used in continual learning suffer from the stability gap, encountered when evaluating the model continually (rather than only on task boundaries). In this article, we study the effect of model ensembling as a way to improve performance and stability in online continual learning. We notice that naively ensembling models coming from a variety of training tasks increases the performance in online continual learning considerably. Starting from this observation, and drawing inspirations from semi-supervised learning ensembling methods, we use a lightweight temporal ensemble that computes the exponential moving average of the weights (EMA) at test time, and show that it can drastically increase the performance and stability when used in combination with several methods from the literature., Comment: CoLLAs 2023 accepted paper
- Published
- 2023
38. Exemplar-free Continual Learning of Vision Transformers via Gated Class-Attention and Cascaded Feature Drift Compensation
- Author
-
Cotogni, Marco, Yang, Fei, Cusano, Claudio, Bagdanov, Andrew D., van de Weijer, Joost, Cotogni, Marco, Yang, Fei, Cusano, Claudio, Bagdanov, Andrew D., and van de Weijer, Joost
- Abstract
We propose a new method for exemplar-free class incremental training of ViTs. The main challenge of exemplar-free continual learning is maintaining plasticity of the learner without causing catastrophic forgetting of previously learned tasks. This is often achieved via exemplar replay which can help recalibrate previous task classifiers to the feature drift which occurs when learning new tasks. Exemplar replay, however, comes at the cost of retaining samples from previous tasks which for many applications may not be possible. To address the problem of continual ViT training, we first propose gated class-attention to minimize the drift in the final ViT transformer block. This mask-based gating is applied to class-attention mechanism of the last transformer block and strongly regulates the weights crucial for previous tasks. Importantly, gated class-attention does not require the task-ID during inference, which distinguishes it from other parameter isolation methods. Secondly, we propose a new method of feature drift compensation that accommodates feature drift in the backbone when learning new tasks. The combination of gated class-attention and cascaded feature drift compensation allows for plasticity towards new tasks while limiting forgetting of previous ones. Extensive experiments performed on CIFAR-100, Tiny-ImageNet and ImageNet100 demonstrate that our exemplar-free method obtains competitive results when compared to rehearsal based ViT methods.
- Published
- 2022
39. Attribution-aware Weight Transfer: A Warm-Start Initialization for Class-Incremental Semantic Segmentation
- Author
-
Goswami, Dipam, Schuster, René, van de Weijer, Joost, Stricker, Didier, Goswami, Dipam, Schuster, René, van de Weijer, Joost, and Stricker, Didier
- Abstract
In class-incremental semantic segmentation (CISS), deep learning architectures suffer from the critical problems of catastrophic forgetting and semantic background shift. Although recent works focused on these issues, existing classifier initialization methods do not address the background shift problem and assign the same initialization weights to both background and new foreground class classifiers. We propose to address the background shift with a novel classifier initialization method which employs gradient-based attribution to identify the most relevant weights for new classes from the classifier's weights for the previous background and transfers these weights to the new classifier. This warm-start weight initialization provides a general solution applicable to several CISS methods. Furthermore, it accelerates learning of new classes while mitigating forgetting. Our experiments demonstrate significant improvement in mIoU compared to the state-of-the-art CISS methods on the Pascal-VOC 2012, ADE20K and Cityscapes datasets., Comment: Accepted at WACV 2023
- Published
- 2022
40. Positive Pair Distillation Considered Harmful: Continual Meta Metric Learning for Lifelong Object Re-Identification
- Author
-
Wang, Kai, Wu, Chenshen, Bagdanov, Andy, Liu, Xialei, Yang, Shiqi, Jui, Shangling, van de Weijer, Joost, Wang, Kai, Wu, Chenshen, Bagdanov, Andy, Liu, Xialei, Yang, Shiqi, Jui, Shangling, and van de Weijer, Joost
- Abstract
Lifelong object re-identification incrementally learns from a stream of re-identification tasks. The objective is to learn a representation that can be applied to all tasks and that generalizes to previously unseen re-identification tasks. The main challenge is that at inference time the representation must generalize to previously unseen identities. To address this problem, we apply continual meta metric learning to lifelong object re-identification. To prevent forgetting of previous tasks, we use knowledge distillation and explore the roles of positive and negative pairs. Based on our observation that the distillation and metric losses are antagonistic, we propose to remove positive pairs from distillation to robustify model updates. Our method, called Distillation without Positive Pairs (DwoPP), is evaluated on extensive intra-domain experiments on person and vehicle re-identification datasets, as well as inter-domain experiments on the LReID benchmark. Our experiments demonstrate that DwoPP significantly outperforms the state-of-the-art. The code is here: https://github.com/wangkai930418/DwoPP_code, Comment: BMVC 2022
- Published
- 2022
41. Attention Distillation: self-supervised vision transformer students need more guidance
- Author
-
Wang, Kai, Yang, Fei, van de Weijer, Joost, Wang, Kai, Yang, Fei, and van de Weijer, Joost
- Abstract
Self-supervised learning has been widely applied to train high-quality vision transformers. Unleashing their excellent performance on memory and compute constraint devices is therefore an important research topic. However, how to distill knowledge from one self-supervised ViT to another has not yet been explored. Moreover, the existing self-supervised knowledge distillation (SSKD) methods focus on ConvNet based architectures are suboptimal for ViT knowledge distillation. In this paper, we study knowledge distillation of self-supervised vision transformers (ViT-SSKD). We show that directly distilling information from the crucial attention mechanism from teacher to student can significantly narrow the performance gap between both. In experiments on ImageNet-Subset and ImageNet-1K, we show that our method AttnDistill outperforms existing self-supervised knowledge distillation (SSKD) methods and achieves state-of-the-art k-NN accuracy compared with self-supervised learning (SSL) methods learning from scratch (with the ViT-S model). We are also the first to apply the tiny ViT-T model on self-supervised learning. Moreover, AttnDistill is independent of self-supervised learning algorithms, it can be adapted to ViT based SSL methods to improve the performance in future research. The code is here: https://github.com/wangkai930418/attndistill, Comment: BMVC 2022
- Published
- 2022
42. Compound internal anaphora : evidence from acceptability judgements on Italian argumental compounds
- Author
-
Lami, Irene, van de Weijer, Joost, Lami, Irene, and van de Weijer, Joost
- Published
- 2022
43. Attracting and Dispersing: A Simple Approach for Source-free Domain Adaptation
- Author
-
Yang, Shiqi, Wang, Yaxing, Wang, Kai, Jui, Shangling, van de Weijer, Joost, Yang, Shiqi, Wang, Yaxing, Wang, Kai, Jui, Shangling, and van de Weijer, Joost
- Abstract
We propose a simple but effective source-free domain adaptation (SFDA) method. Treating SFDA as an unsupervised clustering problem and following the intuition that local neighbors in feature space should have more similar predictions than other features, we propose to optimize an objective of prediction consistency. This objective encourages local neighborhood features in feature space to have similar predictions while features farther away in feature space have dissimilar predictions, leading to efficient feature clustering and cluster assignment simultaneously. For efficient training, we seek to optimize an upper-bound of the objective resulting in two simple terms. Furthermore, we relate popular existing methods in domain adaptation, source-free domain adaptation and contrastive learning via the perspective of discriminability and diversity. The experimental results prove the superiority of our method, and our method can be adopted as a simple but strong baseline for future research in SFDA. Our method can be also adapted to source-free open-set and partial-set DA which further shows the generalization ability of our method. Code is available in https://github.com/Albert0147/AaD_SFDA., Comment: NeurIPS 2022
- Published
- 2022
44. Towards Exemplar-Free Continual Learning in Vision Transformers: an Account of Attention, Functional and Weight Regularization
- Author
-
Pelosin, Francesco, Jha, Saurav, Torsello, Andrea, Raducanu, Bogdan, van de Weijer, Joost, Pelosin, Francesco, Jha, Saurav, Torsello, Andrea, Raducanu, Bogdan, and van de Weijer, Joost
- Abstract
In this paper, we investigate the continual learning of Vision Transformers (ViT) for the challenging exemplar-free scenario, with special focus on how to efficiently distill the knowledge of its crucial self-attention mechanism (SAM). Our work takes an initial step towards a surgical investigation of SAM for designing coherent continual learning methods in ViTs. We first carry out an evaluation of established continual learning regularization techniques. We then examine the effect of regularization when applied to two key enablers of SAM: (a) the contextualized embedding layers, for their ability to capture well-scaled representations with respect to the values, and (b) the prescaled attention maps, for carrying value-independent global contextual information. We depict the perks of each distilling strategy on two image recognition benchmarks (CIFAR100 and ImageNet-32) -- while (a) leads to a better overall accuracy, (b) helps enhance the rigidity by maintaining competitive performances. Furthermore, we identify the limitation imposed by the symmetric nature of regularization losses. To alleviate this, we propose an asymmetric variant and apply it to the pooled output distillation (POD) loss adapted for ViTs. Our experiments confirm that introducing asymmetry to POD boosts its plasticity while retaining stability across (a) and (b). Moreover, we acknowledge low forgetting measures for all the compared methods, indicating that ViTs might be naturally inclined continual learner, Comment: Accepted at Continual Learning Workshop (CVPR 2022)
- Published
- 2022
45. Planckian Jitter: countering the color-crippling effects of color jitter on self-supervised training
- Author
-
Zini, Simone, Gomez-Villa, Alex, Buzzelli, Marco, Twardowski, Bartłomiej, Bagdanov, Andrew D., van de Weijer, Joost, Zini, Simone, Gomez-Villa, Alex, Buzzelli, Marco, Twardowski, Bartłomiej, Bagdanov, Andrew D., and van de Weijer, Joost
- Abstract
Several recent works on self-supervised learning are trained by mapping different augmentations of the same image to the same feature representation. The data augmentations used are of crucial importance to the quality of learned feature representations. In this paper, we analyze how the color jitter traditionally used in data augmentation negatively impacts the quality of the color features in learned feature representations. To address this problem, we propose a more realistic, physics-based color data augmentation - which we call Planckian Jitter - that creates realistic variations in chromaticity and produces a model robust to illumination changes that can be commonly observed in real life, while maintaining the ability to discriminate image content based on color information. Experiments confirm that such a representation is complementary to the representations learned with the currently-used color jitter augmentation and that a simple concatenation leads to significant performance gains on a wide range of downstream datasets. In addition, we present a color sensitivity analysis that documents the impact of different training methods on model neurons and shows that the performance of the learned features is robust with respect to illuminant variations., Comment: Accepted at Eleventh International Conference on Learning Representations (ICLR 2023)
- Published
- 2022
46. Main Product Detection with Graph Networks for Fashion
- Author
-
Yazici, Vacit Oguz, Yu, Longlong, Ramisa, Arnau, Herranz, Luis, van de Weijer, Joost, Yazici, Vacit Oguz, Yu, Longlong, Ramisa, Arnau, Herranz, Luis, and van de Weijer, Joost
- Abstract
Computer vision has established a foothold in the online fashion retail industry. Main product detection is a crucial step of vision-based fashion product feed parsing pipelines, focused in identifying the bounding boxes that contain the product being sold in the gallery of images of the product page. The current state-of-the-art approach does not leverage the relations between regions in the image, and treats images of the same product independently, therefore not fully exploiting visual and product contextual information. In this paper we propose a model that incorporates Graph Convolutional Networks (GCN) that jointly represent all detected bounding boxes in the gallery as nodes. We show that the proposed method is better than the state-of-the-art, especially, when we consider the scenario where title-input is missing at inference time and for cross-dataset evaluation, our method outperforms previous approaches by a large margin.
- Published
- 2022
- Full Text
- View/download PDF
47. OneRing: A Simple Method for Source-free Open-partial Domain Adaptation
- Author
-
Yang, Shiqi, Wang, Yaxing, Wang, Kai, Jui, Shangling, van de Weijer, Joost, Yang, Shiqi, Wang, Yaxing, Wang, Kai, Jui, Shangling, and van de Weijer, Joost
- Abstract
In this paper, we investigate Source-free Open-partial Domain Adaptation (SF-OPDA), which addresses the situation where there exist both domain and category shifts between source and target domains. Under the SF-OPDA setting, which aims to address data privacy concerns, the model cannot access source data anymore during target adaptation. We propose a novel training scheme to learn a (n+1)-way classifier to predict the n source classes and the unknown class, where samples of only known source categories are available for training. Furthermore, for target adaptation, we simply adopt a weighted entropy minimization to adapt the source pretrained model to the unlabeled target domain without source data. In experiments, we show our simple method surpasses current OPDA approaches which demand source data during adaptation. When augmented with a closed-set domain adaptation approach during target adaptation, our source-free method further outperforms the current state-of-the-art OPDA method by 2.5%, 7.2% and 13% on Office-31, Office-Home and VisDA respectively., Comment: Updated. It only focuses on source-free open-partial domain adaptation, to avoid any potential misunderstanding
- Published
- 2022
48. MVMO: A Multi-Object Dataset for Wide Baseline Multi-View Semantic Segmentation
- Author
-
Alvarez-Gila, Aitor, van de Weijer, Joost, Wang, Yaxing, Garrote, Estibaliz, Alvarez-Gila, Aitor, van de Weijer, Joost, Wang, Yaxing, and Garrote, Estibaliz
- Abstract
We present MVMO (Multi-View, Multi-Object dataset): a synthetic dataset of 116,000 scenes containing randomly placed objects of 10 distinct classes and captured from 25 camera locations in the upper hemisphere. MVMO comprises photorealistic, path-traced image renders, together with semantic segmentation ground truth for every view. Unlike existing multi-view datasets, MVMO features wide baselines between cameras and high density of objects, which lead to large disparities, heavy occlusions and view-dependent object appearance. Single view semantic segmentation is hindered by self and inter-object occlusions that could benefit from additional viewpoints. Therefore, we expect that MVMO will propel research in multi-view semantic segmentation and cross-view semantic transfer. We also provide baselines that show that new research is needed in such fields to exploit the complementary information of multi-view setups., Comment: 5 pages
- Published
- 2022
49. Avalanche: An end-to-end library for continual learning
- Author
-
Lomonaco, Vincenzo, Pellegrini, Lorenzo, Cossu, Andrea, Carta, Antonio, Graffieti, Gabriele, Hayes, Tyler L., De Lange, Matthias, Masana, Marc, Pomponi, Jary, van de Ven, Gido M., Mundt, Martin, She, Qi, Cooper, Keiland, Forest, Jeremy, Belouadah, Eden, Calderara, Simone, Parisi, German I., Cuzzolin, Fabio, Tolias, Andreas S., Scardapane, Simone, Antiga, Luca, Ahmad, Subutai, Popescu, Adrian, Kanan, Christopher, van de Weijer, Joost, Tuytelaars, Tinne, Bacciu, Davide, Maltoni, Davide, Lomonaco, Vincenzo, Pellegrini, Lorenzo, Cossu, Andrea, Carta, Antonio, Graffieti, Gabriele, Hayes, Tyler L., De Lange, Matthias, Masana, Marc, Pomponi, Jary, van de Ven, Gido M., Mundt, Martin, She, Qi, Cooper, Keiland, Forest, Jeremy, Belouadah, Eden, Calderara, Simone, Parisi, German I., Cuzzolin, Fabio, Tolias, Andreas S., Scardapane, Simone, Antiga, Luca, Ahmad, Subutai, Popescu, Adrian, Kanan, Christopher, van de Weijer, Joost, Tuytelaars, Tinne, Bacciu, Davide, and Maltoni, Davide
- Abstract
Learning continually from non-stationary data streams is a long-standing goal and a challenging problem in machine learning. Recently, we have witnessed a renewed and fast-growing interest in continual learning, especially within the deep learning community. However, algorithmic solutions are often difficult to re-implement, evaluate and port across different settings, where even results on standard benchmarks are hard to reproduce. In this work, we propose Avalanche, an open-source end-to-end library for continual learning research based on PyTorch. Avalanche is designed to provide a shared and collaborative codebase for fast prototyping, training, and reproducible evaluation of continual learning algorithms.
- Published
- 2021
50. Avalanche: An end-to-end library for continual learning
- Author
-
Lomonaco, Vincenzo, Pellegrini, Lorenzo, Cossu, Andrea, Carta, Antonio, Graffieti, Gabriele, Hayes, Tyler L., De Lange, Matthias, Masana, Marc, Pomponi, Jary, van de Ven, Gido M., Mundt, Martin, She, Qi, Cooper, Keiland, Forest, Jeremy, Belouadah, Eden, Calderara, Simone, Parisi, German I., Cuzzolin, Fabio, Tolias, Andreas S., Scardapane, Simone, Antiga, Luca, Ahmad, Subutai, Popescu, Adrian, Kanan, Christopher, van de Weijer, Joost, Tuytelaars, Tinne, Bacciu, Davide, Maltoni, Davide, Lomonaco, Vincenzo, Pellegrini, Lorenzo, Cossu, Andrea, Carta, Antonio, Graffieti, Gabriele, Hayes, Tyler L., De Lange, Matthias, Masana, Marc, Pomponi, Jary, van de Ven, Gido M., Mundt, Martin, She, Qi, Cooper, Keiland, Forest, Jeremy, Belouadah, Eden, Calderara, Simone, Parisi, German I., Cuzzolin, Fabio, Tolias, Andreas S., Scardapane, Simone, Antiga, Luca, Ahmad, Subutai, Popescu, Adrian, Kanan, Christopher, van de Weijer, Joost, Tuytelaars, Tinne, Bacciu, Davide, and Maltoni, Davide
- Abstract
Learning continually from non-stationary data streams is a long-standing goal and a challenging problem in machine learning. Recently, we have witnessed a renewed and fast-growing interest in continual learning, especially within the deep learning community. However, algorithmic solutions are often difficult to re-implement, evaluate and port across different settings, where even results on standard benchmarks are hard to reproduce. In this work, we propose Avalanche, an open-source end-to-end library for continual learning research based on PyTorch. Avalanche is designed to provide a shared and collaborative codebase for fast prototyping, training, and reproducible evaluation of continual learning algorithms.
- Published
- 2021
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.