882,068 results on '"bottleneck"'
Search Results
2. Parallel shifting bottleneck algorithms for non-permutation flow shop scheduling
- Author
-
Badri, Hossein, Bahreini, Tayebeh, and Grosu, Daniel
- Published
- 2024
- Full Text
- View/download PDF
3. Bottleneck : Moving, Building, and Belonging in an African City
- Author
-
Caroline Melly and Caroline Melly
- Subjects
- Globalization--Social aspects, Sociology, Urban--Senegal--Dakar, Social mobility--Senegal--Dakar
- Abstract
In Bottleneck, anthropologist Caroline Melly uses the problem of traffic bottlenecks to launch a wide-ranging study of mobility in contemporary urban Senegal—a concept that she argues is central to both citizens'and the state's visions of a successful future. Melly opens with an account of the generation of urban men who came of age on the heels of the era of structural adjustment, a diverse cohort with great dreams of building, moving, and belonging, but frustratingly few opportunities to do so. From there, she moves to a close study of taxi drivers and state workers, and shows how bottlenecks—physical and institutional—affect both. The third section of the book covers a seemingly stalled state effort to solve housing problems by building large numbers of concrete houses, while the fourth takes up the thousands of migrants who attempt, sometimes with tragic results, to cross the Mediterranean on rickety boats in search of new opportunities. The resulting book offers a remarkable portrait of contemporary Senegal and a means of theorizing mobility and its impossibilities far beyond the African continent.
- Published
- 2017
4. A Normalized Bottleneck Distance on Persistence Diagrams and Homology Preservation Under Dimension Reduction
- Author
-
May, Nathan H., Krishnamoorthy, Bala, and Gambill, Patrick
- Published
- 2024
- Full Text
- View/download PDF
5. Estimation of Near-Instance-Level Attribute Bottleneck for Zero-Shot Learning
- Author
-
Jiang, Chenyi, Shen, Yuming, Chen, Dubing, Zhang, Haofeng, Shao, Ling, and Torr, Philip H. S.
- Published
- 2024
- Full Text
- View/download PDF
6. Concept Bottleneck Language Models For protein design
- Author
-
Ismail, Aya Abdelsalam, Oikarinen, Tuomas, Wang, Amy, Adebayo, Julius, Stanton, Samuel, Joren, Taylor, Kleinhenz, Joseph, Goodman, Allen, Bravo, Héctor Corrada, Cho, Kyunghyun, and Frey, Nathan C.
- Subjects
Computer Science - Machine Learning - Abstract
We introduce Concept Bottleneck Protein Language Models (CB-pLM), a generative masked language model with a layer where each neuron corresponds to an interpretable concept. Our architecture offers three key benefits: i) Control: We can intervene on concept values to precisely control the properties of generated proteins, achieving a 3 times larger change in desired concept values compared to baselines. ii) Interpretability: A linear mapping between concept values and predicted tokens allows transparent analysis of the model's decision-making process. iii) Debugging: This transparency facilitates easy debugging of trained models. Our models achieve pre-training perplexity and downstream task performance comparable to traditional masked protein language models, demonstrating that interpretability does not compromise performance. While adaptable to any language model, we focus on masked protein language models due to their importance in drug discovery and the ability to validate our model's capabilities through real-world experiments and expert knowledge. We scale our CB-pLM from 24 million to 3 billion parameters, making them the largest Concept Bottleneck Models trained and the first capable of generative language modeling.
- Published
- 2024
7. Minimum Entropy Coupling with Bottleneck
- Author
-
Ebrahimi, M. Reza, Chen, Jun, and Khisti, Ashish
- Subjects
Computer Science - Machine Learning ,Computer Science - Information Theory - Abstract
This paper investigates a novel lossy compression framework operating under logarithmic loss, designed to handle situations where the reconstruction distribution diverges from the source distribution. This framework is especially relevant for applications that require joint compression and retrieval, and in scenarios involving distributional shifts due to processing. We show that the proposed formulation extends the classical minimum entropy coupling framework by integrating a bottleneck, allowing for a controlled degree of stochasticity in the coupling. We explore the decomposition of the Minimum Entropy Coupling with Bottleneck (MEC-B) into two distinct optimization problems: Entropy-Bounded Information Maximization (EBIM) for the encoder, and Minimum Entropy Coupling (MEC) for the decoder. Through extensive analysis, we provide a greedy algorithm for EBIM with guaranteed performance, and characterize the optimal solution near functional mappings, yielding significant theoretical insights into the structural complexity of this problem. Furthermore, we illustrate the practical application of MEC-B through experiments in Markov Coding Games (MCGs) under rate limits. These games simulate a communication scenario within a Markov Decision Process, where an agent must transmit a compressed message from a sender to a receiver through its actions. Our experiments highlight the trade-offs between MDP rewards and receiver accuracy across various compression rates, showcasing the efficacy of our method compared to conventional compression baseline., Comment: 38th Conference on Neural Information Processing Systems (NeurIPS 2024) - Spotlight
- Published
- 2024
8. Multimodal Information Bottleneck for Deep Reinforcement Learning with Multiple Sensors
- Author
-
You, Bang and Liu, Huaping
- Subjects
Computer Science - Machine Learning ,Computer Science - Robotics - Abstract
Reinforcement learning has achieved promising results on robotic control tasks but struggles to leverage information effectively from multiple sensory modalities that differ in many characteristics. Recent works construct auxiliary losses based on reconstruction or mutual information to extract joint representations from multiple sensory inputs to improve the sample efficiency and performance of reinforcement learning algorithms. However, the representations learned by these methods could capture information irrelevant to learning a policy and may degrade the performance. We argue that compressing information in the learned joint representations about raw multimodal observations is helpful, and propose a multimodal information bottleneck model to learn task-relevant joint representations from egocentric images and proprioception. Our model compresses and retains the predictive information in multimodal observations for learning a compressed joint representation, which fuses complementary information from visual and proprioceptive feedback and meanwhile filters out task-irrelevant information in raw multimodal observations. We propose to minimize the upper bound of our multimodal information bottleneck objective for computationally tractable optimization. Experimental evaluations on several challenging locomotion tasks with egocentric images and proprioception show that our method achieves better sample efficiency and zero-shot robustness to unseen white noise than leading baselines. We also empirically demonstrate that leveraging information from egocentric images and proprioception is more helpful for learning policies on locomotion tasks than solely using one single modality., Comment: 31 pages
- Published
- 2024
- Full Text
- View/download PDF
9. Optimizing Chain-of-Thought Reasoning: Tackling Arranging Bottleneck via Plan Augmentation
- Author
-
Qiu, Yuli, Yao, Jiashu, Huang, Heyan, and Guo, Yuhang
- Subjects
Computer Science - Computation and Language - Abstract
Multi-step reasoning ability of large language models is crucial in tasks such as math and tool utilization. Current researches predominantly focus on enhancing model performance in these multi-step reasoning tasks through fine-tuning with Chain-of-Thought (CoT) steps, yet these methods tend to be heuristic, without exploring nor resolving the bottleneck. In this study, we subdivide CoT reasoning into two parts: arranging and executing, and identify that the bottleneck of models mainly lies in arranging rather than executing. Based on this finding, we propose a plan-based training and reasoning method that guides models to generate arranging steps through abstract plans. We experiment on both math (GSM8k) and tool utilization (ToolBench) benchmarks. Results show that compared to fine-tuning directly with CoT data, our approach achieves a better performance on alleviating arranging bottleneck, particularly excelling in long-distance reasoning generalization.
- Published
- 2024
10. Optimal Causal Representations and the Causal Information Bottleneck
- Author
-
Simoes, Francisco N. F. Q., Dastani, Mehdi, and van Ommen, Thijs
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Information Theory ,Statistics - Machine Learning - Abstract
To effectively study complex causal systems, it is often useful to construct representations that simplify parts of the system by discarding irrelevant details while preserving key features. The Information Bottleneck (IB) method is a widely used approach in representation learning that compresses random variables while retaining information about a target variable. Traditional methods like IB are purely statistical and ignore underlying causal structures, making them ill-suited for causal tasks. We propose the Causal Information Bottleneck (CIB), a causal extension of the IB, which compresses a set of chosen variables while maintaining causal control over a target variable. This method produces representations which are causally interpretable, and which can be used when reasoning about interventions. We present experimental results demonstrating that the learned representations accurately capture causality as intended., Comment: Submitted to ICLR 2025. Code available at github.com/francisco-simoes/cib-optimization-psagd
- Published
- 2024
11. Persistent homology based Bottleneck distance in hypergraph products
- Author
-
Babu, Archana and John, Sunil Jacob
- Published
- 2024
- Full Text
- View/download PDF
12. A deep learning strategy for automatic congestive heart failure detection using novel bottleneck attention module
- Author
-
Wang, Jibin and Guo, Xingtian
- Published
- 2024
- Full Text
- View/download PDF
13. Microsatellite-Based Genetic Diversity, Population Structure and Bottleneck Analysis in Peanut: Conservation and Utilization Implications
- Author
-
Sangh, Chandramohan, Pandya, Janki BharatBhai, Zarna, Vora, T, Radhakrishnan, and Bera, S. K.
- Published
- 2024
- Full Text
- View/download PDF
14. Guarding the Gate: ConceptGuard Battles Concept-Level Backdoors in Concept Bottleneck Models
- Author
-
Lai, Songning, Huang, Yu, Yang, Jiayu, Huang, Gaoxiang, Chen, Wenshuo, and Yue, Yutao
- Subjects
Computer Science - Cryptography and Security ,Computer Science - Computer Vision and Pattern Recognition - Abstract
The increasing complexity of AI models, especially in deep learning, has raised concerns about transparency and accountability, particularly in high-stakes applications like medical diagnostics, where opaque models can undermine trust. Explainable Artificial Intelligence (XAI) aims to address these issues by providing clear, interpretable models. Among XAI techniques, Concept Bottleneck Models (CBMs) enhance transparency by using high-level semantic concepts. However, CBMs are vulnerable to concept-level backdoor attacks, which inject hidden triggers into these concepts, leading to undetectable anomalous behavior. To address this critical security gap, we introduce ConceptGuard, a novel defense framework specifically designed to protect CBMs from concept-level backdoor attacks. ConceptGuard employs a multi-stage approach, including concept clustering based on text distance measurements and a voting mechanism among classifiers trained on different concept subgroups, to isolate and mitigate potential triggers. Our contributions are threefold: (i) we present ConceptGuard as the first defense mechanism tailored for concept-level backdoor attacks in CBMs; (ii) we provide theoretical guarantees that ConceptGuard can effectively defend against such attacks within a certain trigger size threshold, ensuring robustness; and (iii) we demonstrate that ConceptGuard maintains the high performance and interpretability of CBMs, crucial for trustworthiness. Through comprehensive experiments and theoretical proofs, we show that ConceptGuard significantly enhances the security and trustworthiness of CBMs, paving the way for their secure deployment in critical applications., Comment: 17pages, 4 figures
- Published
- 2024
15. Which bits went where? Past and future transfer entropy decomposition with the information bottleneck
- Author
-
Murphy, Kieran A., Yin, Zhuowen, and Bassett, Dani S.
- Subjects
Computer Science - Machine Learning ,Computer Science - Information Theory - Abstract
Whether the system under study is a shoal of fish, a collection of neurons, or a set of interacting atmospheric and oceanic processes, transfer entropy measures the flow of information between time series and can detect possible causal relationships. Much like mutual information, transfer entropy is generally reported as a single value summarizing an amount of shared variation, yet a more fine-grained accounting might illuminate much about the processes under study. Here we propose to decompose transfer entropy and localize the bits of variation on both sides of information flow: that of the originating process's past and that of the receiving process's future. We employ the information bottleneck (IB) to compress the time series and identify the transferred entropy. We apply our method to decompose the transfer entropy in several synthetic recurrent processes and an experimental mouse dataset of concurrent behavioral and neural activity. Our approach highlights the nuanced dynamics within information flow, laying a foundation for future explorations into the intricate interplay of temporal processes in complex systems., Comment: NeurIPS 2024 workshop "Machine learning and the physical sciences" Camera ready
- Published
- 2024
16. GaGSL: Global-augmented Graph Structure Learning via Graph Information Bottleneck
- Author
-
Li, Shuangjie, Song, Jiangqing, Zhang, Baoming, Ruan, Gaoli, Xie, Junyuan, and Wang, Chongjun
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence - Abstract
Graph neural networks (GNNs) are prominent for their effectiveness in processing graph data for semi-supervised node classification tasks. Most works of GNNs assume that the observed structure accurately represents the underlying node relationships. However, the graph structure is inevitably noisy or incomplete in reality, which can degrade the quality of graph representations. Therefore, it is imperative to learn a clean graph structure that balances performance and robustness. In this paper, we propose a novel method named \textit{Global-augmented Graph Structure Learning} (GaGSL), guided by the Graph Information Bottleneck (GIB) principle. The key idea behind GaGSL is to learn a compact and informative graph structure for node classification tasks. Specifically, to mitigate the bias caused by relying solely on the original structure, we first obtain augmented features and augmented structure through global feature augmentation and global structure augmentation. We then input the augmented features and augmented structure into a structure estimator with different parameters for optimization and re-definition of the graph structure, respectively. The redefined structures are combined to form the final graph structure. Finally, we employ GIB based on mutual information to guide the optimization of the graph structure to obtain the minimum sufficient graph structure. Comprehensive evaluations across a range of datasets reveal the outstanding performance and robustness of GaGSL compared with the state-of-the-art methods.
- Published
- 2024
17. Bayesian Concept Bottleneck Models with LLM Priors
- Author
-
Feng, Jean, Kothari, Avni, Zier, Luke, Singh, Chandan, and Tan, Yan Shuo
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Statistics - Machine Learning - Abstract
Concept Bottleneck Models (CBMs) have been proposed as a compromise between white-box and black-box models, aiming to achieve interpretability without sacrificing accuracy. The standard training procedure for CBMs is to predefine a candidate set of human-interpretable concepts, extract their values from the training data, and identify a sparse subset as inputs to a transparent prediction model. However, such approaches are often hampered by the tradeoff between enumerating a sufficiently large set of concepts to include those that are truly relevant versus controlling the cost of obtaining concept extractions. This work investigates a novel approach that sidesteps these challenges: BC-LLM iteratively searches over a potentially infinite set of concepts within a Bayesian framework, in which Large Language Models (LLMs) serve as both a concept extraction mechanism and prior. BC-LLM is broadly applicable and multi-modal. Despite imperfections in LLMs, we prove that BC-LLM can provide rigorous statistical inference and uncertainty quantification. In experiments, it outperforms comparator methods including black-box models, converges more rapidly towards relevant concepts and away from spuriously correlated ones, and is more robust to out-of-distribution samples.
- Published
- 2024
18. Concept Complement Bottleneck Model for Interpretable Medical Image Diagnosis
- Author
-
Wang, Hongmei, Hou, Junlin, and Chen, Hao
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
Models based on human-understandable concepts have received extensive attention to improve model interpretability for trustworthy artificial intelligence in the field of medical image analysis. These methods can provide convincing explanations for model decisions but heavily rely on the detailed annotation of pre-defined concepts. Consequently, they may not be effective in cases where concepts or annotations are incomplete or low-quality. Although some methods automatically discover effective and new visual concepts rather than using pre-defined concepts or could find some human-understandable concepts via large Language models, they are prone to veering away from medical diagnostic evidence and are challenging to understand. In this paper, we propose a concept complement bottleneck model for interpretable medical image diagnosis with the aim of complementing the existing concept set and finding new concepts bridging the gap between explainable models. Specifically, we propose to use concept adapters for specific concepts to mine the concept differences and score concepts in their own attention channels to support almost fairly concept learning. Then, we devise a concept complement strategy to learn new concepts while jointly using known concepts to improve model performance. Comprehensive experiments on medical datasets demonstrate that our model outperforms the state-of-the-art competitors in concept detection and disease diagnosis tasks while providing diverse explanations to ensure model interpretability effectively., Comment: 10 pages, 5 figures, submitted to IEEE TRANSACTIONS ON MEDICAL IMAGING
- Published
- 2024
19. Tree-Based Leakage Inspection and Control in Concept Bottleneck Models
- Author
-
Ragkousis, Angelos and Parbhoo, Sonali
- Subjects
Computer Science - Machine Learning - Abstract
As AI models grow larger, the demand for accountability and interpretability has become increasingly critical for understanding their decision-making processes. Concept Bottleneck Models (CBMs) have gained attention for enhancing interpretability by mapping inputs to intermediate concepts before making final predictions. However, CBMs often suffer from information leakage, where additional input data, not captured by the concepts, is used to improve task performance, complicating the interpretation of downstream predictions. In this paper, we introduce a novel approach for training both joint and sequential CBMs that allows us to identify and control leakage using decision trees. Our method quantifies leakage by comparing the decision paths of hard CBMs with their soft, leaky counterparts. Specifically, we show that soft leaky CBMs extend the decision paths of hard CBMs, particularly in cases where concept information is incomplete. Using this insight, we develop a technique to better inspect and manage leakage, isolating the subsets of data most affected by this. Through synthetic and real-world experiments, we demonstrate that controlling leakage in this way not only improves task accuracy but also yields more informative and transparent explanations.
- Published
- 2024
20. Enforcing Interpretability in Time Series Transformers: A Concept Bottleneck Framework
- Author
-
van Sprang, Angela, Acar, Erman, and Zuidema, Willem
- Subjects
Computer Science - Machine Learning - Abstract
There has been a recent push of research on Transformer-based models for long-term time series forecasting, even though they are inherently difficult to interpret and explain. While there is a large body of work on interpretability methods for various domains and architectures, the interpretability of Transformer-based forecasting models remains largely unexplored. To address this gap, we develop a framework based on Concept Bottleneck Models to enforce interpretability of time series Transformers. We modify the training objective to encourage a model to develop representations similar to predefined interpretable concepts. In our experiments, we enforce similarity using Centered Kernel Alignment, and the predefined concepts include time features and an interpretable, autoregressive surrogate model (AR). We apply the framework to the Autoformer model, and present an in-depth analysis for a variety of benchmark tasks. We find that the model performance remains mostly unaffected, while the model shows much improved interpretability. Additionally, interpretable concepts become local, which makes the trained model easily intervenable. As a proof of concept, we demonstrate a successful intervention in the scenario of a time shift in the data, which eliminates the need to retrain.
- Published
- 2024
21. CAT: Concept-level backdoor ATtacks for Concept Bottleneck Models
- Author
-
Lai, Songning, Yang, Jiayu, Huang, Yu, Hu, Lijie, Xue, Tianlang, Hu, Zhangyi, Li, Jiaxu, Liao, Haicheng, and Yue, Yutao
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Cryptography and Security - Abstract
Despite the transformative impact of deep learning across multiple domains, the inherent opacity of these models has driven the development of Explainable Artificial Intelligence (XAI). Among these efforts, Concept Bottleneck Models (CBMs) have emerged as a key approach to improve interpretability by leveraging high-level semantic information. However, CBMs, like other machine learning models, are susceptible to security threats, particularly backdoor attacks, which can covertly manipulate model behaviors. Understanding that the community has not yet studied the concept level backdoor attack of CBM, because of "Better the devil you know than the devil you don't know.", we introduce CAT (Concept-level Backdoor ATtacks), a methodology that leverages the conceptual representations within CBMs to embed triggers during training, enabling controlled manipulation of model predictions at inference time. An enhanced attack pattern, CAT+, incorporates a correlation function to systematically select the most effective and stealthy concept triggers, thereby optimizing the attack's impact. Our comprehensive evaluation framework assesses both the attack success rate and stealthiness, demonstrating that CAT and CAT+ maintain high performance on clean data while achieving significant targeted effects on backdoored datasets. This work underscores the potential security risks associated with CBMs and provides a robust testing methodology for future security assessments.
- Published
- 2024
22. Towards a Theoretical Understanding of Synthetic Data in LLM Post-Training: A Reverse-Bottleneck Perspective
- Author
-
Gan, Zeyu and Liu, Yong
- Subjects
Computer Science - Artificial Intelligence ,Computer Science - Computation and Language ,Computer Science - Machine Learning - Abstract
Synthetic data has become a pivotal resource in post-training tasks for large language models (LLMs) due to the scarcity of high-quality, specific data. While various methods have been developed to generate synthetic data, there remains a discernible gap between the practical effects of synthetic data and our theoretical comprehension. To address this challenge, we commence by presenting a detailed modeling of the prevalent synthetic data generation process. Building upon this modeling, we demonstrate that the generalization capability of the post-trained model is critically determined by the information gain derived from the generative model, as analyzed from a novel reverse-bottleneck perspective. Moreover, we introduce the concept of Generalization Gain via Mutual Information (GGMI) and elucidate the relationship between generalization gain and information gain. This analysis serves as a theoretical foundation for synthetic data generation and further highlights its connection with the generalization capability of post-trained models, offering an understanding about the design of synthetic data generation techniques and the optimization of the post-training process. We open source our code at https://github.com/ZyGan1999/Towards-a-Theoretical-Understanding-of-Synthetic-Data-in-LLM-Post-Training.
- Published
- 2024
23. Enhancing Vision-Language Model Safety through Progressive Concept-Bottleneck-Driven Alignment
- Author
-
Liu, Zhendong, Nie, Yuanbi, Tan, Yingshui, Yue, Xiangyu, Cui, Qiushi, Wang, Chongjun, Zhu, Xiaoyong, and Zheng, Bo
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
Benefiting from the powerful capabilities of Large Language Models (LLMs), pre-trained visual encoder models connected to LLMs form Vision Language Models (VLMs). However, recent research shows that the visual modality in VLMs is highly vulnerable, allowing attackers to bypass safety alignment in LLMs through visually transmitted content, launching harmful attacks. To address this challenge, we propose a progressive concept-based alignment strategy, PSA-VLM, which incorporates safety modules as concept bottlenecks to enhance visual modality safety alignment. By aligning model predictions with specific safety concepts, we improve defenses against risky images, enhancing explainability and controllability while minimally impacting general performance. Our method is obtained through two-stage training. The low computational cost of the first stage brings very effective performance improvement, and the fine-tuning of the language model in the second stage further improves the safety performance. Our method achieves state-of-the-art results on popular VLM safety benchmark., Comment: arXiv admin note: substantial text overlap with arXiv:2405.13581
- Published
- 2024
24. Breaking the Manual Annotation Bottleneck: Creating a Comprehensive Legal Case Criticality Dataset through Semi-Automated Labeling
- Author
-
Stern, Ronja, Kawamura, Ken, Stürmer, Matthias, Chalkidis, Ilias, and Niklaus, Joel
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning ,68T50 ,I.2 ,I.7 - Abstract
Predicting case criticality helps legal professionals in the court system manage large volumes of case law. This paper introduces the Criticality Prediction dataset, a new resource for evaluating the potential influence of Swiss Federal Supreme Court decisions on future jurisprudence. Unlike existing approaches that rely on resource-intensive manual annotations, we semi-automatically derive labels leading to a much larger dataset than otherwise possible. Our dataset features a two-tier labeling system: (1) the LD-Label, which identifies cases published as Leading Decisions (LD), and (2) the Citation-Label, which ranks cases by their citation frequency and recency. This allows for a more nuanced evaluation of case importance. We evaluate several multilingual models, including fine-tuned variants and large language models, and find that fine-tuned models consistently outperform zero-shot baselines, demonstrating the need for task-specific adaptation. Our contributions include the introduction of this task and the release of a multilingual dataset to the research community.
- Published
- 2024
25. Bottleneck
- Author
-
Luke Barnes and Luke Barnes
- Subjects
- Teenage boys--England--Drama, Coming of age--Drama
- Abstract
Am I a virgin? I think I am. I mean it went in her but it was floppy and it wasn't very nice so I think I am a virgin. I'm going to say I am. Will look better on me uni applications. Liverpool, 1989. Greg is thirteen. He has just started secondary school. He earns pocket money sweeping up hair in a barbers. Girls are aliens. Liverpool FC are everything. Greg has an extraordinary story to tell you. Bottleneck is a vibrant coming-of-age story about becoming a man hrough adventures both big and small. It is about a notorious ity; Liverpool. How the outside world views it and how it views the outside world.
- Published
- 2012
26. Wilson’s bottleneck
- Author
-
Kennel, Charles, Falk, Jim, and Victor, David G.
- Published
- 2024
- Full Text
- View/download PDF
27. Lattice-Valued Bottleneck Duality
- Author
-
Ghrist, Robert, Gould, Julian, and Lopez, Miguel
- Subjects
Mathematics - Optimization and Control ,Mathematics - Combinatorics ,06D05, 90C35, 06A07 - Abstract
This note reformulates certain classical combinatorial duality theorems in the context of order lattices. For source-target networks, we generalize bottleneck path-cut and flow-cut duality results to edges with capacities in a distributive lattice. For posets, we generalize a bottleneck version of Dilworth's theorem, again weighted in a distributive lattice. These results are applicable to a wide array of non-numerical network flow problems, as shown. All results, proofs, and applications were created in collaboration with AI language models. An appendix documents their role and impact.
- Published
- 2024
28. Explanation Bottleneck Models
- Author
-
Yamaguchi, Shin'ya and Nishida, Kosuke
- Subjects
Computer Science - Artificial Intelligence ,Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Machine Learning - Abstract
Recent concept-based interpretable models have succeeded in providing meaningful explanations by pre-defined concept sets. However, the dependency on the pre-defined concepts restricts the application because of the limited number of concepts for explanations. This paper proposes a novel interpretable deep neural network called explanation bottleneck models (XBMs). XBMs generate a text explanation from the input without pre-defined concepts and then predict a final task prediction based on the generated explanation by leveraging pre-trained vision-language encoder-decoder models. To achieve both the target task performance and the explanation quality, we train XBMs through the target task loss with the regularization penalizing the explanation decoder via the distillation from the frozen pre-trained decoder. Our experiments, including a comparison to state-of-the-art concept bottleneck models, confirm that XBMs provide accurate and fluent natural language explanations without pre-defined concept sets. Code will be available at https://github.com/yshinya6/xbm/., Comment: 13 pages, 4 figures
- Published
- 2024
29. Revealing the phonon bottleneck limit in negatively charged CdS quantum dots
- Author
-
Sherman, Skylar J., Hou, Bokang, Coley-O'Rourke, Matthew J., Shulenberger, Katherine E., Rabani, Eran, and Dukovic, Gordana
- Subjects
Condensed Matter - Materials Science ,Condensed Matter - Mesoscale and Nanoscale Physics - Abstract
The capture of photoexcited hot electrons in semiconductors before they lose their excess energy to cooling is a long-standing goal in photon energy conversion. Semiconductor nanocrystals have large electron energy spacings that are expected to slow down electron relaxation by phonon emission, but hot electrons in photoexcited nanocrystals nevertheless cool rapidly by energy transfer to holes. This makes the intrinsic phonon-bottleneck limited electron lifetime in nanocrystals elusive. We used a combination of theory and experiments to probe the hot electron dynamics of negatively charged Cadmium Sulfide (CdS) colloidal quantum dots (QDs) in the absence of holes. Experiments found that these hot electrons cooled on a 100 ps timescale. Theoretical simulations predicted that pure phonon-bottleneck limited electron cooling occurs on a similar timescale. This similarity suggests that the experimental measurements reflect the upper limit on hot electron lifetimes in these CdS QDs and the lower limit on the rates of processes that can harvest those hot electrons.
- Published
- 2024
30. Graph Neural Network-State Predictive Information Bottleneck (GNN-SPIB) approach for learning molecular thermodynamics and kinetics
- Author
-
Zou, Ziyue, Wang, Dedi, and Tiwary, Pratyush
- Subjects
Computer Science - Machine Learning ,Condensed Matter - Soft Condensed Matter ,Condensed Matter - Statistical Mechanics - Abstract
Molecular dynamics simulations offer detailed insights into atomic motions but face timescale limitations. Enhanced sampling methods have addressed these challenges but even with machine learning, they often rely on pre-selected expert-based features. In this work, we present the Graph Neural Network-State Predictive Information Bottleneck (GNN-SPIB) framework, which combines graph neural networks and the State Predictive Information Bottleneck to automatically learn low-dimensional representations directly from atomic coordinates. Tested on three benchmark systems, our approach predicts essential structural, thermodynamic and kinetic information for slow processes, demonstrating robustness across diverse systems. The method shows promise for complex systems, enabling effective enhanced sampling without requiring pre-defined reaction coordinates or input features.
- Published
- 2024
31. GINTRIP: Interpretable Temporal Graph Regression using Information bottleneck and Prototype-based method
- Author
-
Royat, Ali, Moghadas, Seyed Mohamad, De Cruz, Lesley, and Munteanu, Adrian
- Subjects
Computer Science - Machine Learning - Abstract
Deep neural networks (DNNs) have demonstrated remarkable performance across various domains, yet their application to temporal graph regression tasks faces significant challenges regarding interpretability. This critical issue, rooted in the inherent complexity of both DNNs and underlying spatio-temporal patterns in the graph, calls for innovative solutions. While interpretability concerns in Graph Neural Networks (GNNs) mirror those of DNNs, to the best of our knowledge, no notable work has addressed the interpretability of temporal GNNs using a combination of Information Bottleneck (IB) principles and prototype-based methods. Our research introduces a novel approach that uniquely integrates these techniques to enhance the interpretability of temporal graph regression models. The key contributions of our work are threefold: We introduce the \underline{G}raph \underline{IN}terpretability in \underline{T}emporal \underline{R}egression task using \underline{I}nformation bottleneck and \underline{P}rototype (GINTRIP) framework, the first combined application of IB and prototype-based methods for interpretable temporal graph tasks. We derive a novel theoretical bound on mutual information (MI), extending the applicability of IB principles to graph regression tasks. We incorporate an unsupervised auxiliary classification head, fostering multi-task learning and diverse concept representation, which enhances the model bottleneck's interpretability. Our model is evaluated on real-world traffic datasets, outperforming existing methods in both forecasting accuracy and interpretability-related metrics., Comment: This work has been submitted to the IEEE for possible publication
- Published
- 2024
32. Prioritized Information Bottleneck Theoretic Framework with Distributed Online Learning for Edge Video Analytics
- Author
-
Fang, Zhengru, Hu, Senkang, Wang, Jingjing, Deng, Yiqin, Chen, Xianhao, and Fang, Yuguang
- Subjects
Computer Science - Networking and Internet Architecture - Abstract
Collaborative perception systems leverage multiple edge devices, such surveillance cameras or autonomous cars, to enhance sensing quality and eliminate blind spots. Despite their advantages, challenges such as limited channel capacity and data redundancy impede their effectiveness. To address these issues, we introduce the Prioritized Information Bottleneck (PIB) framework for edge video analytics. This framework prioritizes the shared data based on the signal-to-noise ratio (SNR) and camera coverage of the region of interest (RoI), reducing spatial-temporal data redundancy to transmit only essential information. This strategy avoids the need for video reconstruction at edge servers and maintains low latency. It leverages a deterministic information bottleneck method to extract compact, relevant features, balancing informativeness and communication costs. For high-dimensional data, we apply variational approximations for practical optimization. To reduce communication costs in fluctuating connections, we propose a gate mechanism based on distributed online learning (DOL) to filter out less informative messages and efficiently select edge servers. Moreover, we establish the asymptotic optimality of DOL by proving the sublinearity of their regrets. To validate the effectiveness of the PIB framework, we conduct real-world experiments on three types of edge devices with varied computing capabilities. Compared to five coding methods for image and video compression, PIB improves mean object detection accuracy (MODA) while reducing 17.8% and reduces communication costs by 82.65% under poor channel conditions., Comment: Major revision in IEEE ToN. We conduct additional real-world experiments on various hardware platforms. arXiv admin note: text overlap with arXiv:2408.17047
- Published
- 2024
33. PIB: Prioritized Information Bottleneck Framework for Collaborative Edge Video Analytics
- Author
-
Fang, Zhengru, Hu, Senkang, Yang, Liyan, Deng, Yiqin, Chen, Xianhao, and Fang, Yuguang
- Subjects
Computer Science - Networking and Internet Architecture - Abstract
Collaborative edge sensing systems, particularly in collaborative perception systems in autonomous driving, can significantly enhance tracking accuracy and reduce blind spots with multi-view sensing capabilities. However, their limited channel capacity and the redundancy in sensory data pose significant challenges, affecting the performance of collaborative inference tasks. To tackle these issues, we introduce a Prioritized Information Bottleneck (PIB) framework for collaborative edge video analytics. We first propose a priority-based inference mechanism that jointly considers the signal-to-noise ratio (SNR) and the camera's coverage area of the region of interest (RoI). To enable efficient inference, PIB reduces video redundancy in both spatial and temporal domains and transmits only the essential information for the downstream inference tasks. This eliminates the need to reconstruct videos on the edge server while maintaining low latency. Specifically, it derives compact, task-relevant features by employing the deterministic information bottleneck (IB) method, which strikes a balance between feature informativeness and communication costs. Given the computational challenges caused by IB-based objectives with high-dimensional data, we resort to variational approximations for feasible optimization. Compared to TOCOM-TEM, JPEG, and HEVC, PIB achieves an improvement of up to 15.1\% in mean object detection accuracy (MODA) and reduces communication costs by 66.7% when edge cameras experience poor channel conditions., Comment: Accepted by Globecom 2024. Code will be available at https://github.com/fangzr/PIB-Prioritized-Information-Bottleneck-Framework
- Published
- 2024
34. AdaCBM: An Adaptive Concept Bottleneck Model for Explainable and Accurate Diagnosis
- Author
-
Chowdhury, Townim F., Phan, Vu Minh Hieu, Liao, Kewen, To, Minh-Son, Xie, Yutong, Hengel, Anton van den, Verjans, Johan W., and Liao, Zhibin
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
The integration of vision-language models such as CLIP and Concept Bottleneck Models (CBMs) offers a promising approach to explaining deep neural network (DNN) decisions using concepts understandable by humans, addressing the black-box concern of DNNs. While CLIP provides both explainability and zero-shot classification capability, its pre-training on generic image and text data may limit its classification accuracy and applicability to medical image diagnostic tasks, creating a transfer learning problem. To maintain explainability and address transfer learning needs, CBM methods commonly design post-processing modules after the bottleneck module. However, this way has been ineffective. This paper takes an unconventional approach by re-examining the CBM framework through the lens of its geometrical representation as a simple linear classification system. The analysis uncovers that post-CBM fine-tuning modules merely rescale and shift the classification outcome of the system, failing to fully leverage the system's learning potential. We introduce an adaptive module strategically positioned between CLIP and CBM to bridge the gap between source and downstream domains. This simple yet effective approach enhances classification performance while preserving the explainability afforded by the framework. Our work offers a comprehensive solution that encompasses the entire process, from concept discovery to model training, providing a holistic recipe for leveraging the strengths of GPT, CLIP, and CBM., Comment: Accepted at MICCAI 2024, the 27th International Conference on Medical Image Computing and Computer Assisted Intervention
- Published
- 2024
35. Contrastive Graph Representation Learning with Adversarial Cross-view Reconstruction and Information Bottleneck
- Author
-
Shou, Yuntao, Lan, Haozhi, and Cao, Xiangyong
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence - Abstract
Graph Neural Networks (GNNs) have received extensive research attention due to their powerful information aggregation capabilities. Despite the success of GNNs, most of them suffer from the popularity bias issue in a graph caused by a small number of popular categories. Additionally, real graph datasets always contain incorrect node labels, which hinders GNNs from learning effective node representations. Graph contrastive learning (GCL) has been shown to be effective in solving the above problems for node classification tasks. Most existing GCL methods are implemented by randomly removing edges and nodes to create multiple contrasting views, and then maximizing the mutual information (MI) between these contrasting views to improve the node feature representation. However, maximizing the mutual information between multiple contrasting views may lead the model to learn some redundant information irrelevant to the node classification task. To tackle this issue, we propose an effective Contrastive Graph Representation Learning with Adversarial Cross-view Reconstruction and Information Bottleneck (CGRL) for node classification, which can adaptively learn to mask the nodes and edges in the graph to obtain the optimal graph structure representation. Furthermore, we innovatively introduce the information bottleneck theory into GCLs to remove redundant information in multiple contrasting views while retaining as much information as possible about node classification. Moreover, we add noise perturbations to the original views and reconstruct the augmented views by constructing adversarial views to improve the robustness of node feature representation. Extensive experiments on real-world public datasets demonstrate that our method significantly outperforms existing state-of-the-art algorithms., Comment: 13 pages, 7 figures
- Published
- 2024
36. IBMEA: Exploring Variational Information Bottleneck for Multi-modal Entity Alignment
- Author
-
Su, Taoyu, Sheng, Jiawei, Wang, Shicheng, Zhang, Xinghua, Xu, Hongbo, and Liu, Tingwen
- Subjects
Computer Science - Computation and Language ,Computer Science - Multimedia - Abstract
Multi-modal entity alignment (MMEA) aims to identify equivalent entities between multi-modal knowledge graphs (MMKGs), where the entities can be associated with related images. Most existing studies integrate multi-modal information heavily relying on the automatically-learned fusion module, rarely suppressing the redundant information for MMEA explicitly. To this end, we explore variational information bottleneck for multi-modal entity alignment (IBMEA), which emphasizes the alignment-relevant information and suppresses the alignment-irrelevant information in generating entity representations. Specifically, we devise multi-modal variational encoders to generate modal-specific entity representations as probability distributions. Then, we propose four modal-specific information bottleneck regularizers, limiting the misleading clues in refining modal-specific entity representations. Finally, we propose a modal-hybrid information contrastive regularizer to integrate all the refined modal-specific representations, enhancing the entity similarity between MMKGs to achieve MMEA. We conduct extensive experiments on two cross-KG and three bilingual MMEA datasets. Experimental results demonstrate that our model consistently outperforms previous state-of-the-art methods, and also shows promising and robust performance in low-resource and high-noise data scenarios., Comment: Accepted by ACM MM 2024
- Published
- 2024
37. VLG-CBM: Training Concept Bottleneck Models with Vision-Language Guidance
- Author
-
Srivastava, Divyansh, Yan, Ge, and Weng, Tsui-Wei
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Machine Learning - Abstract
Concept Bottleneck Models (CBMs) provide interpretable prediction by introducing an intermediate Concept Bottleneck Layer (CBL), which encodes human-understandable concepts to explain models' decision. Recent works proposed to utilize Large Language Models (LLMs) and pre-trained Vision-Language Models (VLMs) to automate the training of CBMs, making it more scalable and automated. However, existing approaches still fall short in two aspects: First, the concepts predicted by CBL often mismatch the input image, raising doubts about the faithfulness of interpretation. Second, it has been shown that concept values encode unintended information: even a set of random concepts could achieve comparable test accuracy to state-of-the-art CBMs. To address these critical limitations, in this work, we propose a novel framework called Vision-Language-Guided Concept Bottleneck Model (VLG-CBM) to enable faithful interpretability with the benefits of boosted performance. Our method leverages off-the-shelf open-domain grounded object detectors to provide visually grounded concept annotation, which largely enhances the faithfulness of concept prediction while further improving the model performance. In addition, we propose a new metric called Number of Effective Concepts (NEC) to control the information leakage and provide better interpretability. Extensive evaluations across five standard benchmarks show that our method, VLG-CBM, outperforms existing methods by at least 4.27% and up to 51.09% on accuracy at NEC=5, and by at least 0.45% and up to 29.78% on average accuracy across different NECs, while preserving both faithfulness and interpretability of the learned concepts as demonstrated in extensive experiments., Comment: Accepted by NeurIPS 2024
- Published
- 2024
38. On efficient algorithms for bottleneck path problems with many sources
- Author
-
Kaymakov, Kirill V. and Malyshev, Dmitry S.
- Published
- 2024
- Full Text
- View/download PDF
39. Masked self-supervised ECG representation learning via multiview information bottleneck
- Author
-
Yang, Shunxiang, Lian, Cheng, Zeng, Zhigang, Xu, Bingrong, Su, Yixin, and Xue, Chenyang
- Published
- 2024
- Full Text
- View/download PDF
40. Suppressing internet bottleneck with Kudryashov’s extended version of self-phase modulation and fractional temporal evolution
- Author
-
Murad, Muhammad Amin S., Arnous, Ahmed H., Biswas, Anjan, Yildirim, Yakup, and Alshomrani, Ali Saleh
- Published
- 2024
- Full Text
- View/download PDF
41. A diversity preserving genetic algorithm with tailor-made variation operators for the quadratic bottleneck knapsack problem
- Author
-
Patel, Pritibahen Sumanbhai and Singh, Alok
- Published
- 2024
- Full Text
- View/download PDF
42. DSBAV-Net: Depthwise Separable Bottleneck Attention V-Shaped Network with Hybrid Convolution for Left Atrium Segmentation
- Author
-
Ocal, Hakan
- Published
- 2024
- Full Text
- View/download PDF
43. DIB-UAP: enhancing the transferability of universal adversarial perturbation via deep information bottleneck
- Author
-
Wang, Yang, Zheng, Yunfei, Chen, Lei, Yang, Zhen, and Cao, Tieyong
- Published
- 2024
- Full Text
- View/download PDF
44. The Devil is in the Sources! Knowledge Enhanced Cross-Domain Recommendation in an Information Bottleneck Perspective
- Author
-
Hu, Binbin, Wang, Weifan, Wang, Hanshu, Liu, Ziqi, Shen, Bin, He, Yong, and Chen, Jiawei
- Subjects
Computer Science - Information Retrieval - Abstract
Cross-domain Recommendation (CDR) aims to alleviate the data sparsity and the cold-start problems in traditional recommender systems by leveraging knowledge from an informative source domain. However, previously proposed CDR models pursue an imprudent assumption that the entire information from the source domain is equally contributed to the target domain, neglecting the evil part that is completely irrelevant to users' intrinsic interest. To address this concern, in this paper, we propose a novel knowledge enhanced cross-domain recommendation framework named CoTrans, which remolds the core procedures of CDR models with: Compression on the knowledge from the source domain and Transfer of the purity to the target domain. Specifically, following the theory of Graph Information Bottleneck, CoTrans first compresses the source behaviors with the perception of information from the target domain. Then to preserve all the important information for the CDR task, the feedback signals from both domains are utilized to promote the effectiveness of the transfer procedure. Additionally, a knowledge-enhanced encoder is employed to narrow gaps caused by the non-overlapped items across separate domains. Comprehensive experiments on three widely used cross-domain datasets demonstrate that CoTrans significantly outperforms both single-domain and state-of-the-art cross-domain recommendation approaches., Comment: Accepted by CIKM 2024
- Published
- 2024
45. EQ-CBM: A Probabilistic Concept Bottleneck with Energy-based Models and Quantized Vectors
- Author
-
Kim, Sangwon, Ahn, Dasom, Ko, Byoung Chul, Jang, In-su, and Kim, Kwang-Ju
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
The demand for reliable AI systems has intensified the need for interpretable deep neural networks. Concept bottleneck models (CBMs) have gained attention as an effective approach by leveraging human-understandable concepts to enhance interpretability. However, existing CBMs face challenges due to deterministic concept encoding and reliance on inconsistent concepts, leading to inaccuracies. We propose EQ-CBM, a novel framework that enhances CBMs through probabilistic concept encoding using energy-based models (EBMs) with quantized concept activation vectors (qCAVs). EQ-CBM effectively captures uncertainties, thereby improving prediction reliability and accuracy. By employing qCAVs, our method selects homogeneous vectors during concept encoding, enabling more decisive task performance and facilitating higher levels of human intervention. Empirical results using benchmark datasets demonstrate that our approach outperforms the state-of-the-art in both concept and task accuracy., Comment: Accepted by ACCV 2024
- Published
- 2024
46. Hysteresis Behind A Bottleneck With Location-Dependent Capacity
- Author
-
Hammerl, Alexander, Seshadri, Ravi, Rasmussen, Thomas Kjær, and Nielsen, Otto Anker
- Subjects
Condensed Matter - Statistical Mechanics - Abstract
Macroscopic fundamental diagrams (MFDs) and related network traffic dynamics models have received both theoretical support and empirical validation with the emergence of new data collection technologies. However, the existence of well-defined MFD curves can only be expected for traffic networks with specific topologies and is subject to various disturbances, most importantly hysteresis phenomena. This study aims to improve the understanding of hysteresis in Macroscopic Fundamental Diagrams and Network Exit Functions (NEFs) during rush hour conditions. We apply the LWR theory to a highway corridor featuring a location-dependent downstream bottleneck to identify a figure-eight hysteresis pattern, clockwise on the top and counter-clockwise on the bottom. We discuss why this general pattern is rare in practical scenarios, where a single clockwise loop is more typical. The paper discusses the impact of the road topology and demand parameters on the formation and intensity of hysteresis loops analytically. To substantiate these findings, we employ numerical simulations using the Cell Transmission Model (CTM). Our simulations show that even a slight reduction in the capacity of the homogeneous section can significantly decrease MFD hysteresis while maintaining outflow at the corridor's downstream end. These reductions can be achieved with minimal intervention through standard traffic control measures, such as dynamic speed limits or ramp metering.
- Published
- 2024
47. Turbo Equalization with Coarse Quantization using the Information Bottleneck Method
- Author
-
Mohr, Philipp, Brüggmann, Jasper, and Bauch, Gerhard
- Subjects
Computer Science - Information Theory ,Electrical Engineering and Systems Science - Signal Processing - Abstract
This paper proposes a turbo equalizer for intersymbol interference channels (ISI) that uses coarsely quantized messages across all receiver components. Lookup tables (LUTs) carry out compression operations designed with the information bottleneck method aiming to maximize relevant mutual information. The turbo setup consists of an equalizer and a decoder that provide extrinsic information to each other over multiple turbo iterations. We develop simplified LUT structures to incorporate the decoder feedback in the equalizer with significantly reduced complexity. The proposed receiver is optimized for selected ISI channels. A conceptual hardware implementation is developed to compare the area efficiency and error correction performance. A thorough analysis reveals that LUT-based configurations with very coarse quantization can achieve higher area efficiency than conventional equalizers. Moreover, the proposed turbo setups can outperform the respective non-turbo setups regarding area efficiency and error correction capability.
- Published
- 2024
48. Statistically Valid Information Bottleneck via Multiple Hypothesis Testing
- Author
-
Farzaneh, Amirmohammad and Simeone, Osvaldo
- Subjects
Computer Science - Information Theory ,Computer Science - Machine Learning - Abstract
The information bottleneck (IB) problem is a widely studied framework in machine learning for extracting compressed features that are informative for downstream tasks. However, current approaches to solving the IB problem rely on a heuristic tuning of hyperparameters, offering no guarantees that the learned features satisfy information-theoretic constraints. In this work, we introduce a statistically valid solution to this problem, referred to as IB via multiple hypothesis testing (IB-MHT), which ensures that the learned features meet the IB constraints with high probability, regardless of the size of the available dataset. The proposed methodology builds on Pareto testing and learn-then-test (LTT), and it wraps around existing IB solvers to provide statistical guarantees on the IB constraints. We demonstrate the performance of IB-MHT on classical and deterministic IB formulations, including experiments on distillation of language models. The results validate the effectiveness of IB-MHT in outperforming conventional methods in terms of statistical robustness and reliability.
- Published
- 2024
49. Task-Oriented Communication for Graph Data: A Graph Information Bottleneck Approach
- Author
-
Li, Shujing, Wang, Yanhu, Guo, Shuaishuai, and Feng, Chenyuan
- Subjects
Computer Science - Machine Learning ,Computer Science - Social and Information Networks ,Electrical Engineering and Systems Science - Signal Processing - Abstract
Graph data, essential in fields like knowledge representation and social networks, often involves large networks with many nodes and edges. Transmitting these graphs can be highly inefficient due to their size and redundancy for specific tasks. This paper introduces a method to extract a smaller, task-focused subgraph that maintains key information while reducing communication overhead. Our approach utilizes graph neural networks (GNNs) and the graph information bottleneck (GIB) principle to create a compact, informative, and robust graph representation suitable for transmission. The challenge lies in the irregular structure of graph data, making GIB optimization complex. We address this by deriving a tractable variational upper bound for the objective function. Additionally, we propose the VQ-GIB mechanism, integrating vector quantization (VQ) to convert subgraph representations into a discrete codebook sequence, compatible with existing digital communication systems. Our experiments show that this GIB-based method significantly lowers communication costs while preserving essential task-related information. The approach demonstrates robust performance across various communication channels, suitable for both continuous and discrete systems.
- Published
- 2024
50. Debiasing Graph Representation Learning based on Information Bottleneck
- Author
-
Zhang, Ziyi, Ouyang, Mingxuan, Lin, Wanyu, Lan, Hao, and Yang, Lei
- Subjects
Computer Science - Machine Learning ,Computer Science - Computers and Society - Abstract
Graph representation learning has shown superior performance in numerous real-world applications, such as finance and social networks. Nevertheless, most existing works might make discriminatory predictions due to insufficient attention to fairness in their decision-making processes. This oversight has prompted a growing focus on fair representation learning. Among recent explorations on fair representation learning, prior works based on adversarial learning usually induce unstable or counterproductive performance. To achieve fairness in a stable manner, we present the design and implementation of GRAFair, a new framework based on a variational graph auto-encoder. The crux of GRAFair is the Conditional Fairness Bottleneck, where the objective is to capture the trade-off between the utility of representations and sensitive information of interest. By applying variational approximation, we can make the optimization objective tractable. Particularly, GRAFair can be trained to produce informative representations of tasks while containing little sensitive information without adversarial training. Experiments on various real-world datasets demonstrate the effectiveness of our proposed method in terms of fairness, utility, robustness, and stability.
- Published
- 2024
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.