5,715 results on '"Chen, Quan"'
Search Results
2. Manifestly unitary higher Hilbert spaces
- Author
-
Chen, Quan, Ferrer, Giovanni, Hungar, Brett, Penneys, David, and Sanford, Sean
- Subjects
Mathematics - Quantum Algebra ,Mathematics - Category Theory ,Mathematics - Operator Algebras ,Primary: 18M40, 18N10, 18N20, Secondary: 18M20, 18M30, 18N25 - Abstract
Higher idempotent completion gives a formal inductive construction of the $n$-category of finite dimensional $n$-vector spaces starting with the complex numbers. We propose a manifestly unitary construction of low dimensional higher Hilbert spaces, formally constructing the $\mathrm{C}^*$-3-category of 3-Hilbert spaces from Baez's 2-Hilbert spaces, which itself forms a 3-Hilbert space. We prove that the forgetful functor from 3-Hilbert spaces to 3-vector spaces is fully faithful., Comment: 71 pages, 5 figures, many tikz diagrams. Comments welcome!
- Published
- 2024
3. A QoE-Aware Split Inference Accelerating Algorithm for NOMA-based Edge Intelligence
- Author
-
Yuan, Xin, Li, Ning, Chen, Quan, Xu, Wenchao, Zhang, Zhaoxin, and Guo, Song
- Subjects
Computer Science - Machine Learning - Abstract
Even the AI has been widely used and significantly changed our life, deploying the large AI models on resource limited edge devices directly is not appropriate. Thus, the model split inference is proposed to improve the performance of edge intelligence, in which the AI model is divided into different sub models and the resource-intensive sub model is offloaded to edge server wirelessly for reducing resource requirements and inference latency. However, the previous works mainly concentrate on improving and optimizing the system QoS, ignore the effect of QoE which is another critical item for the users except for QoS. Even the QoE has been widely learned in EC, considering the differences between task offloading in EC and split inference in EI, and the specific issues in QoE which are still not addressed in EC and EI, these algorithms cannot work effectively in edge split inference scenarios. Thus, an effective resource allocation algorithm is proposed in this paper, for accelerating split inference in EI and achieving the tradeoff between inference delay, QoE, and resource consumption, abbreviated as ERA. Specifically, the ERA takes the resource consumption, QoE, and inference latency into account to find the optimal model split strategy and resource allocation strategy. Since the minimum inference delay and resource consumption, and maximum QoE cannot be satisfied simultaneously, the gradient descent based algorithm is adopted to find the optimal tradeoff between them. Moreover, the loop iteration GD approach is developed to reduce the complexity of the GD algorithm caused by parameter discretization. Additionally, the properties of the proposed algorithms are investigated, including convergence, complexity, and approximation error. The experimental results demonstrate that the performance of ERA is much better than that of the previous studies., Comment: 16pages, 19figures. arXiv admin note: substantial text overlap with arXiv:2312.15850
- Published
- 2024
4. PolicyCraft: Supporting Collaborative and Participatory Policy Design through Case-Grounded Deliberation
- Author
-
Kuo, Tzu-Sheng, Chen, Quan Ze, Zhang, Amy X., Hsieh, Jane, Zhu, Haiyi, and Holstein, Kenneth
- Subjects
Computer Science - Human-Computer Interaction - Abstract
Community and organizational policies are typically designed in a top-down, centralized fashion, with limited input from impacted stakeholders. This can result in policies that are misaligned with community needs or perceived as illegitimate. How can we support more collaborative, participatory approaches to policy design? In this paper, we present PolicyCraft, a system that structures collaborative policy design through case-grounded deliberation. Building on past research that highlights the value of concrete cases in establishing common ground, PolicyCraft supports users in collaboratively proposing, critiquing, and revising policies through discussion and voting on cases. A field study across two university courses showed that students using PolicyCraft reached greater consensus and developed better-supported course policies, compared with those using a baseline system that did not scaffold their use of concrete cases. Reflecting on our findings, we discuss opportunities for future HCI systems to help groups more effectively bridge between abstract policies and concrete cases.
- Published
- 2024
5. Policy Prototyping for LLMs: Pluralistic Alignment via Interactive and Collaborative Policymaking
- Author
-
Feng, K. J. Kevin, Cheong, Inyoung, Chen, Quan Ze, and Zhang, Amy X.
- Subjects
Computer Science - Human-Computer Interaction - Abstract
Emerging efforts in AI alignment seek to broaden participation in shaping model behavior by eliciting and integrating collective input into a policy for model finetuning. While pluralistic, these processes are often linear and do not allow participating stakeholders to confirm whether potential outcomes of their contributions are indeed consistent with their intentions. Design prototyping has long advocated for rapid iteration using tight feedback loops of ideation, experimentation, and evaluation to mitigate these issues. We thus propose policy prototyping for LLMs, a new process that draws inspiration from prototyping practices to enable stakeholders to collaboratively and interactively draft LLM policies. Through learnings from a real-world LLM policymaking initiative at an industrial AI lab, we motivate our approach and characterize policy prototyping with four guiding principles. Because policy prototyping emphasizes a contrasting set of priorities compared to previous approaches, we envision our approach to be a valuable addition to the methodological repertoire for pluralistic alignment.
- Published
- 2024
6. Whole Heart Perfusion with High-Multiband Simultaneous Multislice Imaging via Linear Phase Modulated Extended Field of View (SMILE)
- Author
-
Zhao, Shen, Wang, Junyu, Wang, Xitong, Liu, Sizhuo, Chen, Quan, and Salerno, Michael
- Subjects
Electrical Engineering and Systems Science - Signal Processing - Abstract
Purpose: To develop a simultaneous multislice (SMS) first-pass perfusion technique that can achieve whole heart coverage with high multi-band factors, while avoiding the issue of slice leakage. Methods: The proposed Simultaneous Multislice Imaging via Linear phase modulated Extended field of view (SMILE) treats the SMS acquisition and reconstruction within an extended field of view framework, allowing arbitrarily under-sampling of phase encoding lines of the extended k-space matrix and enabling the direct application of 2D parallel imaging reconstruction techniques. We presented a theoretical framework that offers insights into the performance of SMILE. We performed retrospective comparison on 28 subjects and prospective perfusion experiments on 49 patients undergoing routine clinical CMR studies with SMILE at multiband (MB) factors of 3-5, with a total acceleration factor ($R$) of 8 and 10 respectively, and compared SMILE to conventional SMS techniques using standard FOV 2D CAIPI acquisition and standard 2D slice separation techniques including split-slice GRAPPA and ROCK-SPIRiT. Results: Retrospective studies demonstrated 5.2 to 8.0 dB improvement in signal to error ratio (SER) of SMILE over CAIPI perfusion. Prospective studies showed good image quality with grades of 4.5 $\pm$ 0.5 for MB=3, $R$=8 and 3.6 $\pm$ 0.8 for MB=5, $R$=10. (5-point Likert Scale) Conclusion: The theoretical derivation and experimental results validate the SMILE's improved performance at high acceleration and MB factors as compared to the existing 2D CAIPI SMS acquisition and reconstruction techniques for first-pass myocardial perfusion imaging., Comment: 15 pages, 12 figures
- Published
- 2024
7. End User Authoring of Personalized Content Classifiers: Comparing Example Labeling, Rule Writing, and LLM Prompting
- Author
-
Wang, Leijie, Yurechko, Kathryn, Dani, Pranati, Chen, Quan Ze, and Zhang, Amy X.
- Subjects
Computer Science - Human-Computer Interaction - Abstract
Existing tools for laypeople to create personal classifiers often assume a motivated user working uninterrupted in a single, lengthy session. However, users tend to engage with social media casually, with many short sessions on an ongoing, daily basis. To make creating personal classifiers for content curation easier for such users, tools should support rapid initialization and iterative refinement. In this work, we compare three strategies -- (1) example labeling, (2) rule writing, and (3) large language model (LLM) prompting -- for end users to build personal content classifiers. From an experiment with 37 non-programmers tasked with creating personalized comment moderation filters, we found that with LLM prompting, participants reached 95\% of peak performance in 5 minutes, beating other strategies due to higher recall, but all strategies struggled with iterative refinement. Despite LLM prompting's better performance, participants preferred different strategies in different contexts and, even when prompting, provided examples or wrote rule-like prompts, suggesting hybrid approaches.
- Published
- 2024
8. Video to Music Moment Retrieval
- Author
-
Xin, Zijie, Wang, Minquan, Ma, Ye, Wang, Bo, Chen, Quan, Jiang, Peng, and Li, Xirong
- Subjects
Computer Science - Multimedia - Abstract
Adding proper background music helps complete a short video to be shared. Towards automating the task, previous research focuses on video-to-music retrieval (VMR), aiming to find amidst a collection of music the one best matching the content of a given video. Since music tracks are typically much longer than short videos, meaning the returned music has to be cut to a shorter moment, there is a clear gap between the practical need and VMR. In order to bridge the gap, we propose in this paper video to music moment retrieval (VMMR) as a new task. To tackle the new task, we build a comprehensive dataset Ad-Moment which contains 50K short videos annotated with music moments and develop a two-stage approach. In particular, given a test video, the most similar music is retrieved from a given collection. Then, a Transformer based music moment localization is performed. We term this approach Retrieval and Localization (ReaL). Extensive experiments on real-world datasets verify the effectiveness of the proposed method for VMMR.
- Published
- 2024
9. D&M: Enriching E-commerce Videos with Sound Effects by Key Moment Detection and SFX Matching
- Author
-
Liu, Jingyu, Wang, Minquan, Ma, Ye, Wang, Bo, Chen, Aozhu, Chen, Quan, Jiang, Peng, and Li, Xirong
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Videos showcasing specific products are increasingly important for E-commerce. Key moments naturally exist as the first appearance of a specific product, presentation of its distinctive features, the presence of a buying link, etc. Adding proper sound effects (SFX) to these key moments, or video decoration with SFX (VDSFX), is crucial for enhancing the user engaging experience. Previous studies about adding SFX to videos perform video to SFX matching at a holistic level, lacking the ability of adding SFX to a specific moment. Meanwhile, previous studies on video highlight detection or video moment retrieval consider only moment localization, leaving moment to SFX matching untouched. By contrast, we propose in this paper D&M, a unified method that accomplishes key moment detection and moment to SFX matching simultaneously. Moreover, for the new VDSFX task we build a large-scale dataset SFX-Moment from an E-commerce platform. For a fair comparison, we build competitive baselines by extending a number of current video moment detection methods to the new task. Extensive experiments on SFX-Moment show the superior performance of the proposed method over the baselines. Code and data will be released., Comment: 9 pages, 4 figures
- Published
- 2024
10. ASR-enhanced Multimodal Representation Learning for Cross-Domain Product Retrieval
- Author
-
Zhao, Ruixiang, Jia, Jian, Li, Yan, Bai, Xuehan, Chen, Quan, Li, Han, Jiang, Peng, and Li, Xirong
- Subjects
Computer Science - Multimedia ,Computer Science - Artificial Intelligence ,Computer Science - Computer Vision and Pattern Recognition - Abstract
E-commerce is increasingly multimedia-enriched, with products exhibited in a broad-domain manner as images, short videos, or live stream promotions. A unified and vectorized cross-domain production representation is essential. Due to large intra-product variance and high inter-product similarity in the broad-domain scenario, a visual-only representation is inadequate. While Automatic Speech Recognition (ASR) text derived from the short or live-stream videos is readily accessible, how to de-noise the excessively noisy text for multimodal representation learning is mostly untouched. We propose ASR-enhanced Multimodal Product Representation Learning (AMPere). In order to extract product-specific information from the raw ASR text, AMPere uses an easy-to-implement LLM-based ASR text summarizer. The LLM-summarized text, together with visual data, is then fed into a multi-branch network to generate compact multimodal embeddings. Extensive experiments on a large-scale tri-domain dataset verify the effectiveness of AMPere in obtaining a unified multimodal product representation that clearly improves cross-domain product retrieval., Comment: 10 pages, 5 figures
- Published
- 2024
11. Spatiotemporal Graph Guided Multi-modal Network for Livestreaming Product Retrieval
- Author
-
Hu, Xiaowan, Chen, Yiyi, Li, Yan, Wang, Minquan, Wang, Haoqian, Chen, Quan, Li, Han, and Jiang, Peng
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Multimedia - Abstract
With the rapid expansion of e-commerce, more consumers have become accustomed to making purchases via livestreaming. Accurately identifying the products being sold by salespeople, i.e., livestreaming product retrieval (LPR), poses a fundamental and daunting challenge. The LPR task encompasses three primary dilemmas in real-world scenarios: 1) the recognition of intended products from distractor products present in the background; 2) the video-image heterogeneity that the appearance of products showcased in live streams often deviates substantially from standardized product images in stores; 3) there are numerous confusing products with subtle visual nuances in the shop. To tackle these challenges, we propose the Spatiotemporal Graphing Multi-modal Network (SGMN). First, we employ a text-guided attention mechanism that leverages the spoken content of salespeople to guide the model to focus toward intended products, emphasizing their salience over cluttered background products. Second, a long-range spatiotemporal graph network is further designed to achieve both instance-level interaction and frame-level matching, solving the misalignment caused by video-image heterogeneity. Third, we propose a multi-modal hard example mining, assisting the model in distinguishing highly similar products with fine-grained features across the video-image-text domain. Through extensive quantitative and qualitative experiments, we demonstrate the superior performance of our proposed SGMN model, surpassing the state-of-the-art methods by a substantial margin. The code is available at https://github.com/Huxiaowan/SGMN., Comment: 16 pages, 12 figures
- Published
- 2024
12. Training-free Subject-Enhanced Attention Guidance for Compositional Text-to-image Generation
- Author
-
Liu, Shengyuan, Wang, Bo, Ma, Ye, Yang, Te, Cao, Xipeng, Chen, Quan, Li, Han, Dong, Di, and Jiang, Peng
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Existing subject-driven text-to-image generation models suffer from tedious fine-tuning steps and struggle to maintain both text-image alignment and subject fidelity. For generating compositional subjects, it often encounters problems such as object missing and attribute mixing, where some subjects in the input prompt are not generated or their attributes are incorrectly combined. To address these limitations, we propose a subject-driven generation framework and introduce training-free guidance to intervene in the generative process during inference time. This approach strengthens the attention map, allowing for precise attribute binding and feature injection for each subject. Notably, our method exhibits exceptional zero-shot generation ability, especially in the challenging task of compositional generation. Furthermore, we propose a novel metric GroundingScore to evaluate subject alignment thoroughly. The obtained quantitative results serve as compelling evidence showcasing the effectiveness of our proposed method. The code will be released soon., Comment: 26 pages, 13 figures
- Published
- 2024
13. Knowledge Adaptation from Large Language Model to Recommendation for Practical Industrial Application
- Author
-
Jia, Jian, Wang, Yipei, Li, Yan, Chen, Honggang, Bai, Xuehan, Liu, Zhaocheng, Liang, Jian, Chen, Quan, Li, Han, Jiang, Peng, and Gai, Kun
- Subjects
Computer Science - Information Retrieval ,Computer Science - Artificial Intelligence - Abstract
Contemporary recommender systems predominantly rely on collaborative filtering techniques, employing ID-embedding to capture latent associations among users and items. However, this approach overlooks the wealth of semantic information embedded within textual descriptions of items, leading to suboptimal performance in cold-start scenarios and long-tail user recommendations. Leveraging the capabilities of Large Language Models (LLMs) pretrained on massive text corpus presents a promising avenue for enhancing recommender systems by integrating open-world domain knowledge. In this paper, we propose an Llm-driven knowlEdge Adaptive RecommeNdation (LEARN) framework that synergizes open-world knowledge with collaborative knowledge. We address computational complexity concerns by utilizing pretrained LLMs as item encoders and freezing LLM parameters to avoid catastrophic forgetting and preserve open-world knowledge. To bridge the gap between the open-world and collaborative domains, we design a twin-tower structure supervised by the recommendation task and tailored for practical industrial application. Through offline experiments on the large-scale industrial dataset and online experiments on A/B tests, we demonstrate the efficacy of our approach., Comment: 11 pages, 6 figures
- Published
- 2024
14. Towards Fast Setup and High Throughput of GPU Serverless Computing
- Author
-
Zhao, Han, Cui, Weihao, Chen, Quan, Zhang, Shulai, Li, Zijun, Leng, Jingwen, Li, Chao, Zeng, Deze, and Guo, Minyi
- Subjects
Computer Science - Distributed, Parallel, and Cluster Computing - Abstract
Integrating GPUs into serverless computing platforms is crucial for improving efficiency. However, existing solutions for GPU-enabled serverless computing platforms face two significant problems due to coarse-grained GPU management: long setup time and low function throughput. To address these issues, we propose SAGE, a GPU serverless framework with fast setup and high throughput. First, based on the data knowability of GPU function ahead of actual execution, SAGE first devises the parallelized function setup mechanism, which parallelizes the data preparation and context creation. In this way, SAGE achieves fast setup of GPU function invocations.Second, SAGE further proposes the sharing-based memory management mechanism, which shares the read-only memory and context memory across multiple invocations of the same function. The memory sharing mechanism avoids repeated data preparation and then unnecessary data-loading contention. As a consequence, the function throughput could be improved. Our experimental results show that SAGE reduces function duration by 11.3X and improves function density by 1.22X compared to the state-of-the-art serverless platform.
- Published
- 2024
15. A Codesign of Scheduling and Parallelization for Large Model Training in Heterogeneous Clusters
- Author
-
Xue, Chunyu, Cui, Weihao, Zhao, Han, Chen, Quan, Zhang, Shulai, Yang, Pengyu, Yang, Jing, Li, Shaobo, and Guo, Minyi
- Subjects
Computer Science - Distributed, Parallel, and Cluster Computing ,Computer Science - Machine Learning - Abstract
Joint consideration of scheduling and adaptive parallelism offers great opportunities for improving the training efficiency of large models on heterogeneous GPU clusters. However, integrating adaptive parallelism into a cluster scheduler expands the cluster scheduling space. The new space is the product of the original scheduling space and the parallelism exploration space of adaptive parallelism (also a product of pipeline, data, and tensor parallelism). The exponentially enlarged scheduling space and ever-changing optimal parallelism plan from adaptive parallelism together result in the contradiction between low-overhead and accurate performance data acquisition for efficient cluster scheduling. This paper presents Crius, a training system for efficiently scheduling multiple large models with adaptive parallelism in a heterogeneous cluster. Crius proposes a novel scheduling granularity called Cell. It represents a job with deterministic resources and pipeline stages. The exploration space of Cell is shrunk to the product of only data and tensor parallelism, thus exposing the potential for accurate and low-overhead performance estimation. Crius then accurately estimates Cells and efficiently schedules training jobs. When a Cell is selected as a scheduling choice, its represented job runs with the optimal parallelism plan explored. Experimental results show that Crius reduces job completion time by up to 48.9% and schedules large models with up to 1.49x cluster throughput improvement.
- Published
- 2024
16. Hierarchical Frequency-based Upsampling and Refining for Compressed Video Quality Enhancement
- Author
-
Zhang, Qianyu, Zheng, Bolun, Chen, Xinying, Chen, Quan, Zhu, Zhunjie, Wang, Canjin, Li, Zongpeng, and Yan, Chengang
- Subjects
Electrical Engineering and Systems Science - Image and Video Processing ,Computer Science - Computer Vision and Pattern Recognition - Abstract
Video compression artifacts arise due to the quantization operation in the frequency domain. The goal of video quality enhancement is to reduce compression artifacts and reconstruct a visually-pleasant result. In this work, we propose a hierarchical frequency-based upsampling and refining neural network (HFUR) for compressed video quality enhancement. HFUR consists of two modules: implicit frequency upsampling module (ImpFreqUp) and hierarchical and iterative refinement module (HIR). ImpFreqUp exploits DCT-domain prior derived through implicit DCT transform, and accurately reconstructs the DCT-domain loss via a coarse-to-fine transfer. Consequently, HIR is introduced to facilitate cross-collaboration and information compensation between the scales, thus further refine the feature maps and promote the visual quality of the final output. We demonstrate the effectiveness of the proposed modules via ablation experiments and visualized results. Extensive experiments on public benchmarks show that HFUR achieves state-of-the-art performance for both constant bit rate and constant QP modes.
- Published
- 2024
17. Experimental Quantum Byzantine Agreement on a Three-User Quantum Network with Integrated Photonics
- Author
-
Jing, Xu, Qian, Cheng, Weng, Chen-Xun, Li, Bing-Hong, Chen, Zhe, Wang, Chen-Quan, Tang, Jie, Gu, Xiao-Wen, Kong, Yue-Chan, Chen, Tang-Sheng, Yin, Hua-Lei, Jiang, Dong, Niu, Bin, and Lu, Liang-Liang
- Subjects
Quantum Physics - Abstract
Quantum communication networks are crucial for both secure communication and cryptographic networked tasks. Building quantum communication networks in a scalable and cost-effective way is essential for their widespread adoption, among which a stable and miniaturized high-quality quantum light source is a key component. Here, we establish a complete polarization entanglement-based fully connected network, which features an ultrabright integrated Bragg reflection waveguide quantum source, managed by an untrusted service provider, and a streamlined polarization analysis module, which requires only one single-photon detector for each end user. We perform a continuously working quantum entanglement distribution and create correlated bit strings between users. Within the framework of one-time universal hashing, we provide the first experimental implementation of source-independent quantum digital signatures using imperfect keys circumventing the necessity for private amplification. More importantly, we further beat the 1/3 fault-tolerance bound in Byzantine agreement, achieving unconditional security without relying on sophisticated techniques. Our results offer an affordable and practical route for addressing consensus challenges within the emerging quantum network landscape.
- Published
- 2024
- Full Text
- View/download PDF
18. Knowledge Condensation and Reasoning for Knowledge-based VQA
- Author
-
Hao, Dongze, Jia, Jian, Guo, Longteng, Wang, Qunbo, Yang, Te, Li, Yan, Cheng, Yanhua, Wang, Bo, Chen, Quan, Li, Han, and Liu, Jing
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Knowledge-based visual question answering (KB-VQA) is a challenging task, which requires the model to leverage external knowledge for comprehending and answering questions grounded in visual content. Recent studies retrieve the knowledge passages from external knowledge bases and then use them to answer questions. However, these retrieved knowledge passages often contain irrelevant or noisy information, which limits the performance of the model. To address the challenge, we propose two synergistic models: Knowledge Condensation model and Knowledge Reasoning model. We condense the retrieved knowledge passages from two perspectives. First, we leverage the multimodal perception and reasoning ability of the visual-language models to distill concise knowledge concepts from retrieved lengthy passages, ensuring relevance to both the visual content and the question. Second, we leverage the text comprehension ability of the large language models to summarize and condense the passages into the knowledge essence which helps answer the question. These two types of condensed knowledge are then seamlessly integrated into our Knowledge Reasoning model, which judiciously navigates through the amalgamated information to arrive at the conclusive answer. Extensive experiments validate the superiority of the proposed method. Compared to previous methods, our method achieves state-of-the-art performance on knowledge-based VQA datasets (65.1% on OK-VQA and 60.1% on A-OKVQA) without resorting to the knowledge produced by GPT-3 (175B).
- Published
- 2024
19. SDPL: Shifting-Dense Partition Learning for UAV-View Geo-Localization
- Author
-
Chen, Quan, Wang, Tingyu, Yang, Zihao, Li, Haoran, Lu, Rongfeng, Sun, Yaoqi, Zheng, Bolun, and Yan, Chenggang
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Cross-view geo-localization aims to match images of the same target from different platforms, e.g., drone and satellite. It is a challenging task due to the changing appearance of targets and environmental content from different views. Most methods focus on obtaining more comprehensive information through feature map segmentation, while inevitably destroying the image structure, and are sensitive to the shifting and scale of the target in the query. To address the above issues, we introduce simple yet effective part-based representation learning, shifting-dense partition learning (SDPL). We propose a dense partition strategy (DPS), dividing the image into multiple parts to explore contextual information while explicitly maintaining the global structure. To handle scenarios with non-centered targets, we further propose the shifting-fusion strategy, which generates multiple sets of parts in parallel based on various segmentation centers, and then adaptively fuses all features to integrate their anti-offset ability. Extensive experiments show that SDPL is robust to position shifting, and performs com-petitively on two prevailing benchmarks, University-1652 and SUES-200. In addition, SDPL shows satisfactory compatibility with a variety of backbone networks (e.g., ResNet and Swin). https://github.com/C-water/SDPL release., Comment: IEEE TCSVT 2024
- Published
- 2024
- Full Text
- View/download PDF
20. De novo protein design with a denoising diffusion network independent of pretrained structure prediction models
- Author
-
Liu, Yufeng, Wang, Sheng, Dong, Jixin, Chen, Linghui, Wang, Xinyu, Wang, Lei, Li, Fudong, Wang, Chenchen, Zhang, Jiahai, Wang, Yuzhu, Wei, Si, Chen, Quan, and Liu, Haiyan
- Published
- 2024
- Full Text
- View/download PDF
21. Towards real-time non-preemptive multicast scheduling in reconfigurable data center networks
- Author
-
Zhang, Fanlong, Liu, Jianglong, Wu, Yuhang, Chen, Quan, Chai, Yuan, and Wang, Zhuowei
- Published
- 2024
- Full Text
- View/download PDF
22. Aerial-view geo-localization based on multi-layer local pattern cross-attention network
- Author
-
Li, Haoran, Wang, Tingyu, Chen, Quan, Zhao, Qiang, Jiang, Shaowei, Yan, Chenggang, and Zheng, Bolun
- Published
- 2024
- Full Text
- View/download PDF
23. Distributed and latency-aware beaconing for asynchronous duty-cycled IoT networks
- Author
-
Yi, Ming, Xie, Qinglin, Long, Peng, Wu, Yuhang, Chen, Quan, Zhang, Fanlong, and Xu, Wenchao
- Published
- 2024
- Full Text
- View/download PDF
24. Identification of Anoikis-Related Genes in Spinal Cord Injury: Bioinformatics and Experimental Validation
- Author
-
Yin, Wen, Jiang, Zhipeng, Guo, Youwei, Cao, Yudong, Wu, Zhaoping, Zhou, Yi, Chen, Quan, Liu, Weidong, Jiang, Xingjun, and Ren, Caiping
- Published
- 2024
- Full Text
- View/download PDF
25. Elongational Flow-induced Crystallization of Poly(L-lactic acid) Telechelic Ionomers
- Author
-
Liu, Fan, Huang, Shao-Yong, Tang, Jian, and Chen, Quan
- Published
- 2024
- Full Text
- View/download PDF
26. Linear Viscoelasticity of ABA-type Vitrimer Based on Dioxaborolane Metathesis
- Author
-
Wu, Shi-Long, Yang, Huan-Huan, and Chen, Quan
- Published
- 2024
- Full Text
- View/download PDF
27. Nonlinear damping of associative polymers
- Author
-
Pei, Yuxuan, Zhang, Yanjie, Zheng, Chengzhi, Tang, Jian, and Chen, Quan
- Published
- 2024
- Full Text
- View/download PDF
28. EXACT-Net:EHR-guided lung tumor auto-segmentation for non-small cell lung cancer radiotherapy
- Author
-
Hooshangnejad, Hamed, Feng, Xue, Huang, Gaofeng, Zhang, Rui, Kelly, Katelyn, Chen, Quan, and Ding, Kai
- Subjects
Electrical Engineering and Systems Science - Image and Video Processing ,Computer Science - Computer Vision and Pattern Recognition ,Physics - Medical Physics - Abstract
Lung cancer is a devastating disease with the highest mortality rate among cancer types. Over 60% of non-small cell lung cancer (NSCLC) patients, which accounts for 87% of diagnoses, require radiation therapy. Rapid treatment initiation significantly increases the patient's survival rate and reduces the mortality rate. Accurate tumor segmentation is a critical step in the diagnosis and treatment of NSCLC. Manual segmentation is time and labor-consuming and causes delays in treatment initiation. Although many lung nodule detection methods, including deep learning-based models, have been proposed, there is still a long-standing problem of high false positives (FPs) with most of these methods. Here, we developed an electronic health record (EHR) guided lung tumor auto-segmentation called EXACT-Net (EHR-enhanced eXACtitude in Tumor segmentation), where the extracted information from EHRs using a pre-trained large language model (LLM), was used to remove the FPs and keep the TP nodules only. The auto-segmentation model was trained on NSCLC patients' computed tomography (CT), and the pre-trained LLM was used with the zero-shot learning approach. Our approach resulted in a 250% boost in successful nodule detection using the data from ten NSCLC patients treated in our institution.
- Published
- 2024
29. Accelerating Sparse DNNs Based on Tiled GEMM
- Author
-
Guo, Cong, Xue, Fengchen, Leng, Jingwen, Qiu, Yuxian, Guan, Yue, Cui, Weihao, Chen, Quan, and Guo, Minyi
- Subjects
Computer Science - Distributed, Parallel, and Cluster Computing - Abstract
Network pruning can reduce the computation cost of deep neural network (DNN) models. However, sparse models often produce randomly-distributed weights to maintain accuracy, leading to irregular computations. Consequently, unstructured sparse models cannot achieve meaningful speedup on commodity hardware built for dense matrix computations. Accelerators are usually modified or designed with structured sparsity-optimized architectures for exploiting sparsity. For example, the Ampere architecture introduces a sparse tensor core, which adopts the 2:4 sparsity pattern. We propose a pruning method that builds upon the insight that matrix multiplication generally breaks the large matrix into multiple smaller tiles for parallel execution. We present the tile-wise sparsity pattern, which maintains a structured sparsity pattern at the tile level for efficient execution but allows for irregular pruning at the global scale to maintain high accuracy. In addition, the tile-wise sparsity is implemented at the global memory level, and the 2:4 sparsity executes at the register level inside the sparse tensor core. We can combine these two patterns into a tile-vector-wise (TVW) sparsity pattern to explore more fine-grained sparsity and further accelerate the sparse DNN models. We evaluate the TVW on the GPU, achieving averages of $1.85\times$, $2.75\times$, and $22.18\times$ speedups over the dense model, block sparsity, and unstructured sparsity., Comment: Accepted by IEEE Transactions on Computers. arXiv admin note: substantial text overlap with arXiv:2008.13006
- Published
- 2024
30. (A)I Am Not a Lawyer, But...: Engaging Legal Experts towards Responsible LLM Policies for Legal Advice
- Author
-
Cheong, Inyoung, Xia, King, Feng, K. J. Kevin, Chen, Quan Ze, and Zhang, Amy X.
- Subjects
Computer Science - Computers and Society ,Computer Science - Artificial Intelligence - Abstract
Large language models (LLMs) are increasingly capable of providing users with advice in a wide range of professional domains, including legal advice. However, relying on LLMs for legal queries raises concerns due to the significant expertise required and the potential real-world consequences of the advice. To explore \textit{when} and \textit{why} LLMs should or should not provide advice to users, we conducted workshops with 20 legal experts using methods inspired by case-based reasoning. The provided realistic queries ("cases") allowed experts to examine granular, situation-specific concerns and overarching technical and legal constraints, producing a concrete set of contextual considerations for LLM developers. By synthesizing the factors that impacted LLM response appropriateness, we present a 4-dimension framework: (1) User attributes and behaviors, (2) Nature of queries, (3) AI capabilities, and (4) Social impacts. We share experts' recommendations for LLM response strategies, which center around helping users identify `right questions to ask' and relevant information rather than providing definitive legal judgments. Our findings reveal novel legal considerations, such as unauthorized practice of law, confidentiality, and liability for inaccurate advice, that have been overlooked in the literature. The case-based deliberation method enabled us to elicit fine-grained, practice-informed insights that surpass those from de-contextualized surveys or speculative principles. These findings underscore the applicability of our method for translating domain-specific professional knowledge and practices into policies that can guide LLM behavior in a more responsible direction., Comment: 14 pages
- Published
- 2024
31. An Optimizing Framework on MLIR for Efficient FPGA-based Accelerator Generation
- Author
-
Zhang, Weichuang, Zhao, Jieru, Shen, Guan, Chen, Quan, Chen, Chen, and Guo, Minyi
- Subjects
Computer Science - Hardware Architecture ,Computer Science - Programming Languages - Abstract
With the increasing demand for computing capability given limited resource and power budgets, it is crucial to deploy applications to customized accelerators like FPGAs. However, FPGA programming is non-trivial. Although existing high-level synthesis (HLS) tools improve productivity to a certain extent, they are limited in scope and capability to support sufficient FPGA-oriented optimizations. This paper focuses on FPGA-based accelerators and proposes POM, an optimizing framework built on multi-level intermediate representation (MLIR). POM has several features which demonstrate its scope and capability of performance optimization. First, most HLS tools depend exclusively on a single-level IR to perform all the optimizations, introducing excessive information into the IR and making debugging an arduous task. In contrast, POM introduces three layers of IR to perform operations at suitable abstraction levels, streamlining the implementation and debugging process and exhibiting better flexibility, extensibility, and systematicness. Second, POM integrates the polyhedral model into MLIR, enabling advanced dependence analysis and various FPGA-oriented loop transformations. By representing nested loops with integer sets and maps, loop transformations can be conducted conveniently through manipulations on polyhedral semantics. Finally, to further relieve design effort, POM has a user-friendly programming interface (DSL) that allows a concise description of computation and includes a rich collection of scheduling primitives. An automatic design space exploration (DSE) engine is provided to search for high-performance optimization schemes efficiently and generate optimized accelerators automatically. Experimental results show that POM achieves a $6.46\times$ average speedup on typical benchmark suites and a $6.06\times$ average speedup on real-world applications compared to the state-of-the-art., Comment: Accepted by HPCA2024
- Published
- 2024
32. Towards Efficient and Effective Text-to-Video Retrieval with Coarse-to-Fine Visual Representation Learning
- Author
-
Tian, Kaibin, Cheng, Yanhua, Liu, Yi, Hou, Xinglin, Chen, Quan, and Li, Han
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
In recent years, text-to-video retrieval methods based on CLIP have experienced rapid development. The primary direction of evolution is to exploit the much wider gamut of visual and textual cues to achieve alignment. Concretely, those methods with impressive performance often design a heavy fusion block for sentence (words)-video (frames) interaction, regardless of the prohibitive computation complexity. Nevertheless, these approaches are not optimal in terms of feature utilization and retrieval efficiency. To address this issue, we adopt multi-granularity visual feature learning, ensuring the model's comprehensiveness in capturing visual content features spanning from abstract to detailed levels during the training phase. To better leverage the multi-granularity features, we devise a two-stage retrieval architecture in the retrieval phase. This solution ingeniously balances the coarse and fine granularity of retrieval content. Moreover, it also strikes a harmonious equilibrium between retrieval effectiveness and efficiency. Specifically, in training phase, we design a parameter-free text-gated interaction block (TIB) for fine-grained video representation learning and embed an extra Pearson Constraint to optimize cross-modal representation learning. In retrieval phase, we use coarse-grained video representations for fast recall of top-k candidates, which are then reranked by fine-grained video representations. Extensive experiments on four benchmarks demonstrate the efficiency and effectiveness. Notably, our method achieves comparable performance with the current state-of-the-art methods while being nearly 50 times faster.
- Published
- 2024
33. Ferroptosis in health and disease.
- Author
-
Berndt, Carsten, Alborzinia, Hamed, Amen, Vera, Ayton, Scott, Barayeu, Uladzimir, Bartelt, Alexander, Bayir, Hülya, Bebber, Christina, Birsoy, Kivanc, Böttcher, Jan, Brabletz, Simone, Brabletz, Thomas, Brown, Ashley, Brüne, Bernhard, Bulli, Giorgia, Bruneau, Alix, Chen, Quan, DeNicola, Gina, Dick, Tobias, Distéfano, Ayelén, Dixon, Scott, Engler, Jan, Esser-von Bieren, Julia, Fedorova, Maria, Friedmann Angeli, José, Friese, Manuel, Fuhrmann, Dominic, García-Sáez, Ana, Garbowicz, Karolina, Götz, Magdalena, Gu, Wei, Hammerich, Linda, Hassannia, Behrouz, Jiang, Xuejun, Jeridi, Aicha, Kang, Yun, Kagan, Valerian, Konrad, David, Kotschi, Stefan, Lei, Peng, Le Tertre, Marlène, Lev, Sima, Liang, Deguang, Linkermann, Andreas, Lohr, Carolin, Lorenz, Svenja, Luedde, Tom, Methner, Axel, Michalke, Bernhard, Milton, Anna, Min, Junxia, Mishima, Eikan, Müller, Sebastian, Motohashi, Hozumi, Muckenthaler, Martina, Murakami, Shohei, Olzmann, James, Pagnussat, Gabriela, Pan, Zijan, Papagiannakopoulos, Thales, Pedrera Puentes, Lohans, Pratt, Derek, Proneth, Bettina, Ramsauer, Lukas, Rodriguez, Raphael, Saito, Yoshiro, Schmidt, Felix, Schmitt, Carina, Schulze, Almut, Schwab, Annemarie, Schwantes, Anna, Soula, Mariluz, Spitzlberger, Benedikt, Stockwell, Brent, Thewes, Leonie, Thorn-Seshold, Oliver, Toyokuni, Shinya, Tonnus, Wulf, Trumpp, Andreas, Vandenabeele, Peter, Vanden Berghe, Tom, Venkataramani, Vivek, Vogel, Felix, von Karstedt, Silvia, Wang, Fudi, Westermann, Frank, Wientjens, Chantal, Wilhelm, Christoph, Wölk, Michele, Wu, Katherine, Yang, Xin, Yu, Fan, Zou, Yilong, and Conrad, Marcus
- Subjects
Cancer ,Cell death ,Iron ,Ischemia/reperfusion ,Lipid peroxidation ,Neurodegeneration - Abstract
Ferroptosis is a pervasive non-apoptotic form of cell death highly relevant in various degenerative diseases and malignancies. The hallmark of ferroptosis is uncontrolled and overwhelming peroxidation of polyunsaturated fatty acids contained in membrane phospholipids, which eventually leads to rupture of the plasma membrane. Ferroptosis is unique in that it is essentially a spontaneous, uncatalyzed chemical process based on perturbed iron and redox homeostasis contributing to the cell death process, but that it is nonetheless modulated by many metabolic nodes that impinge on the cells susceptibility to ferroptosis. Among the various nodes affecting ferroptosis sensitivity, several have emerged as promising candidates for pharmacological intervention, rendering ferroptosis-related proteins attractive targets for the treatment of numerous currently incurable diseases. Herein, the current members of a Germany-wide research consortium focusing on ferroptosis research, as well as key external experts in ferroptosis who have made seminal contributions to this rapidly growing and exciting field of research, have gathered to provide a comprehensive, state-of-the-art review on ferroptosis. Specific topics include: basic mechanisms, in vivo relevance, specialized methodologies, chemical and pharmacological tools, and the potential contribution of ferroptosis to disease etiopathology and progression. We hope that this article will not only provide established scientists and newcomers to the field with an overview of the multiple facets of ferroptosis, but also encourage additional efforts to characterize further molecular pathways modulating ferroptosis, with the ultimate goal to develop novel pharmacotherapies to tackle the various diseases associated with - or caused by - ferroptosis.
- Published
- 2024
34. Auto-Segmentation of Elective Nodal Clinical Target Volumes for Anal Cancer Using Artificial Intelligence
- Author
-
Law, Bryant, Aldin, Muhtada, Stone, Payton, Rong, Yi, Chen, Quan, Park, Peter, Hunt, Jon-Paul, and Monjazeb, Arta
- Published
- 2024
35. Mobility and Cost Aware Inference Accelerating Algorithm for Edge Intelligence
- Author
-
Yuan, Xin, Li, Ning, Wei, kang, Xu, Wenchao, Chen, Quan, Chen, Hao, and Guo, Song
- Subjects
Computer Science - Distributed, Parallel, and Cluster Computing ,Computer Science - Artificial Intelligence - Abstract
The edge intelligence (EI) has been widely applied recently. Spliting the model between device, edge server, and cloud can improve the performance of EI greatly. The model segmentation without user mobility has been investigated deeply by previous works. However, in most use cases of EI, the end devices are mobile. Only a few works have been carried out on this aspect. These works still have many issues, such as ignoring the energy consumption of mobile device, inappropriate network assumption, and low effectiveness on adaptiving user mobility, etc. Therefore, for addressing the disadvantages of model segmentation and resource allocation in previous works, we propose mobility and cost aware model segmentation and resource allocation algorithm for accelerating the inference at edge (MCSA). Specfically, in the scenario without user mobility, the loop interation gradient descent (Li-GD) algorithm is provided. When the mobile user has a large model inference task needs to be calculated, it will take the energy consumption of mobile user, the communication and computing resource renting cost, and the inference delay into account to find the optimal model segmentation and resource allocation strategy. In the scenario with user mobility, the mobiity aware Li-GD (MLi-GD) algorithm is proposed to calculate the optimal strategy. Then, the properties of the proposed algorithms are investigated, including convergence, complexity, and approximation ratio. The experimental results demonstrate the effectiveness of the proposed algorithms., Comment: 17 pages, 16 figures. arXiv admin note: substantial text overlap with arXiv:2312.15850
- Published
- 2023
36. FODT: Fast, Online, Distributed and Temporary Failure Recovery Approach for MEC
- Author
-
Yuan, Xin, Li, Ning, Zhang, Zhaoxin, Chen, Quan, and Martinez, Jose Fernan
- Subjects
Computer Science - Distributed, Parallel, and Cluster Computing ,Computer Science - Networking and Internet Architecture - Abstract
Mobile edge computing (MEC) can reduce the latency of cloud computing successfully. However, the edge server may fail due to the hardware of software issues. When the edge server failure happens, the users who offload tasks to this server will be affected. How to recover the services for these affected users quickly and effectively is challenging. Moreover, considering that the server failure is continuous and temporary, and the failed server can be repaired, the previous works cannot handle this problem effectively. Therefore, in this paper, we propose the fast, online, distributed, and temporary failure recovery algorithm (FODT) for MEC. In FODT, when edge sever failure happens, only the affected APs recalculate their user-server allocation strategies and the other APs do not change their strategies. For the affected access points (Aps), the strategies before server failure are reused to reduce complexity and latency. When the failed server is repaired, the influenced APs reuse the strategies before server failure to offload task to this server. Based on this approach, the FODT can achieve better performance than previous works. To the best of knowledge, the FODT is the first failure recovery algorithm, and when compared with previous research, it has higher failure recovery efficiency and lower complexity with acceptable approximate ratio., Comment: 12 pages, 7 figures
- Published
- 2023
37. High-resolution myelin-water fraction and quantitative relaxation mapping using 3D ViSTa-MR fingerprinting
- Author
-
Liao, Congyu, Cao, Xiaozhi, Iyer, Siddharth Srinivasan, Schauman, Sophie, Zhou, Zihan, Yan, Xiaoqian, Chen, Quan, Li, Zhitao, Wang, Nan, Gong, Ting, Wu, Zhe, He, Hongjian, Zhong, Jianhui, Yang, Yang, Kerr, Adam, Grill-Spector, Kalanit, and Setsompop, Kawin
- Subjects
Physics - Medical Physics ,Electrical Engineering and Systems Science - Image and Video Processing - Abstract
Purpose: This study aims to develop a high-resolution whole-brain multi-parametric quantitative MRI approach for simultaneous mapping of myelin-water fraction (MWF), T1, T2, and proton-density (PD), all within a clinically feasible scan time. Methods: We developed 3D ViSTa-MRF, which combined Visualization of Short Transverse relaxation time component (ViSTa) technique with MR Fingerprinting (MRF), to achieve high-fidelity whole-brain MWF and T1/T2/PD mapping on a clinical 3T scanner. To achieve fast acquisition and memory-efficient reconstruction, the ViSTa-MRF sequence leverages an optimized 3D tiny-golden-angle-shuffling spiral-projection acquisition and joint spatial-temporal subspace reconstruction with optimized preconditioning algorithm. With the proposed ViSTa-MRF approach, high-fidelity direct MWF mapping was achieved without a need for multi-compartment fitting that could introduce bias and/or noise from additional assumptions or priors. Results: The in-vivo results demonstrate the effectiveness of the proposed acquisition and reconstruction framework to provide fast multi-parametric mapping with high SNR and good quality. The in-vivo results of 1mm- and 0.66mm-iso datasets indicate that the MWF values measured by the proposed method are consistent with standard ViSTa results that are 30x slower with lower SNR. Furthermore, we applied the proposed method to enable 5-minute whole-brain 1mm-iso assessment of MWF and T1/T2/PD mappings for infant brain development and for post-mortem brain samples. Conclusions: In this work, we have developed a 3D ViSTa-MRF technique that enables the acquisition of whole-brain MWF, quantitative T1, T2, and PD maps at 1mm and 0.66mm isotropic resolution in 5 and 15 minutes, respectively. This advancement allows for quantitative investigations of myelination changes in the brain., Comment: 38 pages, 12 figures and 1 table
- Published
- 2023
- Full Text
- View/download PDF
38. JUNO: Optimizing High-Dimensional Approximate Nearest Neighbour Search with Sparsity-Aware Algorithm and Ray-Tracing Core Mapping
- Author
-
Liu, Zihan, Ni, Wentao, Leng, Jingwen, Feng, Yu, Guo, Cong, Chen, Quan, Li, Chao, Guo, Minyi, and Zhu, Yuhao
- Subjects
Computer Science - Distributed, Parallel, and Cluster Computing - Abstract
Approximate nearest neighbor (ANN) search is a widely applied technique in modern intelligent applications, such as recommendation systems and vector databases. Therefore, efficient and high-throughput execution of ANN search has become increasingly important. In this paper, we first characterize the state-of-the-art product quantization-based method of ANN search and identify a significant source of inefficiency in the form of unnecessary pairwise distance calculations and accumulations. To improve efficiency, we propose JUNO, an end-to-end ANN search system that adopts a carefully designed sparsity- and locality-aware search algorithm. We also present an efficient hardware mapping that utilizes ray tracing cores in modern GPUs with pipelined execution on tensor cores to execute our sparsity-aware ANN search algorithm. Our evaluations on four datasets ranging in size from 1 to 100 million search points demonstrate 2.2x-8.5x improvements in search throughput. Moreover, our algorithmic enhancements alone achieve a maximal 2.6x improvement on the hardware without the acceleration of the RT core.
- Published
- 2023
39. Bright and dark solitons under spatiotemporal modulation in (2+1)-dimensional Spin-1 Bose-Einstein condensates
- Author
-
Li, Nan, Xu, Suyong, Sun, Yunzhou, and Chen, Quan
- Published
- 2025
- Full Text
- View/download PDF
40. Efficient Online Path Selection and Workload Allocation for In-Network Computing in MEC
- Author
-
Ouyang, Sheng, Zhang, Fanlong, Mai, Junyu, Chai, Yuan, Chen, Quan, Tao, Yongchao, Goos, Gerhard, Series Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Cai, Zhipeng, editor, Takabi, Daniel, editor, Guo, Shaoyong, editor, and Zou, Yifei, editor
- Published
- 2025
- Full Text
- View/download PDF
41. TRB3 Promotes Cataract Progression through Endoplasmic Reticulum Stress-mediated Mitochondrial Dysfunction and Cell Apoptosis
- Author
-
Liu, Junyi, Tang, Yongying, Li, Jiang, Zhang, Hong, Zha, Xu, Chen, Quan, Li, Jinghua, and Zhao, Xueying
- Published
- 2024
- Full Text
- View/download PDF
42. PAFAH2 suppresses synchronized ferroptosis to ameliorate acute kidney injury
- Author
-
Zhang, Qianping, Sun, Tiantian, Yu, Fan, Liu, Wei, Gao, Jin, Chen, Jinyu, Zheng, Hao, Liu, Jinming, Miao, Chenjian, Guo, Huanyi, Tian, Wu, Su, Meihui, Guo, Yingjie, Liu, Xi, Pei, Yandong, Wang, Zhuofei, Chen, Shang, Mu, Chenglong, Lam, Sin Man, Shui, Guanghou, Li, Zongjin, Yu, Zhongbo, Zhang, Yan, Chen, Guo, Lu, Congcong, Midgley, Adam C., Li, Changhua, Bian, Xin, Liao, Xudong, Wang, Yong, Xiong, Wei, Zhu, Hongying, Li, Yanjun, and Chen, Quan
- Published
- 2024
- Full Text
- View/download PDF
43. A categorical Connes’ χ(M)
- Author
-
Chen, Quan, Jones, Corey, and Penneys, David
- Published
- 2024
- Full Text
- View/download PDF
44. Research on new edge computing network architecture and task offloading strategy for Internet of Things
- Author
-
Jiang, Congshi, Li, Yihong, Su, Junlong, and Chen, Quan
- Published
- 2024
- Full Text
- View/download PDF
45. A 5,6,11,12,17,18-Hexaazatrinaphthylene-Based Luminescent Metal-Organic Framework as Ornidazole Sensor with Extremely High Sensitivity
- Author
-
Zhang, Jinfang, Chen, Quan, Chen, Ling, Yin, Dejing, and Zhang, Chi
- Published
- 2024
- Full Text
- View/download PDF
46. Case Repositories: Towards Case-Based Reasoning for AI Alignment
- Author
-
Feng, K. J. Kevin, Chen, Quan Ze, Cheong, Inyoung, Xia, King, and Zhang, Amy X.
- Subjects
Computer Science - Artificial Intelligence ,Computer Science - Computers and Society ,Computer Science - Human-Computer Interaction - Abstract
Case studies commonly form the pedagogical backbone in law, ethics, and many other domains that face complex and ambiguous societal questions informed by human values. Similar complexities and ambiguities arise when we consider how AI should be aligned in practice: when faced with vast quantities of diverse (and sometimes conflicting) values from different individuals and communities, with whose values is AI to align, and how should AI do so? We propose a complementary approach to constitutional AI alignment, grounded in ideas from case-based reasoning (CBR), that focuses on the construction of policies through judgments on a set of cases. We present a process to assemble such a case repository by: 1) gathering a set of ``seed'' cases -- questions one may ask an AI system -- in a particular domain from discussions in online communities, 2) eliciting domain-specific key dimensions for cases through workshops with domain experts, 3) using LLMs to generate variations of cases not seen in the wild, and 4) engaging with the public to judge and improve cases. We then discuss how such a case repository could assist in AI alignment, both through directly acting as precedents to ground acceptable behaviors, and as a medium for individuals and communities to engage in moral reasoning around AI., Comment: MP2 workshop @ NeurIPS 2023
- Published
- 2023
47. Adaptive CPU Resource Allocation for Emulator in Kernel-based Virtual Machine
- Author
-
Yang, Yecheng, Pang, Pu, Wang, Jiawen, Chen, Quan, and Guo, Minyi
- Subjects
Computer Science - Operating Systems - Abstract
The technologies of heterogeneous multi-core architectures, co-location, and virtualization can be used to reduce server power consumption and improve system utilization, which are three important technologies for data centers. This article explores the scheduling strategy of Emulator threads within virtual machine processes in a scenario of co-location of multiple virtual machines on heterogeneous multi-core architectures. In this co-location scenario, the scheduling strategy for Emulator threads significantly affects the performance of virtual machines. This article focuses on this thread for the first time in the relevant field. This article found that the scheduling latency metric can well indicate the running status of the vCPU threads and Emulator threads in the virtualization environment, and applied this metric to the design of the scheduling strategy. This article designed an Emulator thread scheduler based on heuristic rules, which, in coordination with the host operating system's scheduler, dynamically adjusts the scheduling scope of Emulator threads to improve the overall performance of virtual machines. The article found that in real application scenarios, the scheduler effectively improved the performance of applications within virtual machines, with a maximum performance improvement of 40.7%.
- Published
- 2023
48. Case Law Grounding: Using Precedents to Align Decision-Making for Humans and AI
- Author
-
Chen, Quan Ze and Zhang, Amy X.
- Subjects
Computer Science - Human-Computer Interaction - Abstract
Communities and groups often need to make decisions based on social norms and preferences, such as when moderating content or building AI systems that reflect human values. The prevailing approach has been to first create high-level guidelines -- ``constitutions'' -- and then decide on new cases using the outlined criteria. However, social norms and preferences vary between groups, decision-makers can interpret guidelines inconsistently, and exceptional situations may be under-specified. In this work, we take inspiration from legal systems and introduce ``case law grounding'' (CLG), a novel workflow that uses past cases and decisions (\textbf{precedents}) to help ground future decisions, for both human and LLM-based decision-makers. We evaluate CLG against a constitution-only approach on two tasks for both types of decision-makers, and find that decisions produced with CLG were more accurately aligned to observed ground truth in all cases, producing a 3.3--23.3 \%-points improvement (across different tasks and groups) for humans and 9.2--30.0 \%-points (across different tasks and groups) for LLM agents. We also discuss other aspects where a case-based approach could augment existing ``constitutional'' approaches when it comes to aligning human and AI decisions.
- Published
- 2023
49. STAG: Enabling Low Latency and Low Staleness of GNN-based Services with Dynamic Graphs
- Author
-
Wang, Jiawen, Chen, Quan, Zeng, Deze, Song, Zhuo, Chen, Chen, and Guo, Minyi
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence - Abstract
Many emerging user-facing services adopt Graph Neural Networks (GNNs) to improve serving accuracy. When the graph used by a GNN model changes, representations (embedding) of nodes in the graph should be updated accordingly. However, the node representation update is too slow, resulting in either long response latency of user queries (the inference is performed after the update completes) or high staleness problem (the inference is performed based on stale data). Our in-depth analysis shows that the slow update is mainly due to neighbor explosion problem in graphs and duplicated computation. Based on such findings, we propose STAG, a GNN serving framework that enables low latency and low staleness of GNN-based services. It comprises a collaborative serving mechanism and an additivity-based incremental propagation strategy. With the collaborative serving mechanism, only part of node representations are updated during the update phase, and the final representations are calculated in the inference phase. It alleviates the neighbor explosion problem. The additivity-based incremental propagation strategy reuses intermediate data during the update phase, eliminating duplicated computation problem. Experimental results show that STAG accelerates the update phase by 1.3x~90.1x, and greatly reduces staleness time with a slight increase in response latency.
- Published
- 2023
50. Confidence Contours: Uncertainty-Aware Annotation for Medical Semantic Segmentation
- Author
-
Ye, Andre, Chen, Quan Ze, and Zhang, Amy
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Human-Computer Interaction - Abstract
Medical image segmentation modeling is a high-stakes task where understanding of uncertainty is crucial for addressing visual ambiguity. Prior work has developed segmentation models utilizing probabilistic or generative mechanisms to infer uncertainty from labels where annotators draw a singular boundary. However, as these annotations cannot represent an individual annotator's uncertainty, models trained on them produce uncertainty maps that are difficult to interpret. We propose a novel segmentation representation, Confidence Contours, which uses high- and low-confidence ``contours'' to capture uncertainty directly, and develop a novel annotation system for collecting contours. We conduct an evaluation on the Lung Image Dataset Consortium (LIDC) and a synthetic dataset. From an annotation study with 30 participants, results show that Confidence Contours provide high representative capacity without considerably higher annotator effort. We also find that general-purpose segmentation models can learn Confidence Contours at the same performance level as standard singular annotations. Finally, from interviews with 5 medical experts, we find that Confidence Contour maps are more interpretable than Bayesian maps due to representation of structural uncertainty., Comment: 10 pages content, 12 pages total. Accepted to HCOMP '23
- Published
- 2023
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.