8,679 results on '"Wang, Hongwei"'
Search Results
2. Probabilistic Prior Driven Attention Mechanism Based on Diffusion Model for Imaging Through Atmospheric Turbulence
- Author
-
Sun, Guodong, Ma, Qixiang, Zhang, Liqiang, Wang, Hongwei, Gao, Zixuan, and Zhang, Haotian
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Atmospheric turbulence introduces severe spatial and geometric distortions, challenging traditional image restoration methods. We propose the Probabilistic Prior Turbulence Removal Network (PPTRN), which combines probabilistic diffusion-based prior modeling with Transformer-driven feature extraction to address this issue. PPTRN employs a two-stage approach: first, a latent encoder and Transformer are jointly trained on clear images to establish robust feature representations. Then, a Denoising Diffusion Probabilistic Model (DDPM) models prior distributions over latent vectors, guiding the Transformer in capturing diverse feature variations essential for restoration. A key innovation in PPTRN is the Probabilistic Prior Driven Cross Attention mechanism, which integrates the DDPM-generated prior with feature embeddings to reduce artifacts and enhance spatial coherence. Extensive experiments validate that PPTRN significantly improves restoration quality on turbulence-degraded images, setting a new benchmark in clarity and structural fidelity.
- Published
- 2024
3. Graph Pre-Training Models Are Strong Anomaly Detectors
- Author
-
Cheng, Jiashun, Zheng, Zinan, Liu, Yang, Tang, Jianheng, Wang, Hongwei, Rong, Yu, Li, Jia, and Tsung, Fugee
- Subjects
Computer Science - Machine Learning - Abstract
Graph Anomaly Detection (GAD) is a challenging and practical research topic where Graph Neural Networks (GNNs) have recently shown promising results. The effectiveness of existing GNNs in GAD has been mainly attributed to the simultaneous learning of node representations and the classifier in an end-to-end manner. Meanwhile, graph pre-training, the two-stage learning paradigm such as DGI and GraphMAE, has shown potential in leveraging unlabeled graph data to enhance downstream tasks, yet its impact on GAD remains under-explored. In this work, we show that graph pre-training models are strong graph anomaly detectors. Specifically, we demonstrate that pre-training is highly competitive, markedly outperforming the state-of-the-art end-to-end training models when faced with limited supervision. To understand this phenomenon, we further uncover pre-training enhances the detection of distant, under-represented, unlabeled anomalies that go beyond 2-hop neighborhoods of known anomalies, shedding light on its superior performance against end-to-end models. Moreover, we extend our examination to the potential of pre-training in graph-level anomaly detection. We envision this work to stimulate a re-evaluation of pre-training's role in GAD and offer valuable insights for future research.
- Published
- 2024
4. LongMemEval: Benchmarking Chat Assistants on Long-Term Interactive Memory
- Author
-
Wu, Di, Wang, Hongwei, Yu, Wenhao, Zhang, Yuwei, Chang, Kai-Wei, and Yu, Dong
- Subjects
Computer Science - Computation and Language - Abstract
Recent large language model (LLM)-driven chat assistant systems have integrated memory components to track user-assistant chat histories, enabling more accurate and personalized responses. However, their long-term memory capabilities in sustained interactions remain underexplored. This paper introduces LongMemEval, a comprehensive benchmark designed to evaluate five core long-term memory abilities of chat assistants: information extraction, multi-session reasoning, temporal reasoning, knowledge updates, and abstention. With 500 meticulously curated questions embedded within freely scalable user-assistant chat histories, LongMemEval presents a significant challenge to existing long-term memory systems, with commercial chat assistants and long-context LLMs showing 30% accuracy drop on memorizing information across sustained interactions. We then present a unified framework that breaks down the long-term memory design into four design choices across the indexing, retrieval, and reading stages. Built upon key experimental insights, we propose several memory designs including session decomposition for optimizing value granularity, fact-augmented key expansion for enhancing the index structure, and time-aware query expansion for refining the search scope. Experiment results show that these optimizations greatly improve both memory recall and downstream question answering on LongMemEval. Overall, our study provides valuable resources and guidance for advancing the long-term memory capabilities of LLM-based chat assistants, paving the way toward more personalized and reliable conversational AI.
- Published
- 2024
5. SAKA: An Intelligent Platform for Semi-automated Knowledge Graph Construction and Application
- Author
-
Zhang, Hanrong, Wang, Xinyue, Pan, Jiabao, and Wang, Hongwei
- Subjects
Computer Science - Artificial Intelligence - Abstract
Knowledge graph (KG) technology is extensively utilized in many areas, and many companies offer applications based on KG. Nonetheless, the majority of KG platforms necessitate expertise and tremendous time and effort of users to construct KG records manually, which poses great difficulties for ordinary people to use. Additionally, audio data is abundant and holds valuable information, but it is challenging to transform it into a KG. What's more, the platforms usually do not leverage the full potential of the KGs constructed by users. In this paper, we propose an intelligent and user-friendly platform for Semi-automated KG Construction and Application (SAKA) to address the problems aforementioned. Primarily, users can semi-automatically construct KGs from structured data of numerous areas by interacting with the platform, based on which multi-versions of KG can be stored, viewed, managed, and updated. Moreover, we propose an Audio-based KG Information Extraction (AGIE) method to establish KGs from audio data. Lastly, the platform creates a semantic parsing-based knowledge base question answering (KBQA) system based on the user-created KGs. We prove the feasibility of the semi-automatic KG construction method on the SAKA platform.
- Published
- 2024
6. Decentralized Clinical Trials in the Era of Real-World Evidence: A Statistical Perspective
- Author
-
Chen, Jie, Di, Junrui, Daizadeh, Nadia, Lu, Ying, Wang, Hongwei, Shen, Yuan-Li, Kirk, Jennifer, Rockhold, Frank W., Pang, Herbert, Zhao, Jing, He, Weili, Potter, Andrew, and Lee, Hana
- Subjects
Statistics - Applications - Abstract
There has been a growing trend that activities relating to clinical trials take place at locations other than traditional trial sites (hence decentralized clinical trials or DCTs), some of which are at settings of real-world clinical practice. Although there are numerous benefits of DCTs, this also brings some implications on a number of issues relating to the design, conduct, and analysis of DCTs. The Real-World Evidence Scientific Working Group of the American Statistical Association Biopharmaceutical Section has been reviewing the field of DCTs and provides in this paper considerations for decentralized trials from a statistical perspective. This paper first discusses selected critical decentralized elements that may have statistical implications on the trial and then summarizes regulatory guidance, framework, and initiatives on DCTs. More discussions are presented by focusing on the design (including construction of estimand), implementation, statistical analysis plan (including missing data handling), and reporting of safety events. Some additional considerations (e.g., ethical considerations, technology infrastructure, study oversight, data security and privacy, and regulatory compliance) are also briefly discussed. This paper is intended to provide statistical considerations for decentralized trials of medical products to support regulatory decision-making.
- Published
- 2024
7. Agent Security Bench (ASB): Formalizing and Benchmarking Attacks and Defenses in LLM-based Agents
- Author
-
Zhang, Hanrong, Huang, Jingyuan, Mei, Kai, Yao, Yifei, Wang, Zhenting, Zhan, Chenlu, Wang, Hongwei, and Zhang, Yongfeng
- Subjects
Computer Science - Cryptography and Security ,Computer Science - Artificial Intelligence - Abstract
Although LLM-based agents, powered by Large Language Models (LLMs), can use external tools and memory mechanisms to solve complex real-world tasks, they may also introduce critical security vulnerabilities. However, the existing literature does not comprehensively evaluate attacks and defenses against LLM-based agents. To address this, we introduce Agent Security Bench (ASB), a comprehensive framework designed to formalize, benchmark, and evaluate the attacks and defenses of LLM-based agents, including 10 scenarios (e.g., e-commerce, autonomous driving, finance), 10 agents targeting the scenarios, over 400 tools, 23 different types of attack/defense methods, and 8 evaluation metrics. Based on ASB, we benchmark 10 prompt injection attacks, a memory poisoning attack, a novel Plan-of-Thought backdoor attack, a mixed attack, and 10 corresponding defenses across 13 LLM backbones with nearly 90,000 testing cases in total. Our benchmark results reveal critical vulnerabilities in different stages of agent operation, including system prompt, user prompt handling, tool usage, and memory retrieval, with the highest average attack success rate of 84.30\%, but limited effectiveness shown in current defenses, unveiling important works to be done in terms of agent security for the community. Our code can be found at https://github.com/agiresearch/ASB.
- Published
- 2024
8. Cognitive Kernel: An Open-source Agent System towards Generalist Autopilots
- Author
-
Zhang, Hongming, Pan, Xiaoman, Wang, Hongwei, Ma, Kaixin, Yu, Wenhao, and Yu, Dong
- Subjects
Computer Science - Artificial Intelligence - Abstract
We introduce Cognitive Kernel, an open-source agent system towards the goal of generalist autopilots. Unlike copilot systems, which primarily rely on users to provide essential state information (e.g., task descriptions) and assist users by answering questions or auto-completing contents, autopilot systems must complete tasks from start to finish independently, which requires the system to acquire the state information from the environments actively. To achieve this, an autopilot system should be capable of understanding user intents, actively gathering necessary information from various real-world sources, and making wise decisions. Cognitive Kernel adopts a model-centric design. In our implementation, the central policy model (a fine-tuned LLM) initiates interactions with the environment using a combination of atomic actions, such as opening files, clicking buttons, saving intermediate results to memory, or calling the LLM itself. This differs from the widely used environment-centric design, where a task-specific environment with predefined actions is fixed, and the policy model is limited to selecting the correct action from a given set of options. Our design facilitates seamless information flow across various sources and provides greater flexibility. We evaluate our system in three use cases: real-time information management, private information management, and long-term memory management. The results demonstrate that Cognitive Kernel achieves better or comparable performance to other closed-source systems in these scenarios. Cognitive Kernel is fully dockerized, ensuring everyone can deploy it privately and securely. We open-source the system and the backbone model to encourage further research on LLM-driven autopilot systems.
- Published
- 2024
9. Line Spectral Estimation with Unlimited Sensing
- Author
-
Wang, Hongwei, Fang, Jun, Li, Hongbin, and Leus, Geert
- Subjects
Electrical Engineering and Systems Science - Signal Processing - Abstract
In the paper, we consider the line spectral estimation problem in an unlimited sensing framework (USF), where a modulo analog-to-digital converter (ADC) is employed to fold the input signal back into a bounded interval before quantization. Such an operation is mathematically equivalent to taking the modulo of the input signal with respect to the interval. To overcome the noise sensitivity of higher-order difference-based methods, we explore the properties of the first-order difference of modulo samples, and develop two line spectral estimation algorithms based on first-order difference, which are robust against noise. Specifically, we show that, with a high probability, the first-order difference of the original samples is equivalent to that of the modulo samples. By utilizing this property, line spectral estimation is solved via a robust sparse signal recovery approach. The second algorithms is built on our finding that, with a sufficiently high sampling rate, the first-order difference of the original samples can be decomposed as a sum of the first-order difference of the modulo samples and a sequence whose elements are confined to be three possible values. This decomposition enables us to formulate the line spectral estimation problem as a mixed integer linear program that can be efficiently solved. Simulation results show that both proposed methods are robust against noise and achieve a significant performance improvement over the higher-order difference-based method.
- Published
- 2024
10. Human-Machine Co-Adaptation for Robot-Assisted Rehabilitation via Dual-Agent Multiple Model Reinforcement Learning (DAMMRL)
- Author
-
An, Yang, Li, Yaqi, Wang, Hongwei, Duffield, Rob, and Su, Steven W.
- Subjects
Computer Science - Robotics - Abstract
This study introduces a novel approach to robot-assisted ankle rehabilitation by proposing a Dual-Agent Multiple Model Reinforcement Learning (DAMMRL) framework, leveraging multiple model adaptive control (MMAC) and co-adaptive control strategies. In robot-assisted rehabilitation, one of the key challenges is modelling human behaviour due to the complexity of human cognition and physiological systems. Traditional single-model approaches often fail to capture the dynamics of human-machine interactions. Our research employs a multiple model strategy, using simple sub-models to approximate complex human responses during rehabilitation tasks, tailored to varying levels of patient incapacity. The proposed system's versatility is demonstrated in real experiments and simulated environments. Feasibility and potential were evaluated with 13 healthy young subjects, yielding promising results that affirm the anticipated benefits of the approach. This study not only introduces a new paradigm for robot-assisted ankle rehabilitation but also opens the way for future research in adaptive, patient-centred therapeutic interventions.
- Published
- 2024
11. DynaThink: Fast or Slow? A Dynamic Decision-Making Framework for Large Language Models
- Author
-
Pan, Jiabao, Zhang, Yan, Zhang, Chen, Liu, Zuozhu, Wang, Hongwei, and Li, Haizhou
- Subjects
Computer Science - Computation and Language - Abstract
Large language models (LLMs) have demonstrated emergent capabilities across diverse reasoning tasks via popular Chains-of-Thought (COT) prompting. However, such a simple and fast COT approach often encounters limitations in dealing with complicated problems, while a thorough method, which considers multiple reasoning pathways and verifies each step carefully, results in slower inference. This paper addresses the challenge of enabling LLMs to autonomously select between fast and slow inference methods, thereby optimizing both efficiency and effectiveness. We introduce a dynamic decision-making framework that categorizes tasks into two distinct pathways: 'Fast', designated for tasks where the LLM quickly identifies a high-confidence solution, and 'Slow', allocated for tasks that the LLM perceives as complex and for which it has low confidence in immediate solutions as well as requiring more reasoning paths to verify. Experiments on five popular reasoning benchmarks demonstrated the superiority of the DynaThink over baselines.
- Published
- 2024
12. Retrieval Augmented Instruction Tuning for Open NER with Large Language Models
- Author
-
Xie, Tingyu, Zhang, Jian, Zhang, Yan, Liang, Yuanyuan, Li, Qi, and Wang, Hongwei
- Subjects
Computer Science - Computation and Language - Abstract
The strong capability of large language models (LLMs) has been applied to information extraction (IE) through either retrieval augmented prompting or instruction tuning (IT). However, the best way to incorporate information with LLMs for IE remains an open question. In this paper, we explore Retrieval Augmented Instruction Tuning (RA-IT) for IE, focusing on the task of open named entity recognition (NER). Specifically, for each training sample, we retrieve semantically similar examples from the training dataset as the context and prepend them to the input of the original instruction. To evaluate our RA-IT approach more thoroughly, we construct a Chinese IT dataset for open NER and evaluate RA-IT in both English and Chinese scenarios. Experimental results verify the effectiveness of RA-IT across various data sizes and in both English and Chinese scenarios. We also conduct thorough studies to explore the impacts of various retrieval strategies in the proposed RA-IT framework. Code and data are available at: https://github.com/Emma1066/Retrieval-Augmented-IT-OpenNER
- Published
- 2024
13. STEVE Series: Step-by-Step Construction of Agent Systems in Minecraft
- Author
-
Zhao, Zhonghan, Chai, Wenhao, Wang, Xuan, Ma, Ke, Chen, Kewei, Guo, Dongxu, Ye, Tian, Zhang, Yanting, Wang, Hongwei, and Wang, Gaoang
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Building an embodied agent system with a large language model (LLM) as its core is a promising direction. Due to the significant costs and uncontrollable factors associated with deploying and training such agents in the real world, we have decided to begin our exploration within the Minecraft environment. Our STEVE Series agents can complete basic tasks in a virtual environment and more challenging tasks such as navigation and even creative tasks, with an efficiency far exceeding previous state-of-the-art methods by a factor of $2.5\times$ to $7.3\times$. We begin our exploration with a vanilla large language model, augmenting it with a vision encoder and an action codebase trained on our collected high-quality dataset STEVE-21K. Subsequently, we enhanced it with a Critic and memory to transform it into a complex system. Finally, we constructed a hierarchical multi-agent system. Our recent work explored how to prune the agent system through knowledge distillation. In the future, we will explore more potential applications of STEVE agents in the real world., Comment: CVPR 2024 Embodied AI Workshop
- Published
- 2024
14. Baicalein ameliorates cadmium-induced hepatic and renal oxidative damage in rats
- Author
-
Wang, Jicang, Zhu, Huali, Zhang, Cai, Wang, Hongwei, and Yang, Zijun
- Published
- 2019
- Full Text
- View/download PDF
15. Towards Imperceptible Backdoor Attack in Self-supervised Learning
- Author
-
Zhang, Hanrong, Wang, Zhenting, Han, Tingxu, Jin, Mingyu, Zhan, Chenlu, Du, Mengnan, Wang, Hongwei, and Ma, Shiqing
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Self-supervised learning models are vulnerable to backdoor attacks. Existing backdoor attacks that are effective in self-supervised learning often involve noticeable triggers, like colored patches, which are vulnerable to human inspection. In this paper, we propose an imperceptible and effective backdoor attack against self-supervised models. We first find that existing imperceptible triggers designed for supervised learning are not as effective in compromising self-supervised models. We then identify this ineffectiveness is attributed to the overlap in distributions between the backdoor and augmented samples used in self-supervised learning. Building on this insight, we design an attack using optimized triggers that are disentangled to the augmented transformation in the self-supervised learning, while also remaining imperceptible to human vision. Experiments on five datasets and seven SSL algorithms demonstrate our attack is highly effective and stealthy. It also has strong resistance to existing backdoor defenses. Our code can be found at https://github.com/Zhang-Henry/IMPERATIVE.
- Published
- 2024
16. Harmonizing Human Insights and AI Precision: Hand in Hand for Advancing Knowledge Graph Task
- Author
-
Wang, Shurong, Zhang, Yufei, Huang, Xuliang, and Wang, Hongwei
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence - Abstract
Knowledge graph embedding (KGE) has caught significant interest for its effectiveness in knowledge graph completion (KGC), specifically link prediction (LP), with recent KGE models cracking the LP benchmarks. Despite the rapidly growing literature, insufficient attention has been paid to the cooperation between humans and AI on KG. However, humans' capability to analyze graphs conceptually may further improve the efficacy of KGE models with semantic information. To this effect, we carefully designed a human-AI team (HAIT) system dubbed KG-HAIT, which harnesses the human insights on KG by leveraging fully human-designed ad-hoc dynamic programming (DP) on KG to produce human insightful feature (HIF) vectors that capture the subgraph structural feature and semantic similarities. By integrating HIF vectors into the training of KGE models, notable improvements are observed across various benchmarks and metrics, accompanied by accelerated model convergence. Our results underscore the effectiveness of human-designed DP in the task of LP, emphasizing the pivotal role of collaboration between humans and AI on KG. We open avenues for further exploration and innovation through KG-HAIT, paving the way towards more effective and insightful KG analysis techniques.
- Published
- 2024
17. Near/Far-Field Channel Estimation For Terahertz Systems With ELAAs: A Block-Sparse-Aware Approach
- Author
-
Wang, Hongwei, Fang, Jun, Duan, Huiping, and Li, Hongbin
- Subjects
Electrical Engineering and Systems Science - Signal Processing - Abstract
Millimeter wave/Terahertz (mmWave/THz) communication with extremely large-scale antenna arrays (ELAAs) offers a promising solution to meet the escalating demand for high data rates in next-generation communications. A large array aperture, along with the ever increasing carrier frequency within the mmWave/THz bands, leads to a large Rayleigh distance. As a result, the traditional plane-wave assumption may not hold valid for mmWave/THz systems featuring ELAAs. In this paper, we consider the problem of hybrid near/far-field channel estimation by taking spherical wave propagation into account. By analyzing the coherence properties of any two near-field steering vectors, we prove that the hybrid near/far-field channel admits a block-sparse representation on a specially designed orthogonal dictionary. Specifically, the percentage of nonzero elements of such a block-sparse representation decreases in the order of $1/\sqrt{N}$, which tends to zero as the number of antennas, $N$, grows. Such a block-sparse representation allows to convert channel estimation into a block-sparse signal recovery problem. Simulation results are provided to verify our theoretical results and illustrate the performance of the proposed channel estimation approach in comparison with existing state-of-the-art methods.
- Published
- 2024
18. Do We Really Need a Complex Agent System? Distill Embodied Agent into a Single Model
- Author
-
Zhao, Zhonghan, Ma, Ke, Chai, Wenhao, Wang, Xuan, Chen, Kewei, Guo, Dongxu, Zhang, Yanting, Wang, Hongwei, and Wang, Gaoang
- Subjects
Computer Science - Artificial Intelligence ,Computer Science - Computer Vision and Pattern Recognition - Abstract
With the power of large language models (LLMs), open-ended embodied agents can flexibly understand human instructions, generate interpretable guidance strategies, and output executable actions. Nowadays, Multi-modal Language Models~(MLMs) integrate multi-modal signals into LLMs, further bringing richer perception to entity agents and allowing embodied agents to perceive world-understanding tasks more delicately. However, existing works: 1) operate independently by agents, each containing multiple LLMs, from perception to action, resulting in gaps between complex tasks and execution; 2) train MLMs on static data, struggling with dynamics in open-ended scenarios; 3) input prior knowledge directly as prompts, suppressing application flexibility. We propose STEVE-2, a hierarchical knowledge distillation framework for open-ended embodied tasks, characterized by 1) a hierarchical system for multi-granular task division, 2) a mirrored distillation method for parallel simulation data, and 3) an extra expert model for bringing additional knowledge into parallel simulation. After distillation, embodied agents can complete complex, open-ended tasks without additional expert guidance, utilizing the performance and knowledge of a versatile MLM. Extensive evaluations on navigation and creation tasks highlight the superior performance of STEVE-2 in open-ended tasks, with $1.4 \times$ - $7.3 \times$ in performance., Comment: arXiv admin note: text overlap with arXiv:2403.08282
- Published
- 2024
19. Conceptual and Unbiased Reasoning in Language Models
- Author
-
Zhou, Ben, Zhang, Hongming, Chen, Sihao, Yu, Dian, Wang, Hongwei, Peng, Baolin, Roth, Dan, and Yu, Dong
- Subjects
Computer Science - Computation and Language - Abstract
Conceptual reasoning, the ability to reason in abstract and high-level perspectives, is key to generalization in human cognition. However, limited study has been done on large language models' capability to perform conceptual reasoning. In this work, we bridge this gap and propose a novel conceptualization framework that forces models to perform conceptual reasoning on abstract questions and generate solutions in a verifiable symbolic space. Using this framework as an analytical tool, we show that existing large language models fall short on conceptual reasoning, dropping 9% to 28% on various benchmarks compared to direct inference methods. We then discuss how models can improve since high-level abstract reasoning is key to unbiased and generalizable decision-making. We propose two techniques to add trustworthy induction signals by generating familiar questions with similar underlying reasoning paths and asking models to perform self-refinement. Experiments show that our proposed techniques improve models' conceptual reasoning performance by 8% to 11%, achieving a more robust reasoning system that relies less on inductive biases., Comment: Preprint under review
- Published
- 2024
20. MedM2G: Unifying Medical Multi-Modal Generation via Cross-Guided Diffusion with Visual Invariant
- Author
-
Zhan, Chenlu, Lin, Yu, Wang, Gaoang, Wang, Hongwei, and Wu, Jian
- Subjects
Electrical Engineering and Systems Science - Image and Video Processing ,Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Machine Learning - Abstract
Medical generative models, acknowledged for their high-quality sample generation ability, have accelerated the fast growth of medical applications. However, recent works concentrate on separate medical generation models for distinct medical tasks and are restricted to inadequate medical multi-modal knowledge, constraining medical comprehensive diagnosis. In this paper, we propose MedM2G, a Medical Multi-Modal Generative framework, with the key innovation to align, extract, and generate medical multi-modal within a unified model. Extending beyond single or two medical modalities, we efficiently align medical multi-modal through the central alignment approach in the unified space. Significantly, our framework extracts valuable clinical knowledge by preserving the medical visual invariant of each imaging modal, thereby enhancing specific medical information for multi-modal generation. By conditioning the adaptive cross-guided parameters into the multi-flow diffusion framework, our model promotes flexible interactions among medical multi-modal for generation. MedM2G is the first medical generative model that unifies medical generation tasks of text-to-image, image-to-text, and unified generation of medical modalities (CT, MRI, X-ray). It performs 5 medical generation tasks across 10 datasets, consistently outperforming various state-of-the-art works., Comment: Accepted by CVPR2024
- Published
- 2024
21. API Is Enough: Conformal Prediction for Large Language Models Without Logit-Access
- Author
-
Su, Jiayuan, Luo, Jing, Wang, Hongwei, and Cheng, Lu
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning - Abstract
This study aims to address the pervasive challenge of quantifying uncertainty in large language models (LLMs) without logit-access. Conformal Prediction (CP), known for its model-agnostic and distribution-free features, is a desired approach for various LLMs and data distributions. However, existing CP methods for LLMs typically assume access to the logits, which are unavailable for some API-only LLMs. In addition, logits are known to be miscalibrated, potentially leading to degraded CP performance. To tackle these challenges, we introduce a novel CP method that (1) is tailored for API-only LLMs without logit-access; (2) minimizes the size of prediction sets; and (3) ensures a statistical guarantee of the user-defined coverage. The core idea of this approach is to formulate nonconformity measures using both coarse-grained (i.e., sample frequency) and fine-grained uncertainty notions (e.g., semantic similarity). Experimental results on both close-ended and open-ended Question Answering tasks show our approach can mostly outperform the logit-based CP baselines.
- Published
- 2024
22. Fibrillarin homologs regulate translation in divergent cell lineages during planarian homeostasis and regeneration
- Author
-
Chen, Jiajia, Li, Yucong, Wang, Yan, Wang, Hui, Yang, Jiaqi, Pan, Xue, Zhao, Yun, Xu, Hao, Jiang, Penglei, Qian, Pengxu, Wang, Hongwei, Xie, Zhi, and Lei, Kai
- Published
- 2024
- Full Text
- View/download PDF
23. Up-Conversion Luminescence and Optical Temperature-Sensing Properties of Yb3+ and Er3+ Co-doped Yttrium Aluminum Garnet Phosphor
- Author
-
Zha, Jiahao, He, Chongjun, Chen, Fangzhou, Wang, Hongwei, Dong, Biao, Liu, Lijuan, Xia, Mingjun, Deng, Chenguang, Li, Qian, Lu, Yuangang, Chen, Huiting, and Liu, Siguo
- Published
- 2024
- Full Text
- View/download PDF
24. Central role of Sigma-1 receptor in ochratoxin A-induced ferroptosis
- Author
-
Chen, Wenying, Han, Lingyun, Yang, Ruiran, Wang, Hongwei, Yao, Song, Deng, Huiqiong, Liu, Shuangchao, Zhou, Yao, and Shen, Xiao Li
- Published
- 2024
- Full Text
- View/download PDF
25. A study on the monitoring of landslide deformation disasters in Wenxian County, Longnan City based on different time-series InSAR techniques
- Author
-
Zhang, Jinlong, Yang, Rui, Qi, Yuan, Zhang, Hui, Zhang, Juan, Guo, Qianhong, Ma, Chao, and Wang, Hongwei
- Published
- 2024
- Full Text
- View/download PDF
26. Synthesis, Structure and Luminescence Properties of Mn-doped MgAl2O4 Red-Emitting Phosphors with Varying Sintering Temperature
- Author
-
Zha, Jiahao, He, Chongjun, Chen, Fangzhou, Wang, Hongwei, Dong, Biao, Liu, Lijuan, Xia, Mingjun, Deng, Chenguang, Li, Qian, Lu, Yuangang, and Chen, Huiting
- Published
- 2024
- Full Text
- View/download PDF
27. Germline loss in C. elegans enhances longevity by disrupting adhesion between niche and stem cells
- Author
-
Liu, Meng, Chen, Jiehui, Cui, Guizhong, Dai, Yumin, Song, Mengjiao, Zhou, Chunyu, Hu, Qingyuan, Chen, Qingxia, Wang, Hongwei, Chen, Wanli, Han, Jingdong Jackie, Peng, Guangdun, Jing, Naihe, and Shen, Yidong
- Published
- 2024
- Full Text
- View/download PDF
28. Stacking faults and their effects on improving plasticity in a Co–Al–W–base superalloy at 800 °C
- Author
-
Wang, Hongwei, Wang, Lei, Liu, Yang, and Song, Xiu
- Published
- 2024
- Full Text
- View/download PDF
29. Enhanced outcomes in residual or recurrent craniopharyngioma: evaluating combined gamma knife and phosphorus-32 brachytherapy
- Author
-
Ma, Jie, Chen, Tao, Zhang, Jianning, Cao, Weidong, Gao, Gan, Yu, Xinguang, and Wang, Hongwei
- Published
- 2024
- Full Text
- View/download PDF
30. The Footprints of Mitochondrial Fission and Apoptosis in Fluoride-Induced Renal Dysfunction
- Author
-
Zuo, Qiyong, Lin, Lin, Zhang, Yuling, Ommati, Mohammad Mehdi, Wang, Hongwei, and Zhao, Jing
- Published
- 2024
- Full Text
- View/download PDF
31. UniDCP: Unifying Multiple Medical Vision-language Tasks via Dynamic Cross-modal Learnable Prompts
- Author
-
Zhan, Chenlu, Zhang, Yufei, Lin, Yu, Wang, Gaoang, and Wang, Hongwei
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
Medical vision-language pre-training (Med-VLP) models have recently accelerated the fast-growing medical diagnostics application. However, most Med-VLP models learn task-specific representations independently from scratch, thereby leading to great inflexibility when they work across multiple fine-tuning tasks. In this work, we propose UniDCP, a Unified medical vision-language model with Dynamic Cross-modal learnable Prompts, which can be plastically applied to multiple medical vision-language tasks. Specifically, we explicitly construct a unified framework to harmonize diverse inputs from multiple pretraining tasks by leveraging cross-modal prompts for unification, which accordingly can accommodate heterogeneous medical fine-tuning tasks. Furthermore, we conceive a dynamic cross-modal prompt optimizing strategy that optimizes the prompts within the shareable space for implicitly processing the shareable clinic knowledge. UniDCP is the first Med-VLP model capable of performing all 8 medical uni-modal and cross-modal tasks over 14 corresponding datasets, consistently yielding superior results over diverse state-of-the-art methods.
- Published
- 2023
32. Dense X Retrieval: What Retrieval Granularity Should We Use?
- Author
-
Chen, Tong, Wang, Hongwei, Chen, Sihao, Yu, Wenhao, Ma, Kaixin, Zhao, Xinran, Zhang, Hongming, and Yu, Dong
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Information Retrieval - Abstract
Dense retrieval has become a prominent method to obtain relevant context or world knowledge in open-domain NLP tasks. When we use a learned dense retriever on a retrieval corpus at inference time, an often-overlooked design choice is the retrieval unit in which the corpus is indexed, e.g. document, passage, or sentence. We discover that the retrieval unit choice significantly impacts the performance of both retrieval and downstream tasks. Distinct from the typical approach of using passages or sentences, we introduce a novel retrieval unit, proposition, for dense retrieval. Propositions are defined as atomic expressions within text, each encapsulating a distinct factoid and presented in a concise, self-contained natural language format. We conduct an empirical comparison of different retrieval granularity. Our experiments reveal that indexing a corpus by fine-grained units such as propositions significantly outperforms passage-level units in retrieval tasks. Moreover, constructing prompts with fine-grained retrieved units for retrieval-augmented language models improves the performance of downstream QA tasks given a specific computation budget.
- Published
- 2023
33. Applications of Mass Spectrometry in Textile Analysis: An Overview
- Author
-
Ruan, Yudi, Meng, Xianshuang, Wang, Jiangang, Wang, Hongwei, Ye, Qiong, Shou, Qianyi, and Ma, Qiang
- Published
- 2024
- Full Text
- View/download PDF
34. Pre-hepatectomy dynamic circulating tumor DNA to predict pathologic response to preoperative chemotherapy and post-hepatectomy recurrence in patients with colorectal liver metastases
- Author
-
Liu, Ming, Bao, Quan, Zhao, Tingting, Huang, Longfei, Zhang, Danhua, Wang, Yanyan, Yan, Xiaoluan, Wang, Hongwei, Jin, Kemin, Liu, Wei, Wang, Kun, and Xing, Baocai
- Published
- 2024
- Full Text
- View/download PDF
35. Self-Improving for Zero-Shot Named Entity Recognition with Large Language Models
- Author
-
Xie, Tingyu, Li, Qi, Zhang, Yan, Liu, Zuozhu, and Wang, Hongwei
- Subjects
Computer Science - Computation and Language - Abstract
Exploring the application of powerful large language models (LLMs) on the named entity recognition (NER) task has drawn much attention recently. This work pushes the performance boundary of zero-shot NER with LLMs by proposing a training-free self-improving framework, which utilizes an unlabeled corpus to stimulate the self-learning ability of LLMs. First, we use the LLM to make predictions on the unlabeled corpus using self-consistency and obtain a self-annotated dataset. Second, we explore various strategies to select reliable annotations to form a reliable self-annotated dataset. Finally, for each test input, we retrieve demonstrations from the reliable self-annotated dataset and perform inference via in-context learning. Experiments on four benchmarks show substantial performance improvements achieved by our framework. Through comprehensive experimental analysis, we find that increasing the size of unlabeled corpus or iterations of self-improving does not guarantee further improvement, but the performance might be boosted via more advanced strategies for reliable annotation selection. Code and data are publicly available at https://github.com/Emma1066/Self-Improve-Zero-Shot-NER, Comment: Accepted to NAACL 2024 (Main Conference)
- Published
- 2023
36. Chain-of-Note: Enhancing Robustness in Retrieval-Augmented Language Models
- Author
-
Yu, Wenhao, Zhang, Hongming, Pan, Xiaoman, Ma, Kaixin, Wang, Hongwei, and Yu, Dong
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence - Abstract
Retrieval-augmented language models (RALMs) represent a substantial advancement in the capabilities of large language models, notably in reducing factual hallucination by leveraging external knowledge sources. However, the reliability of the retrieved information is not always guaranteed. The retrieval of irrelevant data can lead to misguided responses, and potentially causing the model to overlook its inherent knowledge, even when it possesses adequate information to address the query. Moreover, standard RALMs often struggle to assess whether they possess adequate knowledge, both intrinsic and retrieved, to provide an accurate answer. In situations where knowledge is lacking, these systems should ideally respond with "unknown" when the answer is unattainable. In response to these challenges, we introduces Chain-of-Noting (CoN), a novel approach aimed at improving the robustness of RALMs in facing noisy, irrelevant documents and in handling unknown scenarios. The core idea of CoN is to generate sequential reading notes for retrieved documents, enabling a thorough evaluation of their relevance to the given question and integrating this information to formulate the final answer. We employed ChatGPT to create training data for CoN, which was subsequently trained on an LLaMa-2 7B model. Our experiments across four open-domain QA benchmarks show that RALMs equipped with CoN significantly outperform standard RALMs. Notably, CoN achieves an average improvement of +7.9 in EM score given entirely noisy retrieved documents and +10.5 in rejection rates for real-time questions that fall outside the pre-training knowledge scope., Comment: EMNLP 2024 (main conference)
- Published
- 2023
37. Revisit to the yield ratio of triton and $^3$He as an indicator of neutron-rich neck emission
- Author
-
Wang, Yijie, Wan, Mengting, Diao, Xinyue, Xiao, Sheng, Qin, Yuhao, Qin, Zhi, Guo, Dong, Si, Dawei, Zhang, Boyuan, Tian, Baiting, Guan, Fenhai, Wu, Qianghua, Wei, Xianglun, Yang, Herun, Ma, Peng, Hu, Rongjiang, Duan, Limin, Duan, Fangfang, Ma, Junbing, Xu, Shiwei, Hu, Qiang, Bai, Zhen, Yang, Yanyun, Wang, Jiansong, Liu, Wenbo, Su, Wanqing, Wei, Xiaobao, Ma, Chunwang, Li, Xinxiang, Wang, Hongwei, Zhang, Yingxun, Warda, Michał, Dobrowolski, Arthur, Nerlo-Pomorska, Bożena, Pomorski, Krzysztof, Ou, Li, and Xiao, Zhigang
- Subjects
Nuclear Experiment ,Nuclear Theory - Abstract
The neutron rich neck zone created in heavy ion reaction is experimentally probed by the production of the $A=3$ isobars. The energy spectra and angular distributions of triton and $^3$He are measured with the CSHINE detector in $^{86}$Kr +$^{208}$Pb reactions at 25 MeV/u. While the energy spectrum of $^{3}$He is harder than that of triton, known as "$^{3}$He-puzzle", the yield ratio $R({\rm t/^3He})$ presents a robust rising trend with the polar angle in laboratory. Using the fission fragments to reconstruct the fission plane, the enhancement of out-plane $R({\rm t/^3He})$ is confirmed in comparison to the in-plane ratios. Transport model simulations reproduce qualitatively the experimental trends, but the quantitative agreement is not achieved. The results demonstrate that a neutron rich neck zone is formed in the reactions. Further studies are called for to understand the clustering and the isospin dynamics related to neck formation.
- Published
- 2023
38. Sub-Sentence Encoder: Contrastive Learning of Propositional Semantic Representations
- Author
-
Chen, Sihao, Zhang, Hongming, Chen, Tong, Zhou, Ben, Yu, Wenhao, Yu, Dian, Peng, Baolin, Wang, Hongwei, Roth, Dan, and Yu, Dong
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence - Abstract
We introduce sub-sentence encoder, a contrastively-learned contextual embedding model for fine-grained semantic representation of text. In contrast to the standard practice with sentence embeddings, where the meaning of an entire sequence of text is encoded into a fixed-length vector, the sub-sentence encoder learns to produce distinct contextual embeddings corresponding to different atomic propositions, i.e. atomic units of meaning expressed within a text sequence. The sub-sentence embeddings are contrastively learned to recognize (inferred) semantic equivalence between propositions across different text sequences. Our experiments show the effectiveness of sub-sentence encoders in applications, such as retrieving supporting facts for fine-grained text attribution or recognizing the conditional semantic similarity between texts. In practice, we demonstrate that sub-sentence encoders keep the same level of inference cost and space complexity compared to sentence encoders.
- Published
- 2023
39. On-chip topological transport of optical frequency combs in silicon-based valley photonic crystals
- Author
-
Jiang, Zhen, Wang, Hongwei, Yang, Yuechen, Shen, Yang, Ji, Bo, Chen, Yanghe, Zhang, Yong, Sun, Lu, Wang, Zheng, Jiang, Chun, Su, Yikai, and He, Guangqiang
- Subjects
Physics - Optics ,Quantum Physics - Abstract
The generation and control of optical frequency combs in integrated photonic systems enables complex, high-controllable, and large-scale devices. In parallel, harnessing topological physics in multipartite systems has allowed them with compelling features such as robustness against fabrication imperfections. Here we experimentally demonstrate on-chip topological transport for optical frequency combs at telecommunication wavelengths, both in classical and nonclassical domains. We access both the quantum frequency combs and dissipative Kerr soliton combs with a micro-resonator. The quantum frequency comb, that is, a coherent superposition of multiple frequency modes, is proven to be a frequency-entangled qudit state. We also show that dissipative Kerr soliton combs are highly coherent and mode-locked due to the collective coherence or self-organization of solitons. Moreover, the valley kink states allow both quantum frequency combs and dissipative Kerr soliton combs with robustness against sharp bends. Our topologically protected optical frequency combs could enable the inherent robustness in integrated complex photonic systems., Comment: 20 pages,12 figures
- Published
- 2023
40. On the Dimensionality of Sentence Embeddings
- Author
-
Wang, Hongwei, Zhang, Hongming, and Yu, Dong
- Subjects
Computer Science - Computation and Language - Abstract
Learning sentence embeddings is a fundamental problem in natural language processing. While existing research primarily focuses on enhancing the quality of sentence embeddings, the exploration of sentence embedding dimensions is limited. Here we present a comprehensive and empirical analysis of the dimensionality of sentence embeddings. First, we demonstrate that the optimal dimension of sentence embeddings is usually smaller than the default value. Subsequently, to compress the dimension of sentence embeddings with minimum performance degradation, we identify two components contributing to the overall performance loss: the encoder's performance loss and the pooler's performance loss. Therefore, we propose a two-step training method for sentence representation learning models, wherein the encoder and the pooler are optimized separately to mitigate the overall performance loss in low-dimension scenarios. Experimental results on seven STS tasks and seven sentence classification tasks demonstrate that our method significantly improves the performance of low-dimensional sentence embeddings.
- Published
- 2023
41. Weighted Joint Maximum Mean Discrepancy Enabled Multi-Source-Multi-Target Unsupervised Domain Adaptation Fault Diagnosis
- Author
-
Wang, Zixuan, Tang, Haoran, Wang, Haibo, Qin, Bo, Butala, Mark D., Shen, Weiming, and Wang, Hongwei
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence - Abstract
Despite the remarkable results that can be achieved by data-driven intelligent fault diagnosis techniques, they presuppose the same distribution of training and test data as well as sufficient labeled data. Various operating states often exist in practical scenarios, leading to the problem of domain shift that hinders the effectiveness of fault diagnosis. While recent unsupervised domain adaptation methods enable cross-domain fault diagnosis, they struggle to effectively utilize information from multiple source domains and achieve effective diagnosis faults in multiple target domains simultaneously. In this paper, we innovatively proposed a weighted joint maximum mean discrepancy enabled multi-source-multi-target unsupervised domain adaptation (WJMMD-MDA), which realizes domain adaptation under multi-source-multi-target scenarios in the field of fault diagnosis for the first time. The proposed method extracts sufficient information from multiple labeled source domains and achieves domain alignment between source and target domains through an improved weighted distance loss. As a result, domain-invariant and discriminative features between multiple source and target domains are learned with cross-domain fault diagnosis realized. The performance of the proposed method is evaluated in comprehensive comparative experiments on three datasets, and the experimental results demonstrate the superiority of this method.
- Published
- 2023
42. Empirical Study of Zero-Shot NER with ChatGPT
- Author
-
Xie, Tingyu, Li, Qi, Zhang, Jian, Zhang, Yan, Liu, Zuozhu, and Wang, Hongwei
- Subjects
Computer Science - Computation and Language - Abstract
Large language models (LLMs) exhibited powerful capability in various natural language processing tasks. This work focuses on exploring LLM performance on zero-shot information extraction, with a focus on the ChatGPT and named entity recognition (NER) task. Inspired by the remarkable reasoning capability of LLM on symbolic and arithmetic reasoning, we adapt the prevalent reasoning methods to NER and propose reasoning strategies tailored for NER. First, we explore a decomposed question-answering paradigm by breaking down the NER task into simpler subproblems by labels. Second, we propose syntactic augmentation to stimulate the model's intermediate thinking in two ways: syntactic prompting, which encourages the model to analyze the syntactic structure itself, and tool augmentation, which provides the model with the syntactic information generated by a parsing tool. Besides, we adapt self-consistency to NER by proposing a two-stage majority voting strategy, which first votes for the most consistent mentions, then the most consistent types. The proposed methods achieve remarkable improvements for zero-shot NER across seven benchmarks, including Chinese and English datasets, and on both domain-specific and general-domain scenarios. In addition, we present a comprehensive analysis of the error types with suggestions for optimization directions. We also verify the effectiveness of the proposed methods on the few-shot setting and other LLMs., Comment: Accepted to EMNLP 2023 (Main Conference)
- Published
- 2023
43. SemStamp: A Semantic Watermark with Paraphrastic Robustness for Text Generation
- Author
-
Hou, Abe Bohan, Zhang, Jingyu, He, Tianxing, Wang, Yichen, Chuang, Yung-Sung, Wang, Hongwei, Shen, Lingfeng, Van Durme, Benjamin, Khashabi, Daniel, and Tsvetkov, Yulia
- Subjects
Computer Science - Computation and Language - Abstract
Existing watermarking algorithms are vulnerable to paraphrase attacks because of their token-level design. To address this issue, we propose SemStamp, a robust sentence-level semantic watermarking algorithm based on locality-sensitive hashing (LSH), which partitions the semantic space of sentences. The algorithm encodes and LSH-hashes a candidate sentence generated by an LLM, and conducts sentence-level rejection sampling until the sampled sentence falls in watermarked partitions in the semantic embedding space. A margin-based constraint is used to enhance its robustness. To show the advantages of our algorithm, we propose a "bigram" paraphrase attack using the paraphrase that has the fewest bigram overlaps with the original sentence. This attack is shown to be effective against the existing token-level watermarking method. Experimental results show that our novel semantic watermark algorithm is not only more robust than the previous state-of-the-art method on both common and bigram paraphrase attacks, but also is better at preserving the quality of generation., Comment: Accepted to NAACL 24 Main
- Published
- 2023
44. LASER: LLM Agent with State-Space Exploration for Web Navigation
- Author
-
Ma, Kaixin, Zhang, Hongming, Wang, Hongwei, Pan, Xiaoman, Yu, Wenhao, and Yu, Dong
- Subjects
Computer Science - Computation and Language - Abstract
Large language models (LLMs) have been successfully adapted for interactive decision-making tasks like web navigation. While achieving decent performance, previous methods implicitly assume a forward-only execution mode for the model, where they only provide oracle trajectories as in-context examples to guide the model on how to reason in the environment. Consequently, the model could not handle more challenging scenarios not covered in the in-context examples, e.g., mistakes, leading to sub-optimal performance. To address this issue, we propose to model the interactive task as state space exploration, where the LLM agent transitions among a pre-defined set of states by performing actions to complete the task. This formulation enables flexible backtracking, allowing the model to recover from errors easily. We evaluate our proposed LLM Agent with State-Space ExploRation (LASER) on both the WebShop task and amazon.com. Experimental results show that LASER significantly outperforms previous methods and closes the gap with human performance on the web navigation task., Comment: 4 pages, 2 figures
- Published
- 2023
45. Unsupervised Multi-document Summarization with Holistic Inference
- Author
-
Zhang, Haopeng, Cho, Sangwoo, Song, Kaiqiang, Wang, Xiaoyang, Wang, Hongwei, Zhang, Jiawei, and Yu, Dong
- Subjects
Computer Science - Computation and Language - Abstract
Multi-document summarization aims to obtain core information from a collection of documents written on the same topic. This paper proposes a new holistic framework for unsupervised multi-document extractive summarization. Our method incorporates the holistic beam search inference method associated with the holistic measurements, named Subset Representative Index (SRI). SRI balances the importance and diversity of a subset of sentences from the source documents and can be calculated in unsupervised and adaptive manners. To demonstrate the effectiveness of our method, we conduct extensive experiments on both small and large-scale multi-document summarization datasets under both unsupervised and adaptive settings. The proposed method outperforms strong baselines by a significant margin, as indicated by the resulting ROUGE scores and diversity measures. Our findings also suggest that diversity is essential for improving multi-document summary performance., Comment: Findings of IJCNLP-AACL 2023
- Published
- 2023
46. Thermally tunable add-drop filter based on valley photonic crystals for optical communications
- Author
-
Sun Lu, Li Xingfeng, Hu Pan, Wang Hongwei, Zhang Yong, Tang Guojing, He Xintao, Dong Jianwen, and Su Yikai
- Subjects
topological photonics ,valley photonic crystals ,optical communications ,Physics ,QC1-999 - Abstract
Valley photonic crystals (VPCs) provide an intriguing approach to suppress backscattering losses and enable robust transport of light against sharp bends, which could be utilized to realize low-loss and small-footprint devices for on-chip optical communications. However, there are few studies on how to achieve power-efficient tunable devices based on VPCs, which are essential for implementing basic functions such as optical switching and routing. Here, we propose and experimentally demonstrate a thermally tunable add-drop filter (ADF) based on VPCs operating at telecommunication wavelengths. By leveraging the topological protection of the edge state and the distinct property of negligible scattering at sharp bends, a small footprint of 17.4 × 28.2 μm2 and a low insertion loss of 2.7 dB can be achieved for the proposed device. A diamond-shaped microloop resonator is designed to confine the light and enhance its interaction with the thermal field generated by the microheater, leading to a relatively low power of 23.97 mW needed for switching the output signal from one port to the other. Based on the thermally tunable ADF under the protection of band topology, robust data transmission is implemented with an ultrahigh data rate of 132 Gb/s. Our work shows great potential for developing high-performance topological photonic devices with the thermally tunable silicon-based VPCs, which offers unprecedented opportunities for realizing topologically protected and reconfigurable high-speed datalinks on a chip.
- Published
- 2024
- Full Text
- View/download PDF
47. Fragment and Integrate Network (FIN): A Novel Spatial-Temporal Modeling Based on Long Sequential Behavior for Online Food Ordering Click-Through Rate Prediction
- Author
-
Li, Jun, Wang, Jingjian, Wang, Hongwei, Deng, Xing, Chen, Jielong, Cao, Bing, Wang, Zekun, Xu, Guanjie, Zhang, Ge, Shi, Feng, and Liu, Hualei
- Subjects
Computer Science - Information Retrieval ,Computer Science - Machine Learning - Abstract
Spatial-temporal information has been proven to be of great significance for click-through rate prediction tasks in online Location-Based Services (LBS), especially in mainstream food ordering platforms such as DoorDash, Uber Eats, Meituan, and Ele.me. Modeling user spatial-temporal preferences with sequential behavior data has become a hot topic in recommendation systems and online advertising. However, most of existing methods either lack the representation of rich spatial-temporal information or only handle user behaviors with limited length, e.g. 100. In this paper, we tackle these problems by designing a new spatial-temporal modeling paradigm named Fragment and Integrate Network (FIN). FIN consists of two networks: (i) Fragment Network (FN) extracts Multiple Sub-Sequences (MSS) from lifelong sequential behavior data, and captures the specific spatial-temporal representation by modeling each MSS respectively. Here both a simplified attention and a complicated attention are adopted to balance the performance gain and resource consumption. (ii) Integrate Network (IN) builds a new integrated sequence by utilizing spatial-temporal interaction on MSS and captures the comprehensive spatial-temporal representation by modeling the integrated sequence with a complicated attention. Both public datasets and production datasets have demonstrated the accuracy and scalability of FIN. Since 2022, FIN has been fully deployed in the recommendation advertising system of Ele.me, one of the most popular online food ordering platforms in China, obtaining 5.7% improvement on Click-Through Rate (CTR) and 7.3% increase on Revenue Per Mille (RPM)., Comment: Accepted by CIKM 2023 Applied Research Paper
- Published
- 2023
48. Probing high-momentum component in nucleon momentum distribution by neutron-proton bremsstrahlung {\gamma}-rays in heavy ion reactions
- Author
-
Qin, Yuhao, Niu, Qinglin, Guo, Dong, Xiao, Sheng, Tian, Baiting, Wang, Yijie, Qin, Zhi, Diao, Xinyue, Guan, Fenhai, Si, Dawei, Zhang, Boyuan, Zhang, Yaopeng, Wei, Xianglun, Yang, Herun, Ma, Peng, Hu, Rongjiang, Duan, Limin, Duan, Fangfang, Hu, Qiang, Ma, Junbing, Xu, Shiwei, Bai, Zhen, Yang, Yanyun, Wang, Hongwei, Sun, Baohua, Maydanyuk, Sergei P., Xu, Chang, and Xiao, Zhigang
- Subjects
Nuclear Experiment - Abstract
The high momentum tail (HMT) of nucleons, as a signature of the short-range correlations in nuclei, has been investigated by the high-energy bremsstrahlung $\gamma$ rays produced in $^{86}$Kr + $^{124}$Sn at 25 MeV/u. The energetic photons are measured by a CsI(Tl) hodoscope mounted on the spectrometer CSHINE. The energy spectrum above 30 MeV can be reproduced by the IBUU model calculations incorporating the photon production channel from $np$ process in which the HMTs of nucleons is considered. A non-zero HMT ratio of about $15\%$ is favored by the data. The effect of the capture channel $np \to d\gamma$ is demonstrated.
- Published
- 2023
49. Estimands in Real-World Evidence Studies
- Author
-
Chen, Jie, Scharfstein, Daniel, Wang, Hongwei, Yu, Binbing, Song, Yang, He, Weili, Scott, John, Lin, Xiwu, and Lee, Hana
- Subjects
Statistics - Applications - Abstract
A Real-World Evidence (RWE) Scientific Working Group (SWG) of the American Statistical Association Biopharmaceutical Section (ASA BIOP) has been reviewing statistical considerations for the generation of RWE to support regulatory decision-making. As part of the effort, the working group is addressing estimands in RWE studies. Constructing the right estimand -- the target of estimation -- which reflects the research question and the study objective, is one of the key components in formulating a clinical study. ICH E9(R1) describes statistical principles for constructing estimands in clinical trials with a focus on five attributes -- population, treatment, endpoints, intercurrent events, and population-level summary. However, defining estimands for clinical studies using real-world data (RWD), i.e., RWE studies, requires additional considerations due to, for example, heterogeneity of study population, complexity of treatment regimes, different types and patterns of intercurrent events, and complexities in choosing study endpoints. This paper reviews the essential components of estimands and causal inference framework, discusses considerations in constructing estimands for RWE studies, highlights similarities and differences in traditional clinical trial and RWE study estimands, and provides a roadmap for choosing appropriate estimands for RWE studies.
- Published
- 2023
50. Generalized Out-of-distribution Fault Diagnosis (GOOFD) via Internal Contrastive Learning
- Author
-
Wang, Xingyue, Zhang, Hanrong, Qiao, Xinlong, Ma, Ke, Tao, Shuting, Peng, Peng, and Wang, Hongwei
- Subjects
Computer Science - Artificial Intelligence - Abstract
Fault diagnosis is crucial in monitoring machines within industrial processes. With the increasing complexity of working conditions and demand for safety during production, diverse diagnosis methods are required, and an integrated fault diagnosis system capable of handling multiple tasks is highly desired. However, the diagnosis subtasks are often studied separately, and the current methods still need improvement for such a generalized system. To address this issue, we propose the Generalized Out-of-distribution Fault Diagnosis (GOOFD) framework to integrate diagnosis subtasks. Additionally, a unified fault diagnosis method based on internal contrastive learning and Mahalanobis distance is put forward to underpin the proposed generalized framework. The method involves feature extraction through internal contrastive learning and outlier recognition based on the Mahalanobis distance. Our proposed method can be applied to multiple faults diagnosis tasks and achieve better performance than the existing single-task methods. Experiments are conducted on benchmark and practical process datasets, indicating the effectiveness of the proposed framework.
- Published
- 2023
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.