12,164 results on '"Xu, Chang"'
Search Results
2. Generative Physical AI in Vision: A Survey
- Author
-
Liu, Daochang, Zhang, Junyu, Dinh, Anh-Dung, Park, Eunbyung, Zhang, Shichao, and Xu, Chang
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
Generative Artificial Intelligence (AI) has rapidly advanced the field of computer vision by enabling machines to create and interpret visual data with unprecedented sophistication. This transformation builds upon a foundation of generative models to produce realistic images, videos, and 3D or 4D content. Traditionally, generative models primarily focus on visual fidelity while often neglecting the physical plausibility of generated content. This gap limits their effectiveness in applications requiring adherence to real-world physical laws, such as robotics, autonomous systems, and scientific simulations. As generative AI evolves to increasingly integrate physical realism and dynamic simulation, its potential to function as a "world simulator" expands-enabling the modeling of interactions governed by physics and bridging the divide between virtual and physical realities. This survey systematically reviews this emerging field of physics-aware generative AI in computer vision, categorizing methods based on how they incorporate physical knowledge-either through explicit simulation or implicit learning. We analyze key paradigms, discuss evaluation protocols, and identify future research directions. By offering a comprehensive overview, this survey aims to help future developments in physically grounded generation for vision. The reviewed papers are summarized at https://github.com/BestJunYu/Awesome-Physics-aware-Generation.
- Published
- 2025
3. High-speed readout for direct light orbital angular momentum photodetector via photoelastic modulation
- Author
-
Yang, Dehong, Xu, Chang, Lai, Jiawei, Fan, Zipu, Liang, Delang, Wang, Shiyu, Cheng, Jinluo, and Sun, Dong
- Subjects
Physics - Optics ,Physics - Applied Physics - Abstract
Recent progress in direct photodetection of light orbital angular momentum (OAM) based on the orbital photogalvanic effect (OPGE) provides an effective way for on-chip direct electric readout of orbital angular momentum, as well as large-scale integration focal-plane array devices. However, the recognition of OAM order from photocurrent response requires the extraction of circular polarization-dependent response. To date, the operation speed of such detector is currently at the minute level and is limited by slow mechanical polarization modulation and low OAM recognition capability. In this work, we demonstrate that the operation speed can be greatly improved via electrical polarization modulation strategy with photoelasitc modulator accompanied by phase-locked readout approach with lock-in amplifier. We demonstrate an operation speed of up to 1 kHz with this new technology in the mid-infrared region (4 {\mu}m) on an OAM detector using multilayer graphene (MLG) as photosensitive material. In principle, with new modulation and readout scheme, we can potentially increase the operation speed to 50.14 kHz with a PEM that operates at a state-of-the-art speed. Our work paves the way toward high-speed operation of direct OAM detection devices based on OPGE effect and pushes such technology to a more practical stage for focal plane array applications., Comment: 27 pages,5 figures
- Published
- 2025
4. TimeDP: Learning to Generate Multi-Domain Time Series with Domain Prompts
- Author
-
Huang, Yu-Hao, Xu, Chang, Wu, Yueying, Li, Wu-Jun, and Bian, Jiang
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence - Abstract
Time series generation models are crucial for applications like data augmentation and privacy preservation. Most existing time series generation models are typically designed to generate data from one specified domain. While leveraging data from other domain for better generalization is proved to work in other application areas, this approach remains challenging for time series modeling due to the large divergence in patterns among different real world time series categories. In this paper, we propose a multi-domain time series diffusion model with domain prompts, named TimeDP. In TimeDP, we utilize a time series semantic prototype module which defines time series prototypes to represent time series basis, each prototype vector serving as "word" representing some elementary time series feature. A prototype assignment module is applied to extract the extract domain specific prototype weights, for learning domain prompts as generation condition. During sampling, we extract "domain prompt" with few-shot samples from the target domain and use the domain prompts as condition to generate time series samples. Experiments demonstrate that our method outperforms baselines to provide the state-of-the-art in-domain generation quality and strong unseen domain generation capability., Comment: AAAI 2025
- Published
- 2025
5. Color Correction Meets Cross-Spectral Refinement: A Distribution-Aware Diffusion for Underwater Image Restoration
- Author
-
Chang, Laibin, Wang, Yunke, Du, Bo, and Xu, Chang
- Subjects
Electrical Engineering and Systems Science - Image and Video Processing - Abstract
Underwater imaging often suffers from significant visual degradation, which limits its suitability for subsequent applications. While recent underwater image enhancement (UIE) methods rely on the current advances in deep neural network architecture designs, there is still considerable room for improvement in terms of cross-scene robustness and computational efficiency. Diffusion models have shown great success in image generation, prompting us to consider their application to UIE tasks. However, directly applying them to UIE tasks will pose two challenges, \textit{i.e.}, high computational budget and color unbalanced perturbations. To tackle these issues, we propose DiffColor, a distribution-aware diffusion and cross-spectral refinement model for efficient UIE. Instead of diffusing in the raw pixel space, we transfer the image into the wavelet domain to obtain such low-frequency and high-frequency spectra, it inherently reduces the image spatial dimensions by half after each transformation. Unlike single-noise image restoration tasks, underwater imaging exhibits unbalanced channel distributions due to the selective absorption of light by water. To address this, we design the Global Color Correction (GCC) module to handle the diverse color shifts, thereby avoiding potential global degradation disturbances during the denoising process. For the sacrificed image details caused by underwater scattering, we further present the Cross-Spectral Detail Refinement (CSDR) to enhance the high-frequency details, which are integrated with the low-frequency signal as input conditions for guiding the diffusion. This way not only ensures the high-fidelity of sampled content but also compensates for the sacrificed details. Comprehensive experiments demonstrate the superior performance of DiffColor over state-of-the-art methods in both quantitative and qualitative evaluations.
- Published
- 2025
6. TimeRAF: Retrieval-Augmented Foundation model for Zero-shot Time Series Forecasting
- Author
-
Zhang, Huanyu, Xu, Chang, Zhang, Yi-Fan, Zhang, Zhang, Wang, Liang, Bian, Jiang, and Tan, Tieniu
- Subjects
Computer Science - Machine Learning - Abstract
Time series forecasting plays a crucial role in data mining, driving rapid advancements across numerous industries. With the emergence of large models, time series foundation models (TSFMs) have exhibited remarkable generalization capabilities, such as zero-shot learning, through large-scale pre-training. Meanwhile, Retrieval-Augmented Generation (RAG) methods have been widely employed to enhance the performance of foundation models on unseen data, allowing models to access to external knowledge. In this paper, we introduce TimeRAF, a Retrieval-Augmented Forecasting model that enhance zero-shot time series forecasting through retrieval-augmented techniques. We develop customized time series knowledge bases that are tailored to the specific forecasting tasks. TimeRAF employs an end-to-end learnable retriever to extract valuable information from the knowledge base. Additionally, we propose Channel Prompting for knowledge integration, which effectively extracts relevant information from the retrieved knowledge along the channel dimension. Extensive experiments demonstrate the effectiveness of our model, showing significant improvement across various domains and datasets.
- Published
- 2024
7. Flavor Physics at CEPC: a General Perspective
- Author
-
Ai, Xiaocong, Altmannshofer, Wolfgang, Athron, Peter, Bai, Xiaozhi, Calibbi, Lorenzo, Cao, Lu, Che, Yuzhi, Chen, Chunhui, Chen, Ji-Yuan, Chen, Long, Chen, Mingshui, Chen, Shanzhen, Chen, Xuan, Cheng, Shan, Chiang, Cheng-Wei, Crivellin, Andreas, Cui, Hanhua, Deschamps, Olivier, Descotes-Genon, Sébastien, Du, Xiaokang, Fang, Shuangshi, Gao, Yu, Geng, Li-Sheng, Goldenzweig, Pablo, Gu, Jiayin, Guo, Feng-Kun, Guo, Yuchen, Guo, Zhi-Hui, Han, Tao, He, Hong-Jian, He, Jibo, He, Miao, Huang, Yanping, Isidori, Gino, Ji, Quan, Jiang, Jianfeng, Jiang, Xu-Hui, Kamenik, Jernej F., Kwok, Tsz Hong, Li, Gang, Li, Geng, Li, Haibo, Li, Haitao, Li, Hengne, Li, Honglei, Li, Liang, Li, Lingfeng, Li, Qiang, Li, Shu, Li, Xiaomei, Li, Xin-Qiang, Li, Yiming, Li, Yubo, Li, Yuji, Li, Zhao, Liang, Hao, Liang, Zhijun, Liao, Libo, Ligeti, Zoltan, Liu, Jia, Liu, Jianbei, Liu, Tao, Liu, Yi, Liu, Yong, Liu, Zhen, Lou, Xinchou, Lu, Peng-Cheng, Lusiani, Alberto, Ma, Hong-Hao, Ma, Kai, Mao, Yaxian, Marzocca, David, Niu, Juan-Juan, Prell, Soeren, Qi, Huirong, Qian, Sen, Qian, Wenbin, Qian, Zhuoni, Qin, Qin, Rock, Ariel, Rosner, Jonathan L., Ruan, Manqi, Shao, Dingyu, Shen, Chengping, Shen, Xiaoyan, Shi, Haoyu, Shi, Liaoshan, Si, Zong-Guo, Sierra, Cristian, Song, Huayang, Su, Shufang, Su, Wei, Tammaro, Michele, Wang, En, Wang, Fei, Wang, Hengyu, Wang, Jian, Wang, Jianchun, Wang, Kun, Wang, Lian-Tao, Wang, Wei, Wang, Xiaolong, Wang, Xiaoping, Wang, Yadi, Wang, Yifang, Wang, Yuexin, Wu, Xing-Gang, Wu, Yongcheng, Xiao, Rui-Qing, Xie, Ke-Pan, Xie, Yuehong, Xu, Zijun, Yang, Haijun, Yang, Hongtao, Yang, Lin, Yang, Shuo, Yin, Zhongbao, Yu, Fusheng, Yuan, Changzheng, Yuan, Xing-Bo, Yuan, Xuhao, Yue, Chongxing, Zhan, Xi-Jie, Zhang, Kaili, Zhang, Liming, Zhang, Xiaoming, Zhang, Yang, Zhang, Yanxi, Zhang, Yongchao, Zhang, Yu, Zhang, Zhen-Hua, Zhang, Zhong, Zhao, Mingrui, Zhao, Qiang, Zheng, Xu-Chang, Zheng, Yangheng, Zhou, Chen, Zhu, Pengxuan, Zhu, Yongfeng, Zuo, Xunwu, and Zupan, Jure
- Subjects
High Energy Physics - Experiment ,High Energy Physics - Phenomenology - Abstract
We discuss the landscape of flavor physics at the Circular Electron-Positron Collider (CEPC), based on the nominal luminosity outlined in its Technical Design Report. The CEPC is designed to operate in multiple modes to address a variety of tasks. At the $Z$ pole, the expected production of 4 Tera $Z$ bosons will provide unique and highly precise measurements of $Z$ boson couplings, while the substantial number of boosted heavy-flavored quarks and leptons produced in clean $Z$ decays will facilitate investigations into their flavor physics with unprecedented precision. We investigate the prospects of measuring various physics benchmarks and discuss their implications for particle theories and phenomenological models. Our studies indicate that, with its highlighted advantages and anticipated excellent detector performance, the CEPC can explore beauty and $\tau$ physics in ways that are superior to or complementary with the Belle II and Large-Hadron-Collider-beauty experiments, potentially enabling the detection of new physics at energy scales of 10 TeV and above. This potential also extends to the observation of yet-to-be-discovered rare and exotic processes, as well as testing fundamental principles such as lepton flavor universality, lepton and baryon number conservation, etc., making the CEPC a vibrant platform for flavor physics research. The $WW$ threshold scan, Higgs-factory operation and top-pair productions of the CEPC further enhance its merits in this regard, especially for measuring the Cabibbo-Kobayashi-Maskawa matrix elements, and Flavor-Changing-Neutral-Current physics of Higgs boson and top quarks. We outline the requirements for detector performance and considerations for future development to achieve the anticipated scientific goals.
- Published
- 2024
8. Modelling Multi-modal Cross-interaction for ML-FSIC Based on Local Feature Selection
- Author
-
Yan, Kun, Bouraoui, Zied, Wei, Fangyun, Xu, Chang, Wang, Ping, Jameel, Shoaib, and Schockaert, Steven
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
The aim of multi-label few-shot image classification (ML-FSIC) is to assign semantic labels to images, in settings where only a small number of training examples are available for each label. A key feature of the multi-label setting is that images often have several labels, which typically refer to objects appearing in different regions of the image. When estimating label prototypes, in a metric-based setting, it is thus important to determine which regions are relevant for which labels, but the limited amount of training data and the noisy nature of local features make this highly challenging. As a solution, we propose a strategy in which label prototypes are gradually refined. First, we initialize the prototypes using word embeddings, which allows us to leverage prior knowledge about the meaning of the labels. Second, taking advantage of these initial prototypes, we then use a Loss Change Measurement~(LCM) strategy to select the local features from the training images (i.e.\ the support set) that are most likely to be representative of a given label. Third, we construct the final prototype of the label by aggregating these representative local features using a multi-modal cross-interaction mechanism, which again relies on the initial word embedding-based prototypes. Experiments on COCO, PASCAL VOC, NUS-WIDE, and iMaterialist show that our model substantially improves the current state-of-the-art., Comment: Accepted in Transactions on Multimedia Computing Communications and Applications
- Published
- 2024
9. Oriented Tiny Object Detection: A Dataset, Benchmark, and Dynamic Unbiased Learning
- Author
-
Xu, Chang, Zhang, Ruixiang, Yang, Wen, Zhu, Haoran, Xu, Fang, Ding, Jian, and Xia, Gui-Song
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Detecting oriented tiny objects, which are limited in appearance information yet prevalent in real-world applications, remains an intricate and under-explored problem. To address this, we systemically introduce a new dataset, benchmark, and a dynamic coarse-to-fine learning scheme in this study. Our proposed dataset, AI-TOD-R, features the smallest object sizes among all oriented object detection datasets. Based on AI-TOD-R, we present a benchmark spanning a broad range of detection paradigms, including both fully-supervised and label-efficient approaches. Through investigation, we identify a learning bias presents across various learning pipelines: confident objects become increasingly confident, while vulnerable oriented tiny objects are further marginalized, hindering their detection performance. To mitigate this issue, we propose a Dynamic Coarse-to-Fine Learning (DCFL) scheme to achieve unbiased learning. DCFL dynamically updates prior positions to better align with the limited areas of oriented tiny objects, and it assigns samples in a way that balances both quantity and quality across different object shapes, thus mitigating biases in prior settings and sample selection. Extensive experiments across eight challenging object detection datasets demonstrate that DCFL achieves state-of-the-art accuracy, high efficiency, and remarkable versatility. The dataset, benchmark, and code are available at https://chasel-tsui.github.io/AI-TOD-R/.
- Published
- 2024
10. InvDiff: Invariant Guidance for Bias Mitigation in Diffusion Models
- Author
-
Hou, Min, Wu, Yueying, Xu, Chang, Huang, Yu-Hao, Bai, Chenxi, Wu, Le, and Bian, Jiang
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Information Retrieval ,Computer Science - Machine Learning - Abstract
As one of the most successful generative models, diffusion models have demonstrated remarkable efficacy in synthesizing high-quality images. These models learn the underlying high-dimensional data distribution in an unsupervised manner. Despite their success, diffusion models are highly data-driven and prone to inheriting the imbalances and biases present in real-world data. Some studies have attempted to address these issues by designing text prompts for known biases or using bias labels to construct unbiased data. While these methods have shown improved results, real-world scenarios often contain various unknown biases, and obtaining bias labels is particularly challenging. In this paper, we emphasize the necessity of mitigating bias in pre-trained diffusion models without relying on auxiliary bias annotations. To tackle this problem, we propose a framework, InvDiff, which aims to learn invariant semantic information for diffusion guidance. Specifically, we propose identifying underlying biases in the training data and designing a novel debiasing training objective. Then, we employ a lightweight trainable module that automatically preserves invariant semantic information and uses it to guide the diffusion model's sampling process toward unbiased outcomes simultaneously. Notably, we only need to learn a small number of parameters in the lightweight learnable module without altering the pre-trained diffusion model. Furthermore, we provide a theoretical guarantee that the implementation of InvDiff is equivalent to reducing the error upper bound of generalization. Extensive experimental results on three publicly available benchmarks demonstrate that InvDiff effectively reduces biases while maintaining the quality of image generation. Our code is available at https://github.com/Hundredl/InvDiff., Comment: KDD 2025
- Published
- 2024
11. Tiny Object Detection with Single Point Supervision
- Author
-
Zhu, Haoran, Xu, Chang, Zhang, Ruixiang, Xu, Fang, Yang, Wen, Zhang, Haijian, and Xia, Gui-Song
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Tiny objects, with their limited spatial resolution, often resemble point-like distributions. As a result, bounding box prediction using point-level supervision emerges as a natural and cost-effective alternative to traditional box-level supervision. However, the small scale and lack of distinctive features of tiny objects make point annotations prone to noise, posing significant hurdles for model robustness. To tackle these challenges, we propose Point Teacher--the first end-to-end point-supervised method for robust tiny object detection in aerial images. To handle label noise from scale ambiguity and location shifts in point annotations, Point Teacher employs the teacher-student architecture and decouples the learning into a two-phase denoising process. In this framework, the teacher network progressively denoises the pseudo boxes derived from noisy point annotations, guiding the student network's learning. Specifically, in the first phase, random masking of image regions facilitates regression learning, enabling the teacher to transform noisy point annotations into coarse pseudo boxes. In the second phase, these coarse pseudo boxes are refined using dynamic multiple instance learning, which adaptively selects the most reliable instance from dynamically constructed proposal bags around the coarse pseudo boxes. Extensive experiments on three tiny object datasets (i.e., AI-TOD-v2, SODA-A, and TinyPerson) validate the proposed method's effectiveness and robustness against point location shifts. Notably, relying solely on point supervision, our Point Teacher already shows comparable performance with box-supervised learning methods. Codes and models will be made publicly available.
- Published
- 2024
12. Cross-Self KV Cache Pruning for Efficient Vision-Language Inference
- Author
-
Pei, Xiaohuan, Huang, Tao, and Xu, Chang
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
KV cache pruning has emerged as a promising technique for reducing memory and computation costs in long-context auto-regressive generation. Existing methods for vision-language models (VLMs) typically rely on self-attention scores from large language models (LLMs) to identify and prune irrelevant tokens. However, these approaches overlook the inherent distributional discrepancies between modalities, often leading to inaccurate token importance estimation and the over-pruning of critical visual tokens. To address this, we propose decomposing attention scores into intra-modality attention (within the same modality) and inter-modality attention (across modalities), enabling more precise KV cache pruning by independently managing these distinct attention types. Additionally, we introduce an n-softmax function to counteract distribution shifts caused by pruning, preserving the original smoothness of attention scores and ensuring stable performance. Our final training-free method, \textbf{C}ross-\textbf{S}elf \textbf{P}runing (CSP), achieves competitive performance compared to models with full KV caches while significantly outperforming previous pruning methods. Extensive evaluations on MileBench, a benchmark encompassing 29 multimodal datasets, demonstrate CSP's effectiveness, achieving up to a 41\% performance improvement on challenging tasks like conversational embodied dialogue while reducing the KV cache budget by 13.6\%. The code is available at https://github.com/TerryPei/CSP
- Published
- 2024
13. Frequency-Adaptive Low-Latency Object Detection Using Events and Frames
- Author
-
Zhang, Haitian, Wang, Xiangyuan, Xu, Chang, Wang, Xinya, Xu, Fang, Yu, Huai, Yu, Lei, and Yang, Wen
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
Fusing Events and RGB images for object detection leverages the robustness of Event cameras in adverse environments and the rich semantic information provided by RGB cameras. However, two critical mismatches: low-latency Events \textit{vs.}~high-latency RGB frames; temporally sparse labels in training \textit{vs.}~continuous flow in inference, significantly hinder the high-frequency fusion-based object detection. To address these challenges, we propose the \textbf{F}requency-\textbf{A}daptive Low-Latency \textbf{O}bject \textbf{D}etector (FAOD). FAOD aligns low-frequency RGB frames with high-frequency Events through an Align Module, which reinforces cross-modal style and spatial proximity to address the Event-RGB Mismatch. We further propose a training strategy, Time Shift, which enforces the module to align the prediction from temporally shifted Event-RGB pairs and their original representation, that is, consistent with Event-aligned annotations. This strategy enables the network to use high-frequency Event data as the primary reference while treating low-frequency RGB images as supplementary information, retaining the low-latency nature of the Event stream toward high-frequency detection. Furthermore, we observe that these corrected Event-RGB pairs demonstrate better generalization from low training frequency to higher inference frequencies compared to using Event data alone. Extensive experiments on the PKU-DAVIS-SOD and DSEC-Detection datasets demonstrate that our FAOD achieves SOTA performance. Specifically, in the PKU-DAVIS-SOD Dataset, FAOD achieves 9.8 points improvement in terms of the mAP in fully paired Event-RGB data with only a quarter of the parameters compared to SODFormer, and even maintains robust performance (only a 3 points drop in mAP) under 80$\times$ Event-RGB frequency mismatch.
- Published
- 2024
14. UniGraspTransformer: Simplified Policy Distillation for Scalable Dexterous Robotic Grasping
- Author
-
Wang, Wenbo, Wei, Fangyun, Zhou, Lei, Chen, Xi, Luo, Lin, Yi, Xiaohan, Zhang, Yizhong, Liang, Yaobo, Xu, Chang, Lu, Yan, Yang, Jiaolong, and Guo, Baining
- Subjects
Computer Science - Robotics - Abstract
We introduce UniGraspTransformer, a universal Transformer-based network for dexterous robotic grasping that simplifies training while enhancing scalability and performance. Unlike prior methods such as UniDexGrasp++, which require complex, multi-step training pipelines, UniGraspTransformer follows a streamlined process: first, dedicated policy networks are trained for individual objects using reinforcement learning to generate successful grasp trajectories; then, these trajectories are distilled into a single, universal network. Our approach enables UniGraspTransformer to scale effectively, incorporating up to 12 self-attention blocks for handling thousands of objects with diverse poses. Additionally, it generalizes well to both idealized and real-world inputs, evaluated in state-based and vision-based settings. Notably, UniGraspTransformer generates a broader range of grasping poses for objects in various shapes and orientations, resulting in more diverse grasp strategies. Experimental results demonstrate significant improvements over state-of-the-art, UniDexGrasp++, across various object categories, achieving success rate gains of 3.5%, 7.7%, and 10.1% on seen objects, unseen objects within seen categories, and completely unseen objects, respectively, in the vision-based setting. Project page: https://dexhand.github.io/UniGraspTransformer., Comment: Project page: https://dexhand.github.io/UniGraspTransformer
- Published
- 2024
15. Learning Visual Abstract Reasoning through Dual-Stream Networks
- Author
-
Zhao, Kai, Xu, Chang, and Si, Bailu
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning - Abstract
Visual abstract reasoning tasks present challenges for deep neural networks, exposing limitations in their capabilities. In this work, we present a neural network model that addresses the challenges posed by Raven's Progressive Matrices (RPM). Inspired by the two-stream hypothesis of visual processing, we introduce the Dual-stream Reasoning Network (DRNet), which utilizes two parallel branches to capture image features. On top of the two streams, a reasoning module first learns to merge the high-level features of the same image. Then, it employs a rule extractor to handle combinations involving the eight context images and each candidate image, extracting discrete abstract rules and utilizing an multilayer perceptron (MLP) to make predictions. Empirical results demonstrate that the proposed DRNet achieves state-of-the-art average performance across multiple RPM benchmarks. Furthermore, DRNet demonstrates robust generalization capabilities, even extending to various out-of-distribution scenarios. The dual streams within DRNet serve distinct functions by addressing local or spatial information. They are then integrated into the reasoning module, leveraging abstract rules to facilitate the execution of visual reasoning tasks. These findings indicate that the dual-stream architecture could play a crucial role in visual abstract reasoning., Comment: 10 pages, 6 figures
- Published
- 2024
- Full Text
- View/download PDF
16. Reinforcement learning-enhanced genetic algorithm for wind farm layout optimization
- Author
-
Dong, Guodan, Qin, Jianhua, Wu, Chutian, Xu, Chang, and Yang, Xiaolei
- Subjects
Computer Science - Neural and Evolutionary Computing - Abstract
A reinforcement learning-enhanced genetic algorithm (RLGA) is proposed for wind farm layout optimization (WFLO) problems. While genetic algorithms (GAs) are among the most effective and accessible methods for WFLO, their performance and convergence are highly sensitive to parameter selections. To address the issue, reinforcement learning (RL) is introduced to dynamically select optimal parameters throughout the GA process. To illustrate the accuracy and efficiency of the proposed RLGA, we evaluate the WFLO problem for four layouts (aligned, staggered, sunflower, and unstructured) under unidirectional uniform wind, comparing the results with those from the GA. RLGA achieves similar results to GA for aligned and staggered layouts and outperforms GA for sunflower and unstructured layouts, demonstrating its efficiency. The sunflower and unstructured layouts' complexity highlights RLGA's robustness and efficiency in tackling complex problems. To further validate its capabilities, we investigate larger wind farms with varying turbine placements ($\Delta x = \Delta y = 5D$ and 2$D$, where $D$ is the wind turbine diameter) under three wind conditions: unidirectional, omnidirectional, and non-uniform, presenting greater challenges. The proposed RLGA is about three times more efficient than GA, especially for complex problems. This improvement stems from RL's ability to adjust parameters, avoiding local optima and accelerating convergence.
- Published
- 2024
17. Unsupervised Multi-view UAV Image Geo-localization via Iterative Rendering
- Author
-
Li, Haoyuan, Xu, Chang, Yang, Wen, Mi, Li, Yu, Huai, and Zhang, Haijian
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Robotics ,Electrical Engineering and Systems Science - Image and Video Processing - Abstract
Unmanned Aerial Vehicle (UAV) Cross-View Geo-Localization (CVGL) presents significant challenges due to the view discrepancy between oblique UAV images and overhead satellite images. Existing methods heavily rely on the supervision of labeled datasets to extract viewpoint-invariant features for cross-view retrieval. However, these methods have expensive training costs and tend to overfit the region-specific cues, showing limited generalizability to new regions. To overcome this issue, we propose an unsupervised solution that lifts the scene representation to 3d space from UAV observations for satellite image generation, providing robust representation against view distortion. By generating orthogonal images that closely resemble satellite views, our method reduces view discrepancies in feature representation and mitigates shortcuts in region-specific image pairing. To further align the rendered image's perspective with the real one, we design an iterative camera pose updating mechanism that progressively modulates the rendered query image with potential satellite targets, eliminating spatial offsets relative to the reference images. Additionally, this iterative refinement strategy enhances cross-view feature invariance through view-consistent fusion across iterations. As such, our unsupervised paradigm naturally avoids the problem of region-specific overfitting, enabling generic CVGL for UAV images without feature fine-tuning or data-driven training. Experiments on the University-1652 and SUES-200 datasets demonstrate that our approach significantly improves geo-localization accuracy while maintaining robustness across diverse regions. Notably, without model fine-tuning or paired training, our method achieves competitive performance with recent supervised methods., Comment: 13 pages
- Published
- 2024
18. FLMarket: Enabling Privacy-preserved Pre-training Data Pricing for Federated Learning
- Author
-
Wen, Zhenyu, Feng, Wanglei, Wu, Di, Hu, Haozhen, Xu, Chang, Qian, Bin, Hong, Zhen, Wang, Cong, and Ji, Shouling
- Subjects
Computer Science - Machine Learning ,Computer Science - Distributed, Parallel, and Cluster Computing - Abstract
Federated Learning (FL), as a mainstream privacy-preserving machine learning paradigm, offers promising solutions for privacy-critical domains such as healthcare and finance. Although extensive efforts have been dedicated from both academia and industry to improve the vanilla FL, little work focuses on the data pricing mechanism. In contrast to the straightforward in/post-training pricing techniques, we study a more difficult problem of pre-training pricing without direct information from the learning process. We propose FLMarket that integrates a two-stage, auction-based pricing mechanism with a security protocol to address the utility-privacy conflict. Through comprehensive experiments, we show that the client selection according to FLMarket can achieve more than 10% higher accuracy in subsequent FL training compared to state-of-the-art methods. In addition, it outperforms the in-training baseline with more than 2% accuracy increase and 3x run-time speedup.
- Published
- 2024
19. Harnessing Vision Foundation Models for High-Performance, Training-Free Open Vocabulary Segmentation
- Author
-
Shi, Yuheng, Dong, Minjing, and Xu, Chang
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
While Contrastive Language-Image Pre-training (CLIP) has advanced open-vocabulary predictions, its performance on semantic segmentation remains suboptimal. This shortfall primarily stems from its spatial-invariant semantic features and constrained resolution. While previous adaptations addressed spatial invariance semantic by modifying the self-attention in CLIP's image encoder, the issue of limited resolution remains unexplored. Different from previous segment-then-splice methods that segment sub-images via a sliding window and splice the results, we introduce a splice-then-segment paradigm that incorporates Segment-Anything Model (SAM) to tackle the resolution issue since SAM excels at extracting fine-grained semantic correlations from high-resolution images. Specifically, we introduce Trident, a training-free framework that first splices features extracted by CLIP and DINO from sub-images, then leverages SAM's encoder to create a correlation matrix for global aggregation, enabling a broadened receptive field for effective segmentation. Besides, we propose a refinement strategy for CLIP's coarse segmentation outputs by transforming them into prompts for SAM, further enhancing the segmentation performance. Trident achieves a significant improvement in the mIoU across eight benchmarks compared with the current SOTA, increasing from 44.4 to 48.6.Code is available at https://github.com/YuHengsss/Trident., Comment: 12 pages, 5 figures
- Published
- 2024
20. Investigating Memorization in Video Diffusion Models
- Author
-
Chen, Chen, Liu, Enhuai, Liu, Daochang, Shah, Mubarak, and Xu, Chang
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Diffusion models, widely used for image and video generation, face a significant limitation: the risk of memorizing and reproducing training data during inference, potentially generating unauthorized copyrighted content. While prior research has focused on image diffusion models (IDMs), video diffusion models (VDMs) remain underexplored. To address this gap, we first formally define the two types of memorization in VDMs (content memorization and motion memorization) in a practical way that focuses on privacy preservation and applies to all generation types. We then introduce new metrics specifically designed to separately assess content and motion memorization in VDMs. Additionally, we curate a dataset of text prompts that are most prone to triggering memorization when used as conditioning in VDMs. By leveraging these prompts, we generate diverse videos from various open-source VDMs, successfully extracting numerous training videos from each tested model. Through the application of our proposed metrics, we systematically analyze memorization across various pretrained VDMs, including text-conditional and unconditional models, on a variety of datasets. Our comprehensive study reveals that memorization is widespread across all tested VDMs, indicating that VDMs can also memorize image training data in addition to video datasets. Finally, we propose efficient and effective detection strategies for both content and motion memorization, offering a foundational approach for improving privacy in VDMs., Comment: Preprint
- Published
- 2024
21. Exploring Local Memorization in Diffusion Models via Bright Ending Attention
- Author
-
Chen, Chen, Liu, Daochang, Shah, Mubarak, and Xu, Chang
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
In this paper, we identify and leverage a novel `bright ending' (BE) anomaly in diffusion models prone to memorizing training images to address a new task: locating localized memorization regions within these models. BE refers to a distinct cross-attention pattern observed in text-to-image generations using diffusion models. Specifically, memorized image patches exhibit significantly greater attention to the end token during the final inference step compared to non-memorized patches. This attention map effectively highlights regions where the generated image replicates training data. Furthermore, driven by our observation that local memorization significantly underperforms in existing tasks of measuring, detecting, and mitigating memorization in diffusion models compared to global memorization, we propose a simple yet effective method to integrate BE and the results of the new localization task into these existing frameworks. This integration effectively improves their performances by narrowing the performance gap caused by local memorization. Our results not only demonstrate the successful execution of the new localization task but also establish new state-of-the-art performance across all existing tasks, underscoring the significance of the BE phenomenon., Comment: Preprint
- Published
- 2024
22. Diffusion Attribution Score: Evaluating Training Data Influence in Diffusion Model
- Author
-
Lin, Jinxu, Tao, Linwei, Dong, Minjing, and Xu, Chang
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence - Abstract
As diffusion models become increasingly popular, the misuse of copyrighted and private images has emerged as a major concern. One promising solution to mitigate this issue is identifying the contribution of specific training samples in generative models, a process known as data attribution. Existing data attribution methods for diffusion models typically quantify the contribution of a training sample by evaluating the change in diffusion loss when the sample is included or excluded from the training process. However, we argue that the direct usage of diffusion loss cannot represent such a contribution accurately due to the calculation of diffusion loss. Specifically, these approaches measure the divergence between predicted and ground truth distributions, which leads to an indirect comparison between the predicted distributions and cannot represent the variances between model behaviors. To address these issues, we aim to measure the direct comparison between predicted distributions with an attribution score to analyse the training sample importance, which is achieved by Diffusion Attribution Score (DAS). Underpinned by rigorous theoretical analysis, we elucidate the effectiveness of DAS. Additionally, we explore strategies to accelerate DAS calculations, facilitating its application to large-scale diffusion models. Our extensive experiments across various datasets and diffusion models demonstrate that DAS significantly surpasses previous benchmarks in terms of the linear data-modelling score, establishing new state-of-the-art performance.
- Published
- 2024
23. Non-local detection of coherent Yu-Shiba-Rusinov quantum projections
- Author
-
That, Khai Ton, Xu, Chang, Ioannidis, Ioannis, Schneider, Lucas, Posske, Thore, Wiesendanger, Roland, Morr, Dirk K., and Wiebe, Jens
- Subjects
Condensed Matter - Superconductivity ,Quantum Physics - Abstract
Probing spatially confined quantum states from afar - a long-sought goal to minimize external interference - has been proposed to be achievable in condensed matter systems via coherent projection. The latter can be tailored by sculpturing the eigenstates of the electron sea that surrounds the quantum state using atom-by-atom built cages, so-called quantum corrals. However, assuring the coherent nature of the projection, and manipulating its quantum composition, has remained an elusive goal. Here, we experimentally realize the coherent projection of a magnetic impurity-induced, Yu-Shiba-Rusinov quantum state using the eigenmodes of corrals on the surface of a superconductor, which enables us to manipulate the particle-hole composition of the projected state by tuning corral eigenmodes through the Fermi energy. Our results demonstrate a controlled non-local method for the detection of magnet superconductor hybrid quantum states., Comment: 19 pages, 5 figures
- Published
- 2024
24. Feature Clipping for Uncertainty Calibration
- Author
-
Tao, Linwei, Dong, Minjing, and Xu, Chang
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Machine Learning - Abstract
Deep neural networks (DNNs) have achieved significant success across various tasks, but ensuring reliable uncertainty estimates, known as model calibration, is crucial for their safe and effective deployment. Modern DNNs often suffer from overconfidence, leading to miscalibration. We propose a novel post-hoc calibration method called feature clipping (FC) to address this issue. FC involves clipping feature values to a specified threshold, effectively increasing entropy in high calibration error samples while maintaining the information in low calibration error samples. This process reduces the overconfidence in predictions, improving the overall calibration of the model. Our extensive experiments on datasets such as CIFAR-10, CIFAR-100, and ImageNet, and models including CNNs and transformers, demonstrate that FC consistently enhances calibration performance. Additionally, we provide a theoretical analysis that validates the effectiveness of our method. As the first calibration technique based on feature modification, feature clipping offers a novel approach to improving model calibration, showing significant improvements over both post-hoc and train-time calibration methods and pioneering a new avenue for feature-based model calibration.
- Published
- 2024
25. Consistency Calibration: Improving Uncertainty Calibration via Consistency among Perturbed Neighbors
- Author
-
Tao, Linwei, Guo, Haolan, Dong, Minjing, and Xu, Chang
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Computer Vision and Pattern Recognition - Abstract
Calibration is crucial in deep learning applications, especially in fields like healthcare and autonomous driving, where accurate confidence estimates are vital for decision-making. However, deep neural networks often suffer from miscalibration, with reliability diagrams and Expected Calibration Error (ECE) being the only standard perspective for evaluating calibration performance. In this paper, we introduce the concept of consistency as an alternative perspective on model calibration, inspired by uncertainty estimation literature in large language models (LLMs). We highlight its advantages over the traditional reliability-based view. Building on this concept, we propose a post-hoc calibration method called Consistency Calibration (CC), which adjusts confidence based on the model's consistency across perturbed inputs. CC is particularly effective in locally uncertainty estimation, as it requires no additional data samples or label information, instead generating input perturbations directly from the source data. Moreover, we show that performing perturbations at the logit level significantly improves computational efficiency. We validate the effectiveness of CC through extensive comparisons with various post-hoc and training-time calibration methods, demonstrating state-of-the-art performance on standard datasets such as CIFAR-10, CIFAR-100, and ImageNet, as well as on long-tailed datasets like ImageNet-LT.
- Published
- 2024
26. Focal surfaces of lightcone framed surfaces in the Lorentz-Minkowski 3-space
- Author
-
Xu, Chang and Chen, Liang
- Subjects
Mathematics - Differential Geometry - Abstract
In this paper, we consider the differential geometry properties of focal surfaces of lightcone framed surfaces in Lorentz-Minkowski 3-space. In general, a mixed type surface is a connected regular surface with non-empty spacelike and timelike point sets. In order to investigate the mixed type surface with singular points, we introduce the lightcone framed surface. First, we give the lightcone frame, by which we define the lightcone framed surface. Next, we consider the differential geometry properties of lightcone framed surfaces by using the lightcone frame. At last, we give the definition of focal surfaces of lightcone framed surfaces and investigate the differential geometry properties of focal surfaces.
- Published
- 2024
27. Photoacoustic tracking of photo-magnetically powered nanoparticles for cancer therapy
- Author
-
Li, Jiayan, Xu, Chang, Chen, Yingna, Cao, Junmei, Ye, Wanli, Cheng, Yu, and Cheng, Qian
- Subjects
Physics - Biological Physics ,Physics - Medical Physics - Abstract
The in vivo propulsion and monitoring of nanoparticles (NPs) have received tremendous achievements in the past decade. Developing functional NPs that can be efficiently manipulated inside the human body with a non-invasive tracking modality is critical to clinical translation. This study synthesized a photo-magnetically powered nanoparticle (PMN) with a Fe3O4 core and gold spiky surface. The Au-nanotips ensure PMNs have a strong light absorption in the second near-infrared (NIR) window and produce outstanding photoacoustic signals. The Bio-transmission electron microscopy and simulation results prove that the assembly of PMNs under a magnetic field further enhances the photothermal conversion in cells, contributing to the reduction of ambient viscosity. Photoacoustic imaging (PAI) realized real-time monitoring of PMN movements and revealed that laser plus magnetic coupling couldimprove intratumoral distribution and retention. The proposed methods exhibit excellent potential for the clinical research of cancer nanotherapies.
- Published
- 2024
28. MarS: a Financial Market Simulation Engine Powered by Generative Foundation Model
- Author
-
Li, Junjie, Liu, Yang, Liu, Weiqing, Fang, Shikai, Wang, Lewen, Xu, Chang, and Bian, Jiang
- Subjects
Quantitative Finance - Computational Finance ,Computer Science - Artificial Intelligence ,Computer Science - Computational Engineering, Finance, and Science ,Computer Science - Machine Learning ,Quantitative Finance - Trading and Market Microstructure - Abstract
Generative models aim to simulate realistic effects of various actions across different contexts, from text generation to visual effects. Despite efforts to build real-world simulators, leveraging generative models for virtual worlds, like financial markets, remains underexplored. In financial markets, generative models can simulate market effects of various behaviors, enabling interaction with market scenes and players, and training strategies without financial risk. This simulation relies on the finest structured data in financial market like orders thus building the finest realistic simulation. We propose Large Market Model (LMM), an order-level generative foundation model, for financial market simulation, akin to language modeling in the digital world. Our financial Market Simulation engine (MarS), powered by LMM, addresses the need for realistic, interactive and controllable order generation. Key objectives of this paper include evaluating LMM's scaling law in financial markets, assessing MarS's realism, balancing controlled generation with market impact, and demonstrating MarS's potential applications. We showcase MarS as a forecast tool, detection system, analysis platform, and agent training environment. Our contributions include pioneering a generative model for financial markets, designing MarS to meet domain-specific needs, and demonstrating MarS-based applications' industry potential., Comment: 19 pages, 12 figures
- Published
- 2024
29. FusionSAM: Latent Space driven Segment Anything Model for Multimodal Fusion and Segmentation
- Author
-
Li, Daixun, Xie, Weiying, Cao, Mingxiang, Wang, Yunke, Zhang, Jiaqing, Li, Yunsong, Fang, Leyuan, and Xu, Chang
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Multimodal image fusion and segmentation enhance scene understanding in autonomous driving by integrating data from various sensors. However, current models struggle to efficiently segment densely packed elements in such scenes, due to the absence of comprehensive fusion features that can guide mid-process fine-tuning and focus attention on relevant areas. The Segment Anything Model (SAM) has emerged as a transformative segmentation method. It provides more effective prompts through its flexible prompt encoder, compared to transformers lacking fine-tuned control. Nevertheless, SAM has not been extensively studied in the domain of multimodal fusion for natural images. In this paper, we introduce SAM into multimodal image segmentation for the first time, proposing a novel framework that combines Latent Space Token Generation (LSTG) and Fusion Mask Prompting (FMP) modules to enhance SAM's multimodal fusion and segmentation capabilities. Specifically, we first obtain latent space features of the two modalities through vector quantization and embed them into a cross-attention-based inter-domain fusion module to establish long-range dependencies between modalities. Then, we use these comprehensive fusion features as prompts to guide precise pixel-level segmentation. Extensive experiments on several public datasets demonstrate that the proposed method significantly outperforms SAM and SAM2 in multimodal autonomous driving scenarios, achieving at least 3.9$\%$ higher segmentation mIoU than the state-of-the-art approaches.
- Published
- 2024
30. Decoupled Video Generation with Chain of Training-free Diffusion Model Experts
- Author
-
Li, Wenhao, Cao, Yichao, Su, Xiu, Lin, Xi, You, Shan, Zheng, Mingkai, Chen, Yi, and Xu, Chang
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Video generation models hold substantial potential in areas such as filmmaking. However, current video diffusion models need high computational costs and produce suboptimal results due to extreme complexity of video generation task. In this paper, we propose \textbf{ConFiner}, an efficient video generation framework that decouples video generation into easier subtasks: structure \textbf{con}trol and spatial-temporal re\textbf{fine}ment. It can generate high-quality videos with chain of off-the-shelf diffusion model experts, each expert responsible for a decoupled subtask. During the refinement, we introduce coordinated denoising, which can merge multiple diffusion experts' capabilities into a single sampling. Furthermore, we design ConFiner-Long framework, which can generate long coherent video with three constraint strategies on ConFiner. Experimental results indicate that with only 10\% of the inference cost, our ConFiner surpasses representative models like Lavie and Modelscope across all objective and subjective metrics. And ConFiner-Long can generate high-quality and coherent videos with up to 600 frames.
- Published
- 2024
31. Controllable Financial Market Generation with Diffusion Guided Meta Agent
- Author
-
Huang, Yu-Hao, Xu, Chang, Liu, Yang, Liu, Weiqing, Li, Wu-Jun, and Bian, Jiang
- Subjects
Computer Science - Computational Engineering, Finance, and Science ,Quantitative Finance - Trading and Market Microstructure - Abstract
Order flow modeling stands as the most fundamental and essential financial task, as orders embody the minimal unit within a financial market. However, current approaches often result in unsatisfactory fidelity in generating order flow, and their generation lacks controllability, thereby limiting their application scenario. In this paper, we advocate incorporating controllability into the market generation process, and propose a Diffusion Guided meta Agent(DiGA) model to address the problem. Specifically, we utilize a diffusion model to capture dynamics of market state represented by time-evolving distribution parameters about mid-price return rate and order arrival rate, and define a meta agent with financial economic priors to generate orders from the corresponding distributions. Extensive experimental results demonstrate that our method exhibits outstanding controllability and fidelity in generation. Furthermore, we validate DiGA's effectiveness as generative environment for downstream financial applications.
- Published
- 2024
32. Compress Guidance in Conditional Diffusion Sampling
- Author
-
Dinh, Anh-Dung, Liu, Daochang, and Xu, Chang
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,I.4 - Abstract
We found that enforcing guidance throughout the sampling process is often counterproductive due to the model-fitting issue, where samples are 'tuned' to match the classifier's parameters rather than generalizing the expected condition. This work identifies and quantifies the problem, demonstrating that reducing or excluding guidance at numerous timesteps can mitigate this issue. By distributing a small amount of guidance over a large number of sampling timesteps, we observe a significant improvement in image quality and diversity while also reducing the required guidance timesteps by nearly 40%. This approach addresses a major challenge in applying guidance effectively to generative tasks. Consequently, our proposed method, termed Compress Guidance, allows for the exclusion of a substantial number of guidance timesteps while still surpassing baseline models in image quality. We validate our approach through benchmarks on label-conditional and text-to-image generative tasks across various datasets and models., Comment: 10 pages, 5 figures, Computer Vision and Machine Learning
- Published
- 2024
33. Efficient Image-to-Image Diffusion Classifier for Adversarial Robustness
- Author
-
Mei, Hefei, Dong, Minjing, and Xu, Chang
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Diffusion models (DMs) have demonstrated great potential in the field of adversarial robustness, where DM-based defense methods can achieve superior defense capability without adversarial training. However, they all require huge computational costs due to the usage of large-scale pre-trained DMs, making it difficult to conduct full evaluation under strong attacks and compare with traditional CNN-based methods. Simply reducing the network size and timesteps in DMs could significantly harm the image generation quality, which invalidates previous frameworks. To alleviate this issue, we redesign the diffusion framework from generating high-quality images to predicting distinguishable image labels. Specifically, we employ an image translation framework to learn many-to-one mapping from input samples to designed orthogonal image labels. Based on this framework, we introduce an efficient Image-to-Image diffusion classifier with a pruned U-Net structure and reduced diffusion timesteps. Besides the framework, we redesign the optimization objective of DMs to fit the target of image classification, where a new classification loss is incorporated in the DM-based image translation framework to distinguish the generated label from those of other classes. We conduct sufficient evaluations of the proposed classifier under various attacks on popular benchmarks. Extensive experiments show that our method achieves better adversarial robustness with fewer computational costs than DM-based and CNN-based methods. The code is available at https://github.com/hfmei/IDC.
- Published
- 2024
34. VSSD: Vision Mamba with Non-Causal State Space Duality
- Author
-
Shi, Yuheng, Dong, Minjing, Li, Mingjia, and Xu, Chang
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Vision transformers have significantly advanced the field of computer vision, offering robust modeling capabilities and global receptive field. However, their high computational demands limit their applicability in processing long sequences. To tackle this issue, State Space Models (SSMs) have gained prominence in vision tasks as they offer linear computational complexity. Recently, State Space Duality (SSD), an improved variant of SSMs, was introduced in Mamba2 to enhance model performance and efficiency. However, the inherent causal nature of SSD/SSMs restricts their applications in non-causal vision tasks. To address this limitation, we introduce Visual State Space Duality (VSSD) model, which has a non-causal format of SSD. Specifically, we propose to discard the magnitude of interactions between the hidden state and tokens while preserving their relative weights, which relieves the dependencies of token contribution on previous tokens. Together with the involvement of multi-scan strategies, we show that the scanning results can be integrated to achieve non-causality, which not only improves the performance of SSD in vision tasks but also enhances its efficiency. We conduct extensive experiments on various benchmarks including image classification, detection, and segmentation, where VSSD surpasses existing state-of-the-art SSM-based models. Code and weights are available at \url{https://github.com/YuHengsss/VSSD}., Comment: 16 pages, 5 figures, 7 tables
- Published
- 2024
35. Enhancing Fine-grained Object Detection in Aerial Images via Orthogonal Mapping
- Author
-
Zhu, Haoran, Zhou, Yifan, Xu, Chang, Zhang, Ruixiang, and Yang, Wen
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Fine-Grained Object Detection (FGOD) is a critical task in high-resolution aerial image analysis. This letter introduces Orthogonal Mapping (OM), a simple yet effective method aimed at addressing the challenge of semantic confusion inherent in FGOD. OM introduces orthogonal constraints in the feature space by decoupling features from the last layer of the classification branch with a class-wise orthogonal vector basis. This effectively mitigates semantic confusion and enhances classification accuracy. Moreover, OM can be seamlessly integrated into mainstream object detectors. Extensive experiments conducted on three FGOD datasets (FAIR1M, ShipRSImageNet, and MAR20) demonstrate the effectiveness and superiority of the proposed approach. Notably, with just one line of code, OM achieves a 4.08% improvement in mean Average Precision (mAP) over FCOS on the ShipRSImageNet dataset. Codes are released at https://github.com/ZhuHaoranEIS/Orthogonal-FGOD.
- Published
- 2024
36. QCD corrections of $e^+e^- \to J/\psi+c+\bar{c}$ using the principle of maximum conformality
- Author
-
Huang, Xu-Dong, Wu, Xing-Gang, Zheng, Xu-Chang, Gong, Bin, and Wang, Jian-Xiong
- Subjects
High Energy Physics - Phenomenology - Abstract
In this paper, we compute the total and differential cross sections for $e^+e^- \to J/\psi+c+\bar{c}$ at the $B$ factories up to next-to-leading order (NLO) corrections within the framework of nonrelativistic QCD factorization theory. We then obtain improved pQCD series of those cross sections by using the Principle of Maximum Conformality (PMC). We show that the PMC can be applied for any pQCD calculable observable at the total and differential levels via a self-consistent way in perturbation theory. We observe that a more precise prompt total cross section at the NLO level can be achieved after applying the PMC, e.g. $\sigma|_{\rm prompt}^{\rm PMC}= 0.565^{+0.144}_{-0.125}~\text{pb}$. Here the uncertainty is the squared average of those from the $\alpha_s$ fixed-point uncertainty $\Delta\alpha_s(M_Z)$, the uncertainty of charm quark mass $\Delta m_c$, and an estimated contribution of the uncalculated NNLO-terms as predicted by the Pad\'{e} approximation approach. The differential cross sections $d\sigma/dP_{J/\psi}$, $d\sigma/d|\cos \theta|$, and $d\sigma/dz$ for $e^+e^- \to J/\psi+c+\bar{c}$ are further examined. Those results show that by further considering the feed-down contributions, the PMC predictions show better agreement with the Belle measurements., Comment: 13 pages, 11 figures, matches published version, to be published in Phys. Rev. D
- Published
- 2024
- Full Text
- View/download PDF
37. Training-free Composite Scene Generation for Layout-to-Image Synthesis
- Author
-
Liu, Jiaqi, Huang, Tao, and Xu, Chang
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
Recent breakthroughs in text-to-image diffusion models have significantly advanced the generation of high-fidelity, photo-realistic images from textual descriptions. Yet, these models often struggle with interpreting spatial arrangements from text, hindering their ability to produce images with precise spatial configurations. To bridge this gap, layout-to-image generation has emerged as a promising direction. However, training-based approaches are limited by the need for extensively annotated datasets, leading to high data acquisition costs and a constrained conceptual scope. Conversely, training-free methods face challenges in accurately locating and generating semantically similar objects within complex compositions. This paper introduces a novel training-free approach designed to overcome adversarial semantic intersections during the diffusion conditioning phase. By refining intra-token loss with selective sampling and enhancing the diffusion process with attention redistribution, we propose two innovative constraints: 1) an inter-token constraint that resolves token conflicts to ensure accurate concept synthesis; and 2) a self-attention constraint that improves pixel-to-pixel relationships. Our evaluations confirm the effectiveness of leveraging layout information for guiding the diffusion process, generating content-rich images with enhanced fidelity and complexity. Code is available at https://github.com/Papple-F/csg.git., Comment: ECCV 2024
- Published
- 2024
38. Spontaneous Motion of Liquid Droplets on Soft Gradient Surfaces
- Author
-
Zhao, Weiwei, Qian, Wenjie, Xu, Chang, and Xu, Qin
- Subjects
Condensed Matter - Soft Condensed Matter - Abstract
We report an experimental investigation of the spontaneous motion of liquid droplets on soft gels with a crosslinking gradient. By systematically adjusting the spatial difference in crosslinking density, we observed that millimeter-sized liquid droplets moved along the gradient of elastic modulus and even climbed tilted slopes against gravity. Unlike the wetting dynamics of micro-droplets, which are governed by elastocapillary effects, we demonstrated that the observed spontaneous movements of millimeter-sized droplets were attributed to the surface energy difference resulting from the variation in crosslinking density. Using {\em in-situ} confocal microscopy imaging, we analyzed the viscoelastic dissipation induced by the moving wetting ridges near dynamic contact lines. Based on the relationship between the crosslinking density and surface energy of soft gels, our findings reveal a new method for controlling droplet dynamics at soft and dissipative interfaces.
- Published
- 2024
39. HuntFUZZ: Enhancing Error Handling Testing through Clustering Based Fuzzing
- Author
-
Wei, Jin, Chen, Ping, Dai, Jun, Sun, Xiaoyan, Zhang, Zhihao, Xu, Chang, and Wanga, Yi
- Subjects
Computer Science - Cryptography and Security - Abstract
Testing a program's capability to effectively handling errors is a significant challenge, given that program errors are relatively uncommon. To solve this, Software Fault Injection (SFI)-based fuzzing integrates SFI and traditional fuzzing, injecting and triggering errors for testing (error handling) code. However, we observe that current SFI-based fuzzing approaches have overlooked the correlation between paths housing error points. In fact, the execution paths of error points often share common paths. Nonetheless, Fuzzers usually generate test cases repeatedly to test error points on commonly traversed paths. This practice can compromise the efficiency of the fuzzer(s). Thus, this paper introduces HuntFUZZ, a novel SFI-based fuzzing framework that addresses the issue of redundant testing of error points with correlated paths. Specifically, HuntFUZZ clusters these correlated error points and utilizes concolic execution to compute constraints only for common paths within each cluster. By doing so, we provide the fuzzer with efficient test cases to explore related error points with minimal redundancy. We evaluate HuntFUZZ on a diverse set of 42 applications, and HuntFUZZ successfully reveals 162 known bugs, with 62 of them being related to error handling. Additionally, due to its efficient error point detection method, HuntFUZZ discovers 7 unique zero-day bugs, which are all missed by existing fuzzers. Furthermore, we compare HuntFUZZ with 4 existing fuzzing approaches, including AFL, AFL++, AFLGo, and EH-FUZZ. Our evaluation confirms that HuntFUZZ can cover a broader range of error points, and it exhibits better performance in terms of bug finding speed.
- Published
- 2024
40. Precise determination of the bottom-quark on-shell mass using its four-loop relation to the $\overline{\rm MS}$-scheme running mass
- Author
-
Ma, Shun-Yue, Huang, Xu-Dong, Zheng, Xu-Chang, and Wu, Xing-Gang
- Subjects
High Energy Physics - Phenomenology - Abstract
In this paper, we explore the properties of the bottom-quark on-shell mass ($M_b$) by using its relation to the $\overline{\rm MS}$ mass (${\overline m}_b$). At present, this $\overline{\rm MS}$-on-shell relation has been known up to four-loop QCD corrections, which however still has a $\sim 2\%$ scale uncertainty by taking the renormalization scale as ${\overline m}_b({\overline m}_b)$ and varying it within the usual range of $[{\overline m}_b({\overline m}_b)/2, 2 {\overline m}_b({\overline m}_b)]$. The principle of maximum conformality (PMC) has been adopted to achieve a more precise $\overline{\rm MS}$-on-shell relation by eliminating such scale uncertainty. As a step forward, we also estimate the magnitude of the uncalculated higher-order terms by using the Pad\'{e} approximation approach. Numerically, by using the $\overline{\rm MS}$ mass ${\overline m}_b({\overline m}_b)=4.183\pm0.007$ GeV as an input, our predicted value for the bottom-quark on-shell mass becomes $M_b\simeq 5.372^{+0.091}_{-0.075}$ GeV, where the uncertainty is the squared average of the ones caused by $\Delta \alpha_s(M_Z)$, $\Delta {\overline m}_b({\overline m}_b)$, and the estimated magnitude of the higher-order terms., Comment: 5 pages, 2 figures, matches published version
- Published
- 2024
- Full Text
- View/download PDF
41. Surgical Triplet Recognition via Diffusion Model
- Author
-
Liu, Daochang, Hu, Axel, Shah, Mubarak, and Xu, Chang
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
Surgical triplet recognition is an essential building block to enable next-generation context-aware operating rooms. The goal is to identify the combinations of instruments, verbs, and targets presented in surgical video frames. In this paper, we propose DiffTriplet, a new generative framework for surgical triplet recognition employing the diffusion model, which predicts surgical triplets via iterative denoising. To handle the challenge of triplet association, two unique designs are proposed in our diffusion framework, i.e., association learning and association guidance. During training, we optimize the model in the joint space of triplets and individual components to capture the dependencies among them. At inference, we integrate association constraints into each update of the iterative denoising process, which refines the triplet prediction using the information of individual components. Experiments on the CholecT45 and CholecT50 datasets show the superiority of the proposed method in achieving a new state-of-the-art performance for surgical triplet recognition. Our codes will be released.
- Published
- 2024
42. Locating and Extracting Relational Concepts in Large Language Models
- Author
-
Wang, Zijian, White, Britney, and Xu, Chang
- Subjects
Computer Science - Computation and Language - Abstract
Relational concepts are indeed foundational to the structure of knowledge representation, as they facilitate the association between various entity concepts, allowing us to express and comprehend complex world knowledge. By expressing relational concepts in natural language prompts, people can effortlessly interact with large language models (LLMs) and recall desired factual knowledge. However, the process of knowledge recall lacks interpretability, and representations of relational concepts within LLMs remain unknown to us. In this paper, we identify hidden states that can express entity and relational concepts through causal mediation analysis in fact recall processes. Our finding reveals that at the last token position of the input prompt, there are hidden states that solely express the causal effects of relational concepts. Based on this finding, we assume that these hidden states can be treated as relational representations and we can successfully extract them from LLMs. The experimental results demonstrate high credibility of the relational representations: they can be flexibly transplanted into other fact recall processes, and can also be used as robust entity connectors. Moreover, we also show that the relational representations exhibit significant potential for controllable fact recall through relation rewriting.
- Published
- 2024
43. JavaBench: A Benchmark of Object-Oriented Code Generation for Evaluating Large Language Models
- Author
-
Cao, Jialun, Chen, Zhiyong, Wu, Jiarong, Cheung, Shing-chi, and Xu, Chang
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Programming Languages ,Computer Science - Software Engineering - Abstract
Code generation benchmarks such as HumanEval are widely adopted to evaluate LLMs' capabilities. However, after consolidating the latest 24 benchmarks, we noticed three significant imbalances. First, imbalanced programming language. 95.8% of benchmarks involve Python, while only 5 benchmarks involve Java. Second, imbalanced code granularity. Function-/statement-level benchmarks account for over 83.3% of benchmarks. Only a mere handful extends to class-/project-levels, and all are limited to Python. Third, lacking advanced features. Existing benchmarks primarily assess basic coding skills, while overlooking advanced Object-Oriented Programming (OOP) features (i.e., encapsulation, inheritance, and polymorphism). To fill these gaps, we propose JavaBench, a project-level Java benchmark that exercises OOP features. It comprises four Java projects with 389 methods in 106 Java classes. The test coverage is up to 92%, and JavaBench is attested by 282 undergraduate students, reaching a 90.93/100 average score (i.e., pass rate against the test suite), ensuring the quality of documentation, code skeleton, and tests. To better evaluate LLM's capability against JavaBench, we introduce a systematic evaluation design covering three context settings and five synthesis strategies at two granularities using three hierarchical metrics. Our extensive experiment yields several interesting findings. First, we noticed that regarding project-level Java programming, LLMs are far behind undergraduate students (no project can be correctly completed by any studied LLMs, and at most 41.17% Pass@5 in a more relaxed evaluation). Second, using method signature as prompt context may strike an ideal balance for project-level code generation. JavaBench is publicly available at https://github.com/java-bench/JavaBench., Comment: Accepted by ASE 2024
- Published
- 2024
44. GeminiFusion: Efficient Pixel-wise Multimodal Fusion for Vision Transformer
- Author
-
Jia, Ding, Guo, Jianyuan, Han, Kai, Wu, Han, Zhang, Chao, Xu, Chang, and Chen, Xinghao
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Cross-modal transformers have demonstrated superiority in various vision tasks by effectively integrating different modalities. This paper first critiques prior token exchange methods which replace less informative tokens with inter-modal features, and demonstrate exchange based methods underperform cross-attention mechanisms, while the computational demand of the latter inevitably restricts its use with longer sequences. To surmount the computational challenges, we propose GeminiFusion, a pixel-wise fusion approach that capitalizes on aligned cross-modal representations. GeminiFusion elegantly combines intra-modal and inter-modal attentions, dynamically integrating complementary information across modalities. We employ a layer-adaptive noise to adaptively control their interplay on a per-layer basis, thereby achieving a harmonized fusion process. Notably, GeminiFusion maintains linear complexity with respect to the number of input tokens, ensuring this multimodal framework operates with efficiency comparable to unimodal networks. Comprehensive evaluations across multimodal image-to-image translation, 3D object detection and arbitrary-modal semantic segmentation tasks, including RGB, depth, LiDAR, event data, etc. demonstrate the superior performance of our GeminiFusion against leading-edge techniques. The PyTorch code is available at https://github.com/JiaDingCN/GeminiFusion, Comment: Accepted by ICML 2024, code and models are available at https://github.com/JiaDingCN/GeminiFusion
- Published
- 2024
45. Superatomic catalyst for efficient nitric oxide reduction using Al8O3 superatom-oxide framework
- Author
-
Shen, Kaidong, Yi, Jiuqi, Xu, Chang, Luo, Qiquan, Wu, Xiaojun, and Yang, Jinlong
- Published
- 2025
- Full Text
- View/download PDF
46. Correction: CEPC Technical Design Report: Accelerator
- Author
-
Abdallah, Waleed, de Freitas, Tiago CarlosAdorno, Afanaciev, Konstantin, Ahmad, Shakeel, Ahmed, Ijaz, Ai, Xiaocong, Aleem, Abid, Altmannshofer, Wolfgang, Alves, Fabio, An, Weiming, An, Rui, Anderle, Daniele Paolo, Antusch, Stefan, Arai, Yasuo, Arbuzov, Andrej, Arhrib, Abdesslam, Ashry, Mustafa, Bai, Sha, Bai, Yu, Bai, Yang, Bairathi, Vipul, Balazs, Csaba, Bambade, Philip, Ban, Yong, Bandyopadhyay, Triparno, Bao, Shou-Shan, Barber, Desmond P., Bat, Ays¸e, Batozskaya, Varvara, Behera, Subash Chandra, Belyaev, Alexander, Bertucci, Michele, Bi, Xiao-Jun, Bi, Yuanjie, Bian, Tianjian, Bianchi, Fabrizio, Bieko¨tter, Thomas, Biglietti, Michela, Bilanishvili, Shalva, Binglin, Deng, Bodrov, Denis, Bogomyagkov, Anton, Bondarenko, Serge, Boogert, Stewart, Boonekamp, Maarten, Borri, Marcello, Bosotti, Angelo, Boudry, Vincent, Boukidi, Mohammed, Boyko, Igor, Bozovic, Ivanka, Bozzi, Giuseppe, Brient, Jean-Claude, Budzinskaya, Anastasiia, Bukhari, Masroor, Bytev, Vladimir, Cacciapaglia, Giacomo, Cai, Hua, Cai, Wenyong, Cai, Wujun, Cai, Yijian, Cai, Yizhou, Cai, Yuchen, Cai, Haiying, Cai, Huacheng, Calibbi, Lorenzo, Cang, Junsong, Cao, Guofu, Cao, Jianshe, Chance, Antoine, Chang, Xuejun, Chang, Yue, Chang, Zhe, Chang, Xinyuan, Chao, Wei, Chatrabhuti, Auttakit, Che, Yimin, Che, Yuzhi, Chen, Bin, Chen, Danping, Chen, Fuqing, Chen, Fusan, Chen, Gang, Chen, Guoming, Chen, Hua-Xing, Chen, Huirun, Chen, Jinhui, Chen, Ji-Yuan, Chen, Kai, Chen, Mali, Chen, Mingjun, Chen, Mingshui, Chen, Ning, Chen, Shanhong, Chen, Shanzhen, Chen, Shao-Long, Chen, Shaomin, Chen, Shiqiang, Chen, Tianlu, Chen, Wei, Chen, Xiang, Chen, Xiaoyu, Chen, Xin, Chen, Xun, Chen, Xurong, Chen, Ye, Chen, Ying, Chen, Yukai, Chen, Zelin, Chen, Zilin, Chen, Gang, Chen, Boping, Chen, Chunhui, Cheng, Hok Chuen, Cheng, Huajie, Cheng, Shan, Cheng, Tongguang, Chi, Yunlong, Chimenti, Pietro, Chiu, Wen Han, Cho, Guk, Chu, Ming-Chung, Chu, Xiaotong, Chu, Ziliang, Coloretti, Guglielmo, Crivellin, Andreas, Cui, Hanhua, Cui, Xiaohao, Cui, Zhaoyuan, D’Anzi, Brunella, Dai, Ling-Yun, Dai, Xinchen, Dai, Xuwen, De Maria, Antonio, De Filippis, Nicola, De La Taille, Christophe, De Mori, Francesca, De Sio, Chiara, Del Core, Elisa, Deng, Shuangxue, Deng, Wei-Tian, Deng, Zhi, Deng, Ziyan, Dev, Bhupal, Dewen, Tang, Di Micco, Biagio, Ding, Ran, Ding, Siqin, Ding, Yadong, Dong, Haiyi, Dong, Jianing, Dong, Jing, Dong, Lan, Dong, Mingyi, Dong, Xu, Dong, Yipei, Dong, Yubing, Dordevic, Milos, Drewes, Marco, Du, Mingxuan, Du, Mingxuan, Du, Qianqian, Du, Xiaokang, Du, Yanyan, Du, Yong, Du, Yunfei, Duan, Chun-Gui, Duan, Zhe, Dydyshka, Yahor, Egede, Ulrik, Elmetenawee, Walaa, Eo, Yun, Fan, Ka Yan, Fan, Kuanjun, Fan, Yunyun, Fang, Bo, Fang, Shuangshi, Fang, Yuquan, Farilla, Ada, Farinelli, Riccardo, Farooq, Muhammad, Golfe, Angeles Faus, Fazliakhmetov, Almaz, Fei, Rujun, Feng, Bo, Feng, Chong, Feng, Junhua, Feng, Xu, Feng, Zhuoran, ZhuoranFeng, Castillo, Luis Roberto Flores, Forest, Etienne, Fowlie, Andrew, Fox, Harald, Fu, Hai-Bing, Fu, Jinyu, Fuks, Benjamin, Funakoshi, Yoshihiro, Gabrielli, Emidio, Gan, Nan, Gang, Li, Gao, Jie, Gao, Meisen, Gao, Wenbin, Gao, Wenchun, Gao, Yu, Gao, Yuanning, Gao, Zhanxiang, Gao, Yanyan, Ge, Kun, Ge, Shao-Feng, Ge, Zhenwu, Geng, Li-Sheng, Geng, Qinglin, Geng, Chao-Qiang, Ghosh, Swagata, Gioiosa, Antonio, Gladilin, Leonid, Gong, Ti, Gori, Stefania, Gou, Quanbu, Grinstein, Sebastian, Gu, Chenxi, Guillermo, Gerardo, da Costa, Joao Guimaraes, Guo, Dizhou, Guo, Fangyi, Guo, Jiacheng, Guo, Jun, Guo, Lei, Guo, Lei, Guo, Xia, Guo, Xin-Heng, Guo, Xinyang, Guo, Yun, Guo, Yunqiang, Guo, Yuping, Guo, Zhi-Hui, Gutie´rrez-Rodríguez, Alejandro, Ha, Seungkyu, Habib, Noman, Hajer, Jan, Hammer, Francois, Han, Chengcheng, Han, Huayong, Han, Jifeng, Han, Liang, Han, Liangliang, Han, Ruixiong, Han, Yang, Han, Yezi, Han, Yuanying, Han, Tao, Hao, Jiankui, Hao, Xiqing, XiqingHao, He, Chuanqi, He, Dayong, He, Dongbing, He, Guangyuan, He, Hong-Jian, He, Jibo, He, Jun, He, Longyan, He, Xiang, He, Xiao-Gang, He, Zhenqiang, Heinemann, Klaus, Heinemeyer, Sven, Heng, Yuekun, Herna´ndez-Ruíz, María A., Hong, Jiamin, Hor, Yuenkeung, Hou, George W. S., Hou, Xiantao, Hou, Xiaonan, Hou, Zhilong, Hou, Suen, Hu, Caishi, Hu, Chen, Hu, Dake, Hu, Haiming, Hu, Jiagen, Hu, Jun, Hu, Kun, Hu, Shouyang, Hu, Yongcai, Hu, Yu, Hu, Zhen, Hua, Zhehao, Hua, Jianfei, Huang, Chao-Shang, Huang, Fa Peng, Huang, Guangshun, Huang, Jinshu, Huang, Ke, Huang, Liangsheng, Huang, Shuhui, Huang, Xingtao, Huang, Xu-Guang, Huang, Yanping, Huang, Yonggang, Huang, Yongsheng, Huang, Zimiao, Huanyuan, Chen, Huh, Changgi, Hui, Jiaqi, Huo, Lihua, Hussain, Talab, Hwang, Kyuyeong, Ioannisian, Ara, Iqbal, Munawar, Jackson, Paul, Jafarzade, Shahriyar, Jang, Haeun, Jang, Seoyun, Ji, Daheng, Ji, Qingping, Ji, Quan, Ji, Xiaolu, Jia, Jingguang, Jia, Jinsheng, Jia, Xuewei, Jia, Zihang, Jiang, Cailian, Jiang, Han Ren, Jiang, Houbing, Jiang, Jun, Jiang, Xiaowei, Jiang, Xin, Jiang, Xuhui, Jiang, Yongcheng, Jiang, Zhongjian, Jiang, Cheng, Jiao, Ruiqi, Jin, Dapeng, Jin, Shan, Jin, Song, Jin, Yi, Jis, Junji, Jung, Sunghoon, Kacarevic, Goran, Kajfasz, Eric, Kalinovskaya, Lidia, Kampf, Aleksei, Kang, Wen, Kang, Xian-Wei, Kang, Xiaolin, Karmakar, Biswajit, Ke, Zhiyong, Keloth, Rijeesh, Khan, Alamgir, Khanpour, Hamzeh, Khosonthongkee, Khanchai, KhanchaiKhosonthongkee, Kim, Bobae, Kim, Dongwoon, Kim, Mi Ran, Kim, Minsuk, Kim, Sungwon, Kim, On, Klasen, Michael, Ko, Sanghyun, Koop, Ivan, Kornienko, Vitaliy, Kortman, Bryan, Kozlov, Gennady, Kuang, Shiqing, Kumar, Mukesh, Kuo, Chia Ming, Kwok, Tsz Hong, Lagarde, Franc¸ois Sylvain Ren, Lai, Pei-Zhu, Laktineh, Imad, Lan, Xiaofei, Lan, Zuxiu, Lavezzi, Lia, Lee, Justin, Lee, Junghyun, Lee, Sehwook, Lei, Ge, Lemmon, Roy, Leng, Yongxiang, Leung, Sze Ching, Li, Hai Tao, Li, Bingzhi, Li, Bo, Li, Bo, Li, Changhong, Li, Chao, Li, Cheng, Li, Cheng, Li, Chunhua, Li, Cui, Li, Dazhang, Li, Dikai, Li, Fei, Li, Gang, Li, Gang, Li, Gang, Li, Gaosong, Li, Haibo, Li, Haifeng, Li, Hai-Jun, Li, Haotian, Li, Hengne, Li, Honglei, Li, Huijing, Li, Jialin, Li, Jingyi, Li, Jinmian, Li, Jun, Li, Leyi, Li, Liang, Li, Ling, Li, Mei, Li, Meng, Li, Minxian, Li, Pei-Rong, Li, Qiang, Li, Shaopeng, Li, Shenghe, Li, Shu, Li, Shuo, Li, Teng, Li, Tiange, Li, Tong, Li, Weichang, Li, Weidong, Li, Wenjun, Li, Xiaoling, Li, Xiaomei, Li, Xiaonan, Li, Xiaoping, Li, Xiaoting, Li, Xin, Li, Xinqiang, Li, Xuekang, Li, Yang, Li, Yanwei, Li, Yiming, Li, Ying, Li, Ying-Ying, Li, Yonggang, Li, Yonglin, Li, Yufeng, Li, Yuhui, Li, Zhan, Li, Zhao, Li, Zhiji, Li, Tong, Li, Lingfeng, Li, Fei, Liang, Jing, Liang, Jinhan, Liang, Zhijun, Liao, Guangrui, Liao, Hean, Liao, Jiajun, Liao, Libo, Liao, Longzhou, Liao, Yi, Liao, Yipu, Limphirat, Ayut, AyutLimphirat, Lin, Tao, Lin, Weiping, Lin, Yufu, Lin, Yugen, Liu, Beijiang, Liu, Bo, Liu, Danning, Liu, Dong, Liu, Fu-Hu, Liu, Hongbang, Liu, Huangcheng, Liu, Hui, Liu, Huiling, Liu, Jia, Liu, Jia, Liu, Jiaming, Liu, Jianbei, Liu, Jianyi, Liu, Jingdong, Liu, Jinhua, Liu, Kai, Liu, Kang, Liu, Kun, Liu, Mengyao, Liu, Peng, Liu, Pengcheng, Liu, Qibin, Liu, Shan, Liu, Shidong, Liu, Shuang, Liu, Shubin, Liu, Tao, Liu, Tao, Liu, Tong, Liu, Wei, Liu, Xiang, Liu, Xiao-Hai, Liu, Xiaohui, Liu, Xiaoyu, Liu, Xin, Liu, Xinglin, Liu, Xingquan, Liu, Yang, Liu, Yanlin, Liu, Yao-Bei, Liu, Yi, Liu, Yiming, Liu, Yong, Liu, Yonglu, Liu, Yu, Liu, Yubin, Liu, Yudong, Liu, Yulong, Liu, Zhaofeng, Liu, Zhen, Liu, Zhenchao, Liu, Zhi, Liu, Zhi-Feng, Liu, Zhiqing, Liu, Zhongfu, Liu, Zuowei, Liu, Mia, Liu, Zhen, Liu, Xiaoyang, Lou, Xinchou, Lu, Cai-Dian, Lu, Jun-Xu, Lu, Qiu Zhen, Lu, Shang, Lu, Shang, Lu, Wenxi, Lu, Xiaohan, Lu, Yunpeng, Lu, Zhiyong, Lu, Xianguo, Lu, Wei, Lubsandorzhiev, Bayarto, Lubsandorzhiev, Sultim, Lukanov, Arslan, Luo, Jinliang, Luo, Tao, Luo, xiaoan, Luo, Xiaofeng, Luo, Xiaolan, Lv, Jindong, Lyu, Feng, Lyu, Xiao-Rui, Lyu, Kun-Feng, Ma, Ande, Ma, Hong-Hao, Ma, Jun-Li, Ma, Kai, Ma, Lishuang, Ma, Na, Ma, Renjie, Ma, Weihu, Ma, Xinpeng, Ma, Yanling, Ma, Yan-Qing, Ma, Yongsheng, Ma, Zhonghui, Ma, Zhongjian, Ma, Yang, Maity, Mousam, Mao, Lining, Mao, Yanmin, Mao, Yaxian, Martens, Aure´lien, Maria, Caccia Massimo Luigi, Matsumoto, Shigeki, Mellado, Bruce, Meloni, Davide, Men, Lingling, Meng, Cai, Meng, Lingxin, Mi, Zhenghui, Miao, Yuhui, Migliorati, Mauro, Ming, Lei, Mitsou, Vasiliki A., Monaco, Laura, Moraes, Arthur, Mosala, Karabo, Moursy, Ahmad, Mu, Lichao, Mu, Zhihui, Muchnoi, Nickolai, Muenstermann, Daniel, Muenstermann, Daniel, Munbodh, Pankaj, Murray, William John, Nanni, Jérôme, Nanzanov, Dmitry, Nie, Changshan, Nikitin, Sergei, Ning, Feipeng, Ning, Guozhu, Niu, Jia-Shu, Niu, Juan-Juan, Niu, Yan, Nkadimeng, Edward Khomotso, Ohmi, Kazuhito, Oide, Katsunobu, Okawa, Hideki, Ouchemhou, Mohamed, Ouyang, Qun, Paesani, Daniele, Pagani, Carlo, Paganis, Stathes, Pakuza, Collette, Pan, Jiangyang, Pan, Juntong, Pan, Tong, Pan, Xiang, Panda, Papia, Pandey, Saraswati, Pandurovic, Mila, Paparella, Rocco, Pasechnik, Roman, Passemar, Emilie, Pei, Hua, Peng, Xiaohua, Peng, Xinye, Peng, Yuemei, Ping, Jialun, Ping, Ronggang, Adhya, Souvik Priyam, Qi, Baohua, Qi, Hang, Qi, Huirong, Qi, Ming, Qian, Sen, Qian, Zhuoni, Qiao, Congfeng, Qin, Guangyou, Qin, Jiajia, Qin, Laishun, Qin, Liqing, Qin, Qin, Qin, Xiaoshuai, Qin, Zhonghua, Qu, Guofeng, Racioppi, Antonio, Ramsey-Musolf, Michael, Raza, Shabbar, Rekovic, Vladimir, Ren, Jing, Reuter, Ju¨rgen, Robens, Tania, Rossi, Giancarlo, Ruan, Manqi, Ruan, Manqi, Rumyantsev, Leonid, Ryu, Min Sang, Sadykov, Renat, Sang, Minjing, Sanz-Cillero, Juan Jose´, Saur, Miroslav, Savla, Nishil, Schmidt, Michael A., Sertore, Daniele, Settles, Ron, Sha, Peng, Shao, Ding-Yu, Shao, Ligang, Shao, Hua-Sheng, She, Xin, Shen, Chuang, Shen, Hong-Fei, Shen, Jian-Ming, Shen, Peixun, Shen, Qiuping, Shen, Zhongtao, Sheng, Shuqi, Shi, Haoyu, Shi, Hua, Shi, Qi, Shi, Shusu, Shi, Xiaolei, Shi, Xin, Shi, Yukun, Shi, Zhan, Shipsey, Ian, Shiu, Gary, Shu, Chang, Si, Zong-Guo, Sidorenkov, Andrei, Smiljanić, Ivan, Song, Aodong, Song, Huayang, Song, Jiaojiao, Song, Jinxing, Song, Siyuan, Song, Weimin, Song, Weizheng, Song, Zhi, Sourav, Shashwat, Spruzzola, Paolo, Su, Feng, Su, Shengsen, Su, Wei, Su, Shufang, Sui, Yanfeng, Sui, Zexuan, Sullivan, Michael, Sun, Baiyang, Sun, Guoqiang, Sun, Hao, Sun, Hao-Kai, Sun, Junfeng, Sun, Liang, Sun, Mengcheng, Sun, Pengfei, Sun, Sichun, Sun, Xianjing, Sun, Xiaohu, Sun, Xilei, Sun, Xingyang, Sun, Xin-Yuan, Sun, Yanjun, Sun, Yongzhao, Sun, Yue, Sun, Zheng, Sun, Zheng, Suwonjandee, Narumon, Eldin, Elsayed Tag, Tan, Biao, Tang, Bo, Tang, Chuanxiang, Tang, Gao, Tang, Guangyi, Tang, Jian, Tang, Jingyu, Tang, Liang, Tang, Ying’Ao, Tao, Junquan, Tawfik, Abdel Nasser, Taylor, Geoffrey, Telnov, Valery, Tian, Saike, Torre, Riccardo, Trzaska, Wladyslaw Henryk, Tsybychev, Dmitri, Tu, Yanjun, Tuo, Shengquan, Tytgat, Michael, Islam, Ghalib Ul, Ushakov, Nikita, Valencia, German, Velthuis, Jaap, Vicini, Alessandro, Vickey, Trevor, Vidakovic, Ivana, Videau, Henri, Volkas, Raymond, Voronin, Dmitry, Vukasinovic, Natasa, Wan, Xia, Wan, Xuying, Wang, Xiao, Wang, Anqing, Wang, Bin, Wang, Chengtao, Wang, Chuanye, Wang, Ci, Wang, Dayong, Wang, Dou, Wang, En, Wang, Fei, Wang, Fei, Wang, Guanwen, Wang, Guo-Li, Wang, Haijing, Wang, Haolin, Wang, Jia, Wang, Jian, Wang, Jianchun, Wang, Jianli, Wang, Jiawei, Wang, Jin, Wang, Jin-Wei, Wang, Joseph, Wang, Kechen, Wang, Lechun, Wang, Lei, Wang, Liguo, Wang, Lijiao, Wang, Lu, Wang, Meng, Wang, Na, Wang, Pengcheng, Wang, Qian, Wang, Qun, Wang, Shu Lin, Wang, Shudong, Wang, Taofeng, Wang, Tianhong, Wang, Tianyang, Wang, Tong, Wang, Wei, Wang, Wei, Wang, Xiaolong, Wang, Xiaolong, Wang, Xiaoning, Wang, Xiao-Ping, Wang, Xiongfei, Wang, Xujian, Wang, Yaping, Wang, Yaqian, Wang, Yi, Wang, Yiao, Wang, Yifang, Wang, Yilun, Wang, Yiwei, Wang, You-Kai, Wang, Yuanping, Wang, Yuexin, Wang, Yuhao, Wang, Yu-Ming, Wang, Yuting, Wang, Zhen, Wang, Zhigang, Wang, Weiping, Wang, Zeren Simon, Wang, Biao, Wang, Hui, Wang, Lian-Tao, Wang, Zihui, Wang, Zirui, Wang, Jia, Wang, Tong, Wei, Daihui, Wei, Shujun, Wei, Wei, Wei, Xiaomin, Wei, Yuanyuan, Wei, Yingjie, Wen, Liangjian, Wen, Xuejun, Wen, Yufeng, White, Martin, Williams, Peter, Wolffs, Zef, Womersley, William John, Wu, Baona, Wu, Bobing, Wu, Guanjian, Wu, Jinfei, Wu, Lei, Wu, Lina, Wu, Linghui, Wu, Minlin, Wu, Peiwen, Wu, Qi, Wu, Qun, Wu, Tianya, Wu, Xiang, Wu, Xiaohong, Wu, Xing-Gang, Wu, Xuehui, Wu, Yaru, Wu, Yongcheng, Wu, Yuwen, Wu, Zhi, Wu, Xin, Xia, Lei, Xia, Ligang, Xia, Shang, Xiang, Benhou, Xiang, Dao, Xiang, Zhiyu, Xiao, Bo-Wen, Xiao, Chu-Wen, Xiao, Dong, Xiao, Guangyan, Xiao, Han, Xiao, Meng, Xiao, Ouzheng, Xiao, Rui-Qing, Xiao, Xiang, Xiao, Yichen, Xiao, Ying, Xiao, Yu, Xiao, Yunlong, Xiao, Zhenjun, Xiao, Hengyuan, Xie, Nian, Xie, Yuehong, Xin, Tianmu, Xing, Ye, Xing, Zhizhong, Xu, Da, Xu, Fang, Xu, Fanrong, Xu, Haisheng, Xu, Haocheng, Xu, Ji, Xu, Miaofu, Xu, Qingjin, Xu, Qingnian, Xu, Wei, Xu, Wei, Xu, Weixi, Xu, Xinping, Xu, Zhen, Xu, Zijun, Xu, Zehua, Xu, Yaoyuan, Xue, Feifei, Yan, Baojun, Yan, Bin, Yan, Fen, Yan, Fucheng, Yan, Jiaming, Yan, Liang, Yan, Luping, Yan, Qi-Shu, Yan, Wenbiao, Yan, Yupeng, Yan, Luping, Yan, Haoyue, Yang, Dong, Yang, Fengying, Yang, Guicheng, Yang, Haijun, Yang, Jin Min, Yang, Jing, Yang, Lan, Yang, Li, Yang, Li Lin, Yang, Lili, Yang, Litao, Yang, Mei, Yang, Qiaoli, Yang, Tiansen, Yang, Xiaochen, Yang, Yingjun, Yang, Yueling, Yang, Zhengyong, Yang, Zhenwei, Yang, Youhua, Yang, Xiancong, Yao, De-Liang, Yao, Shi, Ye, Lei, Ye, Lingxi, Ye, Mei, Ye, Rui, Ye, Rui, Ye, Yecheng, Yermolchyk, Vitaly, Yi, Kai, Yi, Li, Yi, Yang, Yin, Di, Yin, Peng-Fei, Yin, Shenghua, Yin, Ze, Yin, Zhongbao, Yinhong, Zhang, Yoo, Hwi Dong, You, Zhengyun, Young, Charles, Yu, Boxiang, Yu, Chenghui, Yu, Fusheng, Yu, Jie-Sheng, Yu, Jinqing, Yu, Lingda, Yu, Zhao-Huan, Yu, Felix, Yu, Bingrong, Yuan, Changzheng, Yuan, Li, Yuan, Xing-Bo, Yuan, Youjin, Yue, Junhui, Yue, Qian, Yue, Baobiao, Zaib, Un Nisa, Zanzottera, Riccardo, Zeng, Hao, Zeng, Ming, Zhai, Jian, Zhai, Jiyuan, Zhai, Xin Zhe, Zhan, Xi-Jie, Zhang, Ben-Wei, Zhang, Bolun, Zhang, Di, Zhang, Guangyi, Zhang, Hao, Zhang, Hong-Hao, Zhang, Huaqiao, Zhang, Hui, Zhang, Jialiang, Zhang, Jianyu, Zhang, Jianzhong, Zhang, Jiehao, Zhang, Jielei, Zhang, Jingru, Zhang, Jinxian, Zhang, Junsong, Zhang, Junxing, Zhang, Lei, Zhang, Lei, Zhang, Liang, Zhang, Licheng, Zhang, Liming, Zhang, Linhao, Zhang, Luyan, Zhang, Mengchao, Zhang, Rao, Zhang, Shulei, Zhang, Wan, Zhang, Wenchao, Zhang, Xiangzhen, Zhang, Xiaomei, Zhang, Xiaoming, Zhang, Xiaoxu, Zhang, Xiaoyu, Zhang, Xuantong, Zhang, Xueyao, Zhang, Yang, Zhang, Yang, Zhang, Yanxi, Zhang, Yao, Zhang, Ying, Zhang, Yixiang, Zhang, Yizhou, Zhang, Yongchao, Zhang, Yu, Zhang, Yuan, Zhang, Yujie, Zhang, Yulei, Zhang, Yumei, Zhang, Yunlong, Zhang, Zhandong, Zhang, Zhaoru, Zhang, Zhen-Hua, Zhang, Zhenyu, Zhang, Zhichao, Zhang, Zhi-Qing, Zhang, Zhuo, Zhang, Zhiqing, Zhang, Cong, Zhang, Tianliang, Zhang, Luyan, Zhao, Guang, Zhao, Hongyun, Zhao, Jie, Zhao, Jingxia, Zhao, Jingyi, Zhao, Ling, Zhao, Luyang, Zhao, Mei, Zhao, Minggang, Zhao, Mingrui, Zhao, Qiang, Zhao, Ruiguang, Zhao, Tongxian, Zhao, Yaliang, Zhao, Ying, Zhao, Yue, Zhao, Zhiyu, Zhao, Zhuo, Zhemchugov, Alexey, Zheng, Hongjuan, Zheng, Jinchao, Zheng, Liang, Zheng, Ran, zheng, shanxi, Zheng, Xu-Chang, Zhile, Wang, Zhong, Weicai, Zhong, Yi-Ming, Zhou, Chen, Zhou, Daicui, Zhou, Jianxin, Zhou, Jing, Zhou, Jing, Zhou, Ning, Zhou, Qi-Dong, Zhou, Shiyu, Zhou, Shun, Zhou, Sihong, Zhou, Xiang, Zhou, Xingyu, Zhou, Yang, Zhou, Yong, Zhou, Yu-Feng, Zhou, Zusheng, Zhou, Demin, Zhu, Dechong, Zhu, Hongbo, Zhu, Huaxing, Zhu, Jingya, Zhu, Kai, Zhu, Pengxuan, Zhu, Ruilin, Zhu, Xianglei, Zhu, Yingshun, Zhu, Yongfeng, Zhuang, Xiao, Zhuang, Xuai, Zobov, Mikhail, Zong, Zhanguo, Zou, Cong, and Zou, Hongying
- Published
- 2024
- Full Text
- View/download PDF
47. How to balance the live birth rate and the multiple pregnancy rate by selecting the cleavage-stage embryo number and quality for POSEIDON Group 1 and Group 2? A retrospective study
- Author
-
He, Huiqing, Liu, Rang, Zhang, Qiuju, Geng, Lan, Hou, Zhenhui, Xu, Chang, Cao, Yanpei, and Xia, Xi
- Published
- 2024
- Full Text
- View/download PDF
48. Enhancement of ε-poly-L-lysine production by Streptomyces albulus FQF-24 with feeding strategies using cassava starch as carbon source
- Author
-
Li, Boyan, Wu, Chenqi, Bai, Senmeng, Zhang, Di, Xu, Chang, Yuan, Xiaofeng, Tian, Jiayi, Bai, Jing, Li, Liangzhi, and Fu, Jiaolong
- Published
- 2024
- Full Text
- View/download PDF
49. Exploring the effect of directive and reflective feedback on elementary school students’ scientific conceptual understanding, epistemological beliefs, and inquiry performance in online inquiry activities
- Author
-
Zheng, Yafeng, Sun, Shebing, Yang, Yang, and Xu, Chang
- Published
- 2024
- Full Text
- View/download PDF
50. Relationship Between COVID-19 Infection and Risk Perception, Knowledge, Attitude, and Four Nonpharmaceutical Interventions During the Late Period of the COVID-19 Epidemic in China: Online Cross-Sectional Survey of 8158 Adults
- Author
-
Xu, Hong, Gan, Yong, Zheng, Daikun, Wu, Bo, Zhu, Xian, Xu, Chang, Liu, Chenglu, Tao, Zhou, Hu, Yaoyue, Chen, Min, Li, Mingjing, Lu, Zuxun, and Chen, Jack
- Subjects
Computer applications to medicine. Medical informatics ,R858-859.7 ,Public aspects of medicine ,RA1-1270 - Abstract
BackgroundSo far, there have been no published population studies on the relationship between a COVID-19 infection and public risk perception, information source, knowledge, attitude, and behaviors during the COVID-19 outbreak in China. ObjectiveThis study aims to understand the relationships between COVID-19 infection; four personal nonpharmaceutical interventions (NPIs; handwashing, proper coughing habits, social distancing, and mask wearing); and public risk perception, knowledge, attitude, and other social demographic variables. MethodsAn online survey of 8158 Chinese adults between February 22 and March 5, 2020, was conducted. Bivariate associations between categorical variables were examined using Fisher exact test. We also explored the determinants of four NPIs as well as their association with COVID-19 infection using logistic regression. ResultsOf 8158 adults included, 57 (0.73%) were infected with COVID-19. The overwhelming majority of respondents showed a positive attitude (n=8094, 99.2%), positive risk perception (n=8146, 99.9%), and high knowledge levels that were among the strongest predictors of the four adopted NPIs (handwashing: n=7895, 96.8%; proper coughing: 5997/6444, 93.1%; social distancing: n=7104/8158, 87.1%; and mask wearing: 5011/5120, 97.9%). There was an increased risk of COVID-19 infection for those who did not wash their hands (2.28% vs 0.65%; risk ratio [RR] 3.53, 95% CI 1.53-8.15; P=.009), did not practice proper coughing (1.79% vs 0.73%; RR 2.44, 95% CI 1.15-5.15; P=.03), did not practice social distancing (1.52% vs 0.58%; RR 2.63, 95% CI 1.48-4.67; P=.002), and did not wear a mask (7.41% vs 0.6%; RR 12.38, 95% CI 5.81-26.36; P
- Published
- 2020
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.