5,502 results on '"Peng, Xin"'
Search Results
2. Optimum Seed Harvest Time of Trifolium lupinaster L. in Relation to Flowering, Pods and Seed Characteristics
- Author
-
Peng, Xin, Wu, Feifei, Fu, Nana, Shi, Fengling, and Zhang, Yutong
- Published
- 2024
- Full Text
- View/download PDF
3. Chinatown Film Culture: The Appearance of Cinema in San Francisco's Chinese Neighborhood by Kim K. Fahlstedt (review)
- Author
-
Peng, Xin
- Published
- 2022
- Full Text
- View/download PDF
4. Lifting the Veil on the Large Language Model Supply Chain: Composition, Risks, and Mitigations
- Author
-
Huang, Kaifeng, Chen, Bihuan, Lu, You, Wu, Susheng, Wang, Dingji, Huang, Yiheng, Jiang, Haowen, Zhou, Zhuotong, Cao, Junming, and Peng, Xin
- Subjects
Computer Science - Software Engineering - Abstract
Large language models (LLM) have sparked significant impact with regard to both intelligence and productivity. In recent years, a great surge has been witnessed in the introduction of both commercial and open-source LLMs. Many businesses have adopted the LLMs into their applications to solve their own domain-specific tasks. However, integrating LLMs into specific business scenarios requires more than just utilizing the models themselves. Instead, it is a systematic process that involves substantial components, which are collectively referred to as the LLM supply chain. The LLM supply chain inherently carries risks. Therefore, it is essential to understand the types of components that may be introduced into the supply chain and the associated risks, enabling different stakeholders to implement effective mitigation measures. While some literature discusses risks associated with LLMs, there is currently no paper that clearly outlines the LLM supply chain from the perspective of both providing and consuming its components. As LLMs have become essential infrastructure in the new era, we believe that a thorough review of the LLM supply chain, along with its inherent risks and mitigation strategies, would be valuable for industry practitioners to avoid potential damages and losses, and enlightening for academic researchers to rethink existing approaches and explore new avenues of research. Our paper provides a comprehensive overview of the LLM supply chain, detailing the stakeholders, composing artifacts, and the supplying types. We developed taxonomies of risk types, risky actions, and mitigations related to various supply chain stakeholders and components. In summary, our work explores the technical and operational aspects of the LLM supply chain, offering valuable insights for researchers and engineers in the evolving LLM landscape., Comment: 17 pages
- Published
- 2024
5. Deep Learning-based Software Engineering: Progress, Challenges, and Opportunities
- Author
-
Chen, Xiangping, Hu, Xing, Huang, Yuan, Jiang, He, Ji, Weixing, Jiang, Yanjie, Jiang, Yanyan, Liu, Bo, Liu, Hui, Li, Xiaochen, Lian, Xiaoli, Meng, Guozhu, Peng, Xin, Sun, Hailong, Shi, Lin, Wang, Bo, Wang, Chong, Wang, Jiayi, Wang, Tiantian, Xuan, Jifeng, Xia, Xin, Yang, Yibiao, Yang, Yixin, Zhang, Li, Zhou, Yuming, and Zhang, Lu
- Subjects
Computer Science - Software Engineering - Abstract
Researchers have recently achieved significant advances in deep learning techniques, which in turn has substantially advanced other research disciplines, such as natural language processing, image processing, speech recognition, and software engineering. Various deep learning techniques have been successfully employed to facilitate software engineering tasks, including code generation, software refactoring, and fault localization. Many papers have also been presented in top conferences and journals, demonstrating the applications of deep learning techniques in resolving various software engineering tasks. However, although several surveys have provided overall pictures of the application of deep learning techniques in software engineering, they focus more on learning techniques, that is, what kind of deep learning techniques are employed and how deep models are trained or fine-tuned for software engineering tasks. We still lack surveys explaining the advances of subareas in software engineering driven by deep learning techniques, as well as challenges and opportunities in each subarea. To this end, in this paper, we present the first task-oriented survey on deep learning-based software engineering. It covers twelve major software engineering subareas significantly impacted by deep learning techniques. Such subareas spread out the through the whole lifecycle of software development and maintenance, including requirements engineering, software development, testing, maintenance, and developer collaboration. As we believe that deep learning may provide an opportunity to revolutionize the whole discipline of software engineering, providing one survey covering as many subareas as possible in software engineering can help future research push forward the frontier of deep learning-based software engineering more systematically., Comment: Accepted in SCIENCE CHINA Information Sciences
- Published
- 2024
- Full Text
- View/download PDF
6. TRANSAGENT: An LLM-Based Multi-Agent System for Code Translation
- Author
-
Yuan, Zhiqiang, Chen, Weitong, Wang, Hanlin, Yu, Kai, Peng, Xin, and Lou, Yiling
- Subjects
Computer Science - Software Engineering ,Computer Science - Artificial Intelligence - Abstract
Code translation converts code from one programming language to another while maintaining its original functionality, which is crucial for software migration, system refactoring, and cross-platform development. Traditional rule-based methods rely on manually-written rules, which can be time-consuming and often result in less readable code. To overcome this, learning-based methods have been developed, leveraging parallel data to train models for automated code translation. More recently, the advance of Large Language Models (LLMs) further boosts learning-based code translation. Although promising, LLM-translated program still suffers from diverse quality issues (e.g., syntax errors and semantic errors). In particular, it can be challenging for LLMs to self-debug these errors when simply provided with the corresponding error messages. In this work, we propose a novel LLM-based multi-agent system TRANSAGENT, which enhances LLM-based code translation by fixing the syntax errors and semantic errors with the synergy between four LLM-based agents, including Initial Code Translator, Syntax Error Fixer, Code Aligner, and Semantic Error Fixer. The main insight of TRANSAGENT is to first localize the error code block in the target program based on the execution alignment between the target and source program, which can narrow down the fixing space and thus lower down the fixing difficulties. To evaluate TRANSAGENT, we first construct a new benchmark from recent programming tasks to mitigate the potential data leakage issue. On our benchmark, TRANSAGENT outperforms the latest LLM-based code translation technique UniTrans in both translation effectiveness and efficiency; additionally, our evaluation on different LLMs show the generalization of TRANSAGENT and our ablation study shows the contribution of each agent.
- Published
- 2024
7. Properties of the QCD Matter: A Review of Selected Results from the ALICE Experiment
- Author
-
Shou, Qi-Ye, Ma, Yu-Gang, Zhang, Song, Zhu, Jian-Hui, Mao, Ya-Xian, Pei, Hua, Yin, Zhong-Bao, Zhang, Xiao-Ming, Zhou, Dai-Cui, Peng, Xin-Ye, Bai, Xiao-Zhi, Tang, Ze-Bo, Zhang, Yi-Fei, and Li, Xiao-Mei
- Subjects
Nuclear Experiment ,High Energy Physics - Experiment - Abstract
The Large Hadron Collider (LHC), the world's largest and most powerful particle accelerator, has been a pivotal tool in advancing our understanding of fundamental physics. By colliding heavy ions (such as lead ions), the LHC recreates conditions similar to those just after the Big Bang. This allows scientists to study the Quark-Gluon Plasma (QGP), a state of matter where quarks and gluons are not confined within protons and neutrons. These studies provide insights into the strong force and the early universe's behavior. In this paper, we provide a comprehensive overview of recent significant findings from A Large Ion Collider Experiment (ALICE) at LHC. The topics encompass measurements regarding to properties of the QGP, particle production, flow and correlations, dileptons, quarkonia and electromagnetic probes, heavy flavor, and jets. Additionally, we introduce future plans for detector upgrades of the ALICE experiment., Comment: 29 pages, 32 figures. This review is dedicated to Professor Wenqing Shen in honor of his leadership and significant impact on the Chinese heavy-ion physics community. All authors contributed equally to this work
- Published
- 2024
8. SpikeGS: Learning 3D Gaussian Fields from Continuous Spike Stream
- Author
-
Yu, Jinze, Peng, Xin, Lu, Zhengda, Kneip, Laurent, and Wang, Yiqun
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
A spike camera is a specialized high-speed visual sensor that offers advantages such as high temporal resolution and high dynamic range compared to conventional frame cameras. These features provide the camera with significant advantages in many computer vision tasks. However, the tasks of novel view synthesis based on spike cameras remain underdeveloped. Although there are existing methods for learning neural radiance fields from spike stream, they either lack robustness in extremely noisy, low-quality lighting conditions or suffer from high computational complexity due to the deep fully connected neural networks and ray marching rendering strategies used in neural radiance fields, making it difficult to recover fine texture details. In contrast, the latest advancements in 3DGS have achieved high-quality real-time rendering by optimizing the point cloud representation into Gaussian ellipsoids. Building on this, we introduce SpikeGS, the method to learn 3D Gaussian fields solely from spike stream. We designed a differentiable spike stream rendering framework based on 3DGS, incorporating noise embedding and spiking neurons. By leveraging the multi-view consistency of 3DGS and the tile-based multi-threaded parallel rendering mechanism, we achieved high-quality real-time rendering results. Additionally, we introduced a spike rendering loss function that generalizes under varying illumination conditions. Our method can reconstruct view synthesis results with fine texture details from a continuous spike stream captured by a moving spike camera, while demonstrating high robustness in extremely noisy low-light scenarios. Experimental results on both real and synthetic datasets demonstrate that our method surpasses existing approaches in terms of rendering quality and speed. Our code will be available at https://github.com/520jz/SpikeGS., Comment: Accepted by ACCV 2024
- Published
- 2024
9. Large Language Model-Based Agents for Software Engineering: A Survey
- Author
-
Liu, Junwei, Wang, Kaixin, Chen, Yixuan, Peng, Xin, Chen, Zhenpeng, Zhang, Lingming, and Lou, Yiling
- Subjects
Computer Science - Software Engineering ,Computer Science - Artificial Intelligence - Abstract
The recent advance in Large Language Models (LLMs) has shaped a new paradigm of AI agents, i.e., LLM-based agents. Compared to standalone LLMs, LLM-based agents substantially extend the versatility and expertise of LLMs by enhancing LLMs with the capabilities of perceiving and utilizing external resources and tools. To date, LLM-based agents have been applied and shown remarkable effectiveness in Software Engineering (SE). The synergy between multiple agents and human interaction brings further promise in tackling complex real-world SE problems. In this work, we present a comprehensive and systematic survey on LLM-based agents for SE. We collect 106 papers and categorize them from two perspectives, i.e., the SE and agent perspectives. In addition, we discuss open challenges and future directions in this critical domain. The repository of this survey is at https://github.com/FudanSELab/Agent4SE-Paper-List.
- Published
- 2024
10. Flow Perturbation to Accelerate Unbiased Sampling of Boltzmann distribution
- Author
-
Peng, Xin and Gao, Ang
- Subjects
Statistics - Machine Learning ,Computer Science - Machine Learning ,Physics - Chemical Physics - Abstract
Flow-based generative models have been employed for sampling the Boltzmann distribution, but their application to high-dimensional systems is hindered by the significant computational cost of obtaining the Jacobian of the flow. To overcome this challenge, we introduce the flow perturbation method, which incorporates optimized stochastic perturbations into the flow. By reweighting trajectories generated by the perturbed flow, our method achieves unbiased sampling of the Boltzmann distribution with orders of magnitude speedup compared to both brute force Jacobian calculations and the Hutchinson estimator. Notably, it accurately sampled the Chignolin protein with all atomic Cartesian coordinates explicitly represented, which, to our best knowledge, is the largest molecule ever Boltzmann sampled in such detail using generative models.
- Published
- 2024
11. TIGER: A Generating-Then-Ranking Framework for Practical Python Type Inference
- Author
-
Wang, Chong, Zhang, Jian, Lou, Yiling, Liu, Mingwei, Sun, Weisong, Liu, Yang, and Peng, Xin
- Subjects
Computer Science - Software Engineering - Abstract
Python's dynamic typing system offers flexibility and expressiveness but can lead to type-related errors, prompting the need for automated type inference to enhance type hinting. While existing learning-based approaches show promising inference accuracy, they struggle with practical challenges in comprehensively handling various types, including complex generic types and (unseen) user-defined types. In this paper, we introduce TIGER, a two-stage generating-then-ranking (GTR) framework, designed to effectively handle Python's diverse type categories. TIGER leverages fine-tuned pre-trained code models to train a generative model with a span masking objective and a similarity model with a contrastive training objective. This approach allows TIGER to generate a wide range of type candidates, including complex generics in the generating stage, and accurately rank them with user-defined types in the ranking stage. Our evaluation on the ManyTypes4Py dataset shows TIGER's advantage over existing methods in various type categories, notably improving accuracy in inferring user-defined and unseen types by 11.2% and 20.1% respectively in Top-5 Exact Match. Moreover, the experimental results not only demonstrate TIGER's superior performance and efficiency, but also underscore the significance of its generating and ranking stages in enhancing automated type inference., Comment: Accepted by ICSE'25
- Published
- 2024
12. Vul-RAG: Enhancing LLM-based Vulnerability Detection via Knowledge-level RAG
- Author
-
Du, Xueying, Zheng, Geng, Wang, Kaixin, Feng, Jiayi, Deng, Wentai, Liu, Mingwei, Chen, Bihuan, Peng, Xin, Ma, Tao, and Lou, Yiling
- Subjects
Computer Science - Software Engineering ,Computer Science - Artificial Intelligence - Abstract
Vulnerability detection is essential for software quality assurance. In recent years, deep learning models (especially large language models) have shown promise in vulnerability detection. In this work, we propose a novel LLM-based vulnerability detection technique Vul-RAG, which leverages knowledge-level retrieval-augmented generation (RAG) framework to detect vulnerability for the given code in three phases. First, Vul-RAG constructs a vulnerability knowledge base by extracting multi-dimension knowledge via LLMs from existing CVE instances; second, for a given code snippet, Vul-RAG} retrieves the relevant vulnerability knowledge from the constructed knowledge base based on functional semantics; third, Vul-RAG leverages LLMs to check the vulnerability of the given code snippet by reasoning the presence of vulnerability causes and fixing solutions of the retrieved vulnerability knowledge. Our evaluation of Vul-RAG on our constructed benchmark PairVul shows that Vul-RAG substantially outperforms all baselines by 12.96\%/110\% relative improvement in accuracy/pairwise-accuracy. In addition, our user study shows that the vulnerability knowledge generated by Vul-RAG can serve as high-quality explanations which can improve the manual detection accuracy from 0.60 to 0.77.
- Published
- 2024
13. STALL+: Boosting LLM-based Repository-level Code Completion with Static Analysis
- Author
-
Liu, Junwei, Chen, Yixuan, Liu, Mingwei, Peng, Xin, and Lou, Yiling
- Subjects
Computer Science - Software Engineering - Abstract
Repository-level code completion is challenging as it involves complicated contexts from multiple files in the repository. To date, researchers have proposed two technical categories to enhance LLM-based repository-level code completion, i.e., retrieval-augmented generation (RAG) and static analysis integration. This work performs the first study on the static analysis integration in LLM-based repository-level code completion by investigating both the effectiveness and efficiency of static analysis integration strategies across different phases of code completion. We first implement a framework STALL+, which supports an extendable and customizable integration of multiple static analysis strategies into the complete pipeline of LLM-based repository-level code completion; and based on STALL+, we perform extensive experiments by including different code LLMs on the latest repository-level code completion benchmark CrossCodeEval. Our findings show that integrating file-level dependencies in prompting phase performs the best while the integration in post-processing phase performs the worse. Additionally, we observe different improvements from static analysis between dynamic languages and static languages, i.e., the best combination is prompting-phase with decoding-phase integration for Java while the best combination is prompting-phase with post-processing-phase integration for Python given the limitations of statically analyzing dynamic languages. Additionally, we find the complementarity between RAG and static analysis integration as well as their cost-effectiveness after combination., Comment: 12 pages, 5 figures
- Published
- 2024
14. How and Why LLMs Use Deprecated APIs in Code Completion? An Empirical Study
- Author
-
Wang, Chong, Huang, Kaifeng, Zhang, Jian, Feng, Yebo, Zhang, Lyuye, Liu, Yang, and Peng, Xin
- Subjects
Computer Science - Software Engineering - Abstract
Large language models (LLMs), pre-trained or fine-tuned on large code corpora, have shown effectiveness in generating code completions. However, in LLM-based code completion, LLMs may struggle to use correct and up-to-date Application Programming Interfaces (APIs) due to the rapid and continuous evolution of libraries. While existing studies have highlighted issues with predicting incorrect APIs, the specific problem of deprecated API usage in LLM-based code completion has not been thoroughly investigated. To address this gap, we conducted the first evaluation study on deprecated API usage in LLM-based code completion. This study involved seven advanced LLMs, 145 API mappings from eight popular Python libraries, and 28,125 completion prompts. The study results reveal the \textit{status quo} and \textit{root causes} of deprecated API usage in LLM-based code completion from the perspectives of \textit{model}, \textit{prompt}, and \textit{library}. Based on these findings, we propose two lightweight fixing approaches, \textsc{ReplaceAPI} and \textsc{InsertPrompt}, which can serve as baseline approaches for future research on mitigating deprecated API usage in LLM-based completion. Additionally, we provide implications for future research on integrating library evolution with LLM-driven software development.
- Published
- 2024
15. Unlocking Parameter-Efficient Fine-Tuning for Low-Resource Language Translation
- Author
-
Su, Tong, Peng, Xin, Thillainathan, Sarubi, Guzmán, David, Ranathunga, Surangika, and Lee, En-Shiun Annie
- Subjects
Computer Science - Computation and Language - Abstract
Parameter-efficient fine-tuning (PEFT) methods are increasingly vital in adapting large-scale pre-trained language models for diverse tasks, offering a balance between adaptability and computational efficiency. They are important in Low-Resource Language (LRL) Neural Machine Translation (NMT) to enhance translation accuracy with minimal resources. However, their practical effectiveness varies significantly across different languages. We conducted comprehensive empirical experiments with varying LRL domains and sizes to evaluate the performance of 8 PEFT methods with in total of 15 architectures using the SacreBLEU score. We showed that 6 PEFT architectures outperform the baseline for both in-domain and out-domain tests and the Houlsby+Inversion adapter has the best performance overall, proving the effectiveness of PEFT methods., Comment: Accepted to the Findings of NAACL 2024
- Published
- 2024
16. A general approach to enhance the survivability of backdoor attacks by decision path coupling
- Author
-
Zhao, Yufei, Wang, Dingji, Chen, Bihuan, Chen, Ziqian, and Peng, Xin
- Subjects
Computer Science - Artificial Intelligence ,Computer Science - Cryptography and Security - Abstract
Backdoor attacks have been one of the emerging security threats to deep neural networks (DNNs), leading to serious consequences. One of the mainstream backdoor defenses is model reconstruction-based. Such defenses adopt model unlearning or pruning to eliminate backdoors. However, little attention has been paid to survive from such defenses. To bridge the gap, we propose Venom, the first generic backdoor attack enhancer to improve the survivability of existing backdoor attacks against model reconstruction-based defenses. We formalize Venom as a binary-task optimization problem. The first is the original backdoor attack task to preserve the original attack capability, while the second is the attack enhancement task to improve the attack survivability. To realize the second task, we propose attention imitation loss to force the decision path of poisoned samples in backdoored models to couple with the crucial decision path of benign samples, which makes backdoors difficult to eliminate. Our extensive evaluation on two DNNs and three datasets has demonstrated that Venom significantly improves the survivability of eight state-of-the-art attacks against eight state-of-the-art defenses without impacting the capability of the original attacks.
- Published
- 2024
17. Dual-comb spectroscopy over a 100 km open-air path
- Author
-
Han, Jin-Jian, Zhong, Wei, Zhao, Ruo-Can, Zeng, Ting, Li, Min, Lu, Jian, Peng, Xin-Xin, Shi, Xi-Ping, Yin, Qin, Wang, Yong, Esamdin, Ali, Shen, Qi, Guan, Jian-Yu, Hou, Lei, Ren, Ji-Gang, Jia, Jian-Jun, Wang, Yu, Jiang, Hai-Feng, Xue, Xiang-Hui, Zhang, Qiang, Dou, Xian-Kang, and Pan, Jian-Wei
- Published
- 2024
- Full Text
- View/download PDF
18. Numerical approximation of the solution of Koiter’s model for an elliptic membrane shell in absence of friction subjected to an obstacle via the penalty method
- Author
-
Peng, Xin, Piersanti, Paolo, and Shen, Xiaoqin
- Published
- 2024
- Full Text
- View/download PDF
19. Structural basis for linker histone H5–nucleosome binding and chromatin fiber compaction
- Author
-
Li, Wenyan, Hu, Jie, Song, Feng, Yu, Juan, Peng, Xin, Zhang, Shuming, Wang, Lin, Hu, Mingli, Liu, Jia-Cheng, Wei, Yu, Xiao, Xue, Li, Yan, Li, Dongyu, Wang, Hui, Zhou, Bing-Rui, Dai, Linchang, Mou, Zongjun, Zhou, Min, Zhang, Haonan, Zhou, Zheng, Zhang, Huidong, Bai, Yawen, Zhou, Jin-Qiu, Li, Wei, Li, Guohong, and Zhu, Ping
- Published
- 2024
- Full Text
- View/download PDF
20. Functional Liver Imaging Score Derived from Gadoxetic Acid-enhanced MRI Predicts Cachexia and Prognosis in Hepatocellular Carcinoma Patients
- Author
-
Li, Xin-xiang, Liu, Bing, Zhao, Yu-fei, Jiang, Yang, Cui, Ying, and Peng, Xin-gui
- Published
- 2024
- Full Text
- View/download PDF
21. Isolation, identification, and evaluation of intestinal bacteria in Macrobrachium rosenbergii
- Author
-
Zhao, Xiuxin, Luo, Jinping, Liu, Peimin, Huang, Hao, Cheng, Zhenheng, Peng, Xin, Tang, Qiongying, Yang, Guoliang, Yi, Shaokui, and Gao, Quanxin
- Published
- 2024
- Full Text
- View/download PDF
22. Intravoxel incoherent motion diffusion-weighted imaging and dynamic contrast-enhanced MRI for predicting parametrial invasion in cervical cancer
- Author
-
Li, Xin-xiang, Liu, Bing, Cui, Ying, Zhao, Yu-fei, Jiang, Yang, and Peng, Xin-gui
- Published
- 2024
- Full Text
- View/download PDF
23. Teaching Code LLMs to Use Autocompletion Tools in Repository-Level Code Generation
- Author
-
Wang, Chong, Zhang, Jian, Feng, Yebo, Li, Tianlin, Sun, Weisong, Liu, Yang, and Peng, Xin
- Subjects
Computer Science - Software Engineering - Abstract
Code large language models (LLMs) face limitations in repository-level code generation due to their lack of awareness of repository-level dependencies (e.g., user-defined attributes), resulting in dependency errors such as undefined-variable and no-member errors. In this work, we introduce ToolGen, an approach that integrates autocompletion tools into the code LLM generation process to address these dependencies. ToolGen comprises two main phases: Trigger Insertion and Model Fine-tuning (Offline), and Tool-integrated Code Generation (Online). During the offline phase, ToolGen augments functions within a given code corpus with a special mark token, indicating positions to trigger autocompletion tools. These augmented functions, along with their corresponding docstrings, are then used to fine-tune a selected code LLM. In the online phase, ToolGen iteratively generates functions by predicting tokens step-by-step using the fine-tuned LLM. Whenever a mark token is encountered, ToolGen invokes the autocompletion tool to suggest code completions and selects the most appropriate one. We conduct comprehensive experiments to evaluate ToolGen's effectiveness in repository-level code generation. To facilitate this evaluation, we create a benchmark comprising 671 real-world code repositories and introduce two new dependency-based metrics: Dependency Coverage and Static Validity Rate. The results demonstrate that ToolGen significantly improves Dependency Coverage by 31.4% to 39.1% and Static Validity Rate by 44.9% to 57.7% across the three LLMs, while maintaining competitive or improved performance in widely recognized similarity metrics such as BLEU-4, CodeBLEU, Edit Similarity, and Exact Match. On the CoderEval dataset, ToolGen achieves improvements of 40.0% and 25.0% in Pass@1 for CodeT5 and CodeLlama, respectively.
- Published
- 2024
24. YAYI 2: Multilingual Open-Source Large Language Models
- Author
-
Luo, Yin, Kong, Qingchao, Xu, Nan, Cao, Jia, Hao, Bao, Qu, Baoyu, Chen, Bo, Zhu, Chao, Zhao, Chenyang, Zhang, Donglei, Feng, Fan, Zhao, Feifei, Sun, Hailong, Yang, Hanxuan, Pan, Haojun, Liu, Hongyu, Guo, Jianbin, Du, Jiangtao, Wang, Jingyi, Li, Junfeng, Sun, Lei, Liu, Liduo, Dong, Lifeng, Liu, Lili, Wang, Lin, Zhang, Liwen, Wang, Minzheng, Wang, Pin, Yu, Ping, Li, Qingxiao, Yan, Rui, Zou, Rui, Li, Ruiqun, Huang, Taiwen, Wang, Xiaodong, Wu, Xiaofei, Peng, Xin, Zhang, Xina, Fang, Xing, Xiao, Xinglin, Hao, Yanni, Dong, Yao, Wang, Yigang, Liu, Ying, Jiang, Yongyu, Wang, Yungan, Wang, Yuqi, Wang, Zhangsheng, Yu, Zhaoxin, Luo, Zhen, Mao, Wenji, Wang, Lei, and Zeng, Dajun
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence - Abstract
As the latest advancements in natural language processing, large language models (LLMs) have achieved human-level language understanding and generation abilities in many real-world tasks, and even have been regarded as a potential path to the artificial general intelligence. To better facilitate research on LLMs, many open-source LLMs, such as Llama 2 and Falcon, have recently been proposed and gained comparable performances to proprietary models. However, these models are primarily designed for English scenarios and exhibit poor performances in Chinese contexts. In this technical report, we propose YAYI 2, including both base and chat models, with 30 billion parameters. YAYI 2 is pre-trained from scratch on a multilingual corpus which contains 2.65 trillion tokens filtered by our pre-training data processing pipeline. The base model is aligned with human values through supervised fine-tuning with millions of instructions and reinforcement learning from human feedback. Extensive experiments on multiple benchmarks, such as MMLU and CMMLU, consistently demonstrate that the proposed YAYI 2 outperforms other similar sized open-source models.
- Published
- 2023
25. Resolving Crash Bugs via Large Language Models: An Empirical Study
- Author
-
Du, Xueying, Liu, Mingwei, Li, Juntao, Wang, Hanlin, Peng, Xin, and Lou, Yiling
- Subjects
Computer Science - Software Engineering ,Computer Science - Artificial Intelligence ,Computer Science - Computation and Language - Abstract
Crash bugs cause unexpected program behaviors or even termination, requiring high-priority resolution. However, manually resolving crash bugs is challenging and labor-intensive, and researchers have proposed various techniques for their automated localization and repair. ChatGPT, a recent large language model (LLM), has garnered significant attention due to its exceptional performance across various domains. This work performs the first investigation into ChatGPT's capability in resolve real-world crash bugs, focusing on its effectiveness in both localizing and repairing code-related and environment-related crash bugs. Specifically, we initially assess ChatGPT's fundamental ability to resolve crash bugs with basic prompts in a single iteration. We observe that ChatGPT performs better at resolving code-related crash bugs compared to environment-related ones, and its primary challenge in resolution lies in inaccurate localization. Additionally, we explore ChatGPT's potential with various advanced prompts. Furthermore, by stimulating ChatGPT's self-planning, it methodically investigates each potential crash-causing environmental factor through proactive inquiry, ultimately identifying the root cause of the crash. Based on our findings, we propose IntDiagSolver, an interaction methodology designed to facilitate precise crash bug resolution through continuous interaction with LLMs. Evaluating IntDiagSolver on multiple LLMs reveals consistent enhancement in the accuracy of crash bug resolution, including ChatGPT, Claude, and CodeLlama.
- Published
- 2023
26. Enhancing Robot Program Synthesis Through Environmental Context
- Author
-
Chen, Tianyi, Wang, Qidi, Dong, Zhen, Shen, Liwei, and Peng, Xin
- Subjects
Computer Science - Robotics - Abstract
Program synthesis aims to automatically generate an executable program that conforms to the given specification. Recent advancements have demonstrated that deep neural methodologies and large-scale pretrained language models are highly proficient in capturing program semantics. For robot programming, prior works have facilitated program synthesis by incorporating global environments. However, the assumption of acquiring a comprehensive understanding of the entire environment is often excessively challenging to achieve. In this work, we present a framework that learns to synthesize a program by rectifying potentially erroneous code segments, with the aid of partially observed environments. To tackle the issue of inadequate attention to partial observations, we propose to first learn an environment embedding space that can implicitly evaluate the impacts of each program token based on the precondition. Furthermore, by employing a graph structure, the model can aggregate both environmental and syntactic information flow and furnish smooth program rectification guidance. Extensive experimental evaluations and ablation studies on the partially observed VizDoom domain authenticate that our method offers superior generalization capability across various tasks and greater robustness when encountering noises.
- Published
- 2023
27. EnvGuard: Guaranteeing Environment-Centric Safety and Security Properties in Web of Things
- Author
-
Sun, Bingkun, Shen, Liwei, Ren, Jialin, Dong, Zhen, Wang, Siao, and Peng, Xin
- Subjects
Computer Science - Software Engineering - Abstract
Web of Things (WoT) technology facilitates the standardized integration of IoT devices ubiquitously deployed in daily environments, promoting diverse WoT applications to automatically sense and regulate the environment. In WoT environment, heterogeneous applications, user activities, and environment changes collectively influence device behaviors, posing risks of unexpected violations of safety and security properties. Existing work on violation identification primarily focuses on the analysis of automated applications, lacking consideration of the intricate interactions in the environment. Moreover, users' intention for violation resolving strategy is much less investigated. To address these limitations, we introduce EnvGuard, an environment-centric approach for property customizing, violation identification and resolution execution in WoT environment. We evaluated EnvGuard in two typical WoT environments. By conducting user studies and analyzing collected real-world environment data, we assess the performance of EnvGuard, and construct a dataset from the collected data to support environment-level violation identification. The results demonstrate the superiority of EnvGuard compared to previous state-of-the-art work, and confirm its usability, feasibility and runtime efficiency.
- Published
- 2023
28. Using alpha, beta, and zeta diversity to map the structure and function of fish community in the central East China Sea
- Author
-
Zhang, Pengzhan, Kong, Yefu, Wang, Linlong, Peng, Xin, and Kang, Bin
- Published
- 2024
- Full Text
- View/download PDF
29. Semantic mask-based two-step approach: a general framework for X-ray diffraction peak search in high-throughput molecular sieve synthetic system
- Author
-
Wei, Zhangpeng, Peng, Xin, Du, Wenli, Qian, Feng, and Yuan, Zhiqing
- Published
- 2024
- Full Text
- View/download PDF
30. Preparation of (S)-epichlorohydrin using a novel halohydrin dehalogenase by selective conformation adjustment
- Author
-
Zhang, Xiao-Jian, Huang, Meng-Yu, Peng, Xin-Xin, Cao, Min, Deng, Han-Zhong, Gong, Yi-Chuan, Tang, Xiao-Ling, Liu, Zhi-Qiang, and Zheng, Yu-Guo
- Published
- 2024
- Full Text
- View/download PDF
31. Rhenium-boosted electrocatalytic activity and durability of pyrolytic IrO2 for acidic oxygen evolution
- Author
-
Lv, Hong-Wei, Zhao, Hong-Bin, Peng, Xin-Yuan, Ye, Zhi-Guo, Huang, Quan-Bo, Yuan, Xue-Tao, Li, Duo-Sheng, and Jin, Zhong
- Published
- 2024
- Full Text
- View/download PDF
32. A facile strategy to construct MOF-based nanocatalyst with enhanced activity and selectivity in oxytetracycline degradation
- Author
-
Ke, Yanjing, Zhang, Jiaxing, Peng, Xin, Zhang, Zhiyi, Wang, Xu, Qi, Wei, and Wang, Mengfan
- Published
- 2024
- Full Text
- View/download PDF
33. The Inhibition Mechanisms of Three Structurally Different Salvianolic Acids on the Non-Enzymatic Glycation of Bovine Serum Albumin
- Author
-
Feng, Guo, Yan, Yu, Wang, Mengfan, Gao, Zhao, Zhao, Yinan, and Peng, Xin
- Published
- 2024
- Full Text
- View/download PDF
34. Inferring Resource-Oriented Intentions using LLMs for Static Resource Leak Detection
- Author
-
Wang, Chong, Liu, Jianan, Peng, Xin, Liu, Yang, and Lou, Yiling
- Subjects
Computer Science - Software Engineering - Abstract
Resource leaks, caused by resources not being released after acquisition, often lead to performance issues and system crashes. Existing static detection techniques rely on mechanical matching of predefined resource acquisition/release APIs and null-checking conditions to find unreleased resources, suffering from both (1) false negatives caused by the incompleteness of predefined resource acquisition/release APIs and (2) false positives caused by the incompleteness of resource reachability validation identification. To overcome these challenges, we propose InferROI, a novel approach that leverages the exceptional code comprehension capability of large language models (LLMs) to directly infer resource-oriented intentions (acquisition, release, and reachability validation) in code. InferROI first prompts the LLM to infer involved intentions for a given code snippet, and then incorporates a two-stage static analysis approach to check control-flow paths for resource leak detection based on the inferred intentions. We evaluate the effectiveness of InferROI in both resource-oriented intention inference and resource leak detection. Experimental results on the DroidLeaks and JLeaks datasets demonstrate InferROI achieves promising bug detection rate (59.3% and 64.8%) and false alarm rate (18.6% and 24.0%). Compared to three industrial static detectors, InferROI detects 14~45 and 167~503 more bugs in DroidLeaks and JLeaks, respectively. When applied to real-world open-source projects, InferROI identifies 26 unknown resource leak bugs, with 7 of them being confirmed by developers. Finally, manual annotation indicated that InferROI achieved a precision of 74.6% and a recall of 81.8% in intention inference, covering more than 60% resource types involved in the datasets. The results of an ablation study underscores the importance of combining LLM-based inference with static analysis.
- Published
- 2023
35. Dual-comb spectroscopy over 100km open-air path
- Author
-
Han, Jin-Jian, Zhong, Wei, Zhao, Ruo-Can, Zeng, Ting, Li, Min, Lu, Jian, Peng, Xin-Xin, Shi, Xi-Ping, Yin, Qin, Wang, Yong, Esamdin, Ali, Shen, Qi, Guan, Jian-Yu, Hou, Lei, Ren, Ji-Gang, Jia, Jian-Jun, Wang, Yu, Jiang, Hai-Feng, Xue, XiangHui, Zhang, Qiang, Dou, Xian-Kang, and Pan, Jian-Wei
- Subjects
Physics - Optics ,Quantum Physics - Abstract
Satellite-based greenhouse gases (GHG) sensing technologies play a critical role in the study of global carbon emissions and climate change. However, none of the existing satellite-based GHG sensing technologies can achieve the measurement of broad bandwidth, high temporal-spatial resolution, and high sensitivity at the same time. Recently, dual-comb spectroscopy (DCS) has been proposed as a superior candidate technology for GHG sensing because it can measure broadband spectra with high temporal-spatial resolution and high sensitivity. The main barrier to DCS's display on satellites is its short measurement distance in open air achieved thus far. Prior research has not been able to implement DCS over 20 km of open-air path. Here, by developing a bistatic setup using time-frequency dissemination and high-power optical frequency combs, we have implemented DCS over a 113 km turbulent horizontal open-air path. Our experiment successfully measured GHG with 7 nm spectral bandwidth and a 10 kHz frequency and achieved a CO2 sensing precision of <2 ppm in 5 minutes and <0.6 ppm in 36 minutes. Our results represent a significant step towards advancing the implementation of DCS as a satellite-based technology and improving technologies for GHG monitoring, Comment: 24 pages, 6 figures
- Published
- 2023
36. Detecting and Fixing Violations of Modification Terms in Open Source Licenses during Forking
- Author
-
Huang, Kaifeng, Xia, Yingfeng, Chen, Bihuan, Zhou, Zhuotong, Guo, Jin, and Peng, Xin
- Subjects
Computer Science - Software Engineering - Abstract
Open source software brings benefit to software community, but also introduces legal risks caused by license violations, which result in serious consequences such as lawsuits and financial losses. To mitigate legal risks, some approaches have been proposed to identify licenses, detect license incompatibilities and inconsistencies, and recommend licenses. As far as we know, however, there is no prior work to understand modification terms in open source licenses or to detect and fix violations of modification terms. To bridge this gap, we first empirically characterize modification terms in 47 open source licenses. These licenses all require certain forms of "notice" to describe the modifications made to the original work. Inspired by our study, we then design LiVo to automatically detect and fix violations of modification terms in open source licenses during forking. Our evaluation has shown the effectiveness and efficiency of LiVo. 18 pull requests of fixing modification term violations have received positive responses. 8 have been merged., Comment: 12 pages
- Published
- 2023
37. Exploring the Potential of ChatGPT in Automated Code Refinement: An Empirical Study
- Author
-
Guo, Qi, Cao, Junming, Xie, Xiaofei, Liu, Shangqing, Li, Xiaohong, Chen, Bihuan, and Peng, Xin
- Subjects
Computer Science - Software Engineering - Abstract
Code review is an essential activity for ensuring the quality and maintainability of software projects. However, it is a time-consuming and often error-prone task that can significantly impact the development process. Recently, ChatGPT, a cutting-edge language model, has demonstrated impressive performance in various natural language processing tasks, suggesting its potential to automate code review processes. However, it is still unclear how well ChatGPT performs in code review tasks. To fill this gap, in this paper, we conduct the first empirical study to understand the capabilities of ChatGPT in code review tasks, specifically focusing on automated code refinement based on given code reviews. To conduct the study, we select the existing benchmark CodeReview and construct a new code review dataset with high quality. We use CodeReviewer, a state-of-the-art code review tool, as a baseline for comparison with ChatGPT. Our results show that ChatGPT outperforms CodeReviewer in code refinement tasks. Specifically, our results show that ChatGPT achieves higher EM and BLEU scores of 22.78 and 76.44 respectively, while the state-of-the-art method achieves only 15.50 and 62.88 on a high-quality code review dataset. We further identify the root causes for ChatGPT's underperformance and propose several strategies to mitigate these challenges. Our study provides insights into the potential of ChatGPT in automating the code review process, and highlights the potential research directions.
- Published
- 2023
38. Reinforcement Learning Based Gasoline Blending Optimization: Achieving More Efficient Nonlinear Online Blending of Fuels
- Author
-
Huang, Muyi, He, Renchu, Dai, Xin, Peng, Xin, Du, Wenli, and Qian, Feng
- Subjects
Computer Science - Computational Engineering, Finance, and Science - Abstract
The online optimization of gasoline blending benefits refinery economies. However, the nonlinear blending mechanism, the oil property fluctuations, and the blending model mismatch bring difficulties to the optimization. To solve the above issues, this paper proposes a novel online optimization method based on deep reinforcement learning algorithm (DRL). The Markov decision process (MDP) expression are given considering a practical gasoline blending system. Then, the environment simulator of gasoline blending process is established based on the MDP expression and the one-year measurement data of a real-world refinery. The soft actor-critic (SAC) DRL algorithm is applied to improve the DRL agent policy by using the data obtained from the interaction between DRL agent and environment simulator. Compared with a traditional method, the proposed method has better economic performance. Meanwhile, it is more robust under property fluctuations and component oil switching. Furthermore, the proposed method maintains performance by automatically adapting to system drift., Comment: 30 pages,13 figures
- Published
- 2023
39. Malicious Package Detection in NPM and PyPI using a Single Model of Malicious Behavior Sequence
- Author
-
Zhang, Junan, Huang, Kaifeng, Chen, Bihuan, Wang, Chong, Tian, Zhenhao, and Peng, Xin
- Subjects
Computer Science - Cryptography and Security ,Computer Science - Software Engineering - Abstract
Open-source software (OSS) supply chain enlarges the attack surface, which makes package registries attractive targets for attacks. Recently, package registries NPM and PyPI have been flooded with malicious packages. The effectiveness of existing malicious NPM and PyPI package detection approaches is hindered by two challenges. The first challenge is how to leverage the knowledge of malicious packages from different ecosystems in a unified way such that multi-lingual malicious package detection can be feasible. The second challenge is how to model malicious behavior in a sequential way such that maliciousness can be precisely captured. To address the two challenges, we propose and implement Cerebro to detect malicious packages in NPM and PyPI. We curate a feature set based on a high-level abstraction of malicious behavior to enable multi-lingual knowledge fusing. We organize extracted features into a behavior sequence to model sequential malicious behavior. We fine-tune the BERT model to understand the semantics of malicious behavior. Extensive evaluation has demonstrated the effectiveness of Cerebro over the state-of-the-art as well as the practically acceptable efficiency. Cerebro has successfully detected 306 and 196 new malicious packages in PyPI and NPM, and received 385 thank letters from the official PyPI and NPM teams.
- Published
- 2023
40. Influence of hip prosthesis position on postoperative gait in symptomatic hip osteoarthritis secondary to hip dysplasia patients after primary total hip arthroplasty: a short-term follow-up study
- Author
-
Yiming Wang, Han Yu, Jianfeng Yang, Kai Xu, Long Cheng, Peng Xin, Jingya Liu, Haichao Ren, Xiaoyu Li, Qingqing Qi, Yan Wang, and Chao Xue
- Subjects
Hip prosthesis position ,Postoperative gait ,Femoral anteversion angle ,Acetabular anteversion angle ,Diseases of the musculoskeletal system ,RC925-935 - Abstract
Abstract Background The aim of this study was to analyze the influence of the positioning of the components of total hip arthroplasty (THA) evaluated by the acetabular anteversion (AA) and femoral anteversion (FA) angle on postoperative gait in patients with symptomatic hip osteoarthritis secondary to hip dysplasia undergoing THA. Methods Between May 2023 and May 2024, patients with symptomatic hip osteoarthritis secondary to hip dysplasia (Crowe Type I and IV) who underwent THA were enrolled in the study. The AA angle and FA angle were measured by computer tomography (CT). Gait data were determined by using the Dynamic Right Gait & Posture analysis system. The relationship between FA, AA and gait data was analyzed by Pearson correlation test, subgroup Pearson correlation test, multiple linear regression. Results A total of 40 patients (45hips) were included in the study. Compared with preoperative, the patient’s postoperative foot progression angle, foot contact angle, plantarflexion velocity, swing foot speed, gait velocity, cadence, stride length were significantly improved. Preoperative FA is significantly different from postoperative FA (P
- Published
- 2024
- Full Text
- View/download PDF
41. Ferroptosis Is Crucial for Cisplatin Induced Sertoli Cell Injury via N6-Methyladenosine Dependent Manner
- Author
-
Zhongru Fan, Peng Xin, Lin Zhao, Chuize Kong, Chiyuan Piao, Zhengqi Wu, Zhongkai Qiu, Wei Zhao, and Zhe Zhang
- Subjects
blood-testis barrier ,cisplatin ,ferroptosis ,rna methylation ,sertoli cells ,Medicine ,Diseases of the genitourinary system. Urology ,RC870-923 - Abstract
Purpose: This study aimed to investigate the effect of the N6-methyladenosine (m6A) dependent ferroptosis on cisplatininduced Sertoli cell injury. Materials and Methods: A cisplatin exposure mouse model was established by intraperitoneal injection of cisplatin in our study. TM4 cell lines was used for in vitro study. Ferroptosis was detected according to metabolomic analysis and a series of assays, including malondialdehyde, glutathione, and glutathione disulfide concentration detection, 2',7'-dichlorodihydrofluorescein diacetate and BODIPY 581/591 C11 probe detection, and transmission electron microscope imaging. Key ferroptosisrelated genes were identified via transcriptomic analysis, western blot and immunohistochemistry. The m6A modification was demonstrated via m6A RNA immunoprecipitation and luciferase reporter assays. Immune cell infiltration was detected by mass cytometry, and verified by flow cytometry and immunofluorescence. Results: Ferroptosis, but not other types of programmed cell death, is a significant phenomenon in cisplatin-induced testis damage and Sertoli cell loss. Ferroptosis induced by cisplatin in Sertoli cell/TM4 cell is GPX4 independent but is regulated by SLC7A11 and ALOX12. Both SLC7A11 and ALOX12 are regulated via m6A dependent manner by METTL3. Furthermore, overexpressed ALOX12-12HETE pathway may result in macrophage polarization and inflammatory response in cisplatin exposure testis. Conclusions: Cisplatin-induced Sertoli cell injury via ferroptosis and promoted ferroptosis in an m6A dependent manner. m6A modification of both SLC7A11 and ALOX12 mRNA could result in ferroptosis in our in vitro model. Further, overexpressed ALOX12 can cause more production of 12-HETE, which may be responsible for testis inflammation caused by cisplatin.
- Published
- 2024
- Full Text
- View/download PDF
42. Recommending Analogical APIs via Knowledge Graph Embedding
- Author
-
Liu, Mingwei, Yang, Yanjun, Lou, Yiling, Peng, Xin, Zhou, Zhong, Du, Xueying, and Yang, Tianyong
- Subjects
Computer Science - Software Engineering - Abstract
Library migration, which re-implements the same software behavior by using a different library instead of using the current one, has been widely observed in software evolution. One essential part of library migration is to find an analogical API that could provide the same functionality as current ones. However, given the large number of libraries/APIs, manually finding an analogical API could be very time-consuming and error-prone. Researchers have developed multiple automated analogical API recommendation techniques. Documentation-based methods have particularly attracted significant interest. Despite their potential, these methods have limitations, such as a lack of comprehensive semantic understanding in documentation and scalability challenges. In this work, we propose KGE4AR, a novel documentation-based approach that leverages knowledge graph (KG) embedding to recommend analogical APIs during library migration. Specifically, KGE4AR proposes a novel unified API KG to comprehensively and structurally represent three types of knowledge in documentation, which can better capture the high-level semantics. Moreover, KGE4AR then proposes to embed the unified API KG into vectors, enabling more effective and scalable similarity calculation. We build KGE4AR' s unified API KG for 35,773 Java libraries and assess it in two API recommendation scenarios: with and without target libraries. Our results show that KGE4AR substantially outperforms state-of-the-art documentation-based techniques in both evaluation scenarios in terms of all metrics (e.g., 47.1%-143.0% and 11.7%-80.6% MRR improvements in each scenario). Additionally, we explore KGE4AR' s scalability, confirming its effective scaling with the growing number of libraries., Comment: Accepted by FSE 2023
- Published
- 2023
43. OpenGCD: Assisting Open World Recognition with Generalized Category Discovery
- Author
-
Gao, Fulin, Zhong, Weimin, Cao, Zhixing, Peng, Xin, and Li, Zhi
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
A desirable open world recognition (OWR) system requires performing three tasks: (1) Open set recognition (OSR), i.e., classifying the known (classes seen during training) and rejecting the unknown (unseen$/$novel classes) online; (2) Grouping and labeling these unknown as novel known classes; (3) Incremental learning (IL), i.e., continual learning these novel classes and retaining the memory of old classes. Ideally, all of these steps should be automated. However, existing methods mostly assume that the second task is completely done manually. To bridge this gap, we propose OpenGCD that combines three key ideas to solve the above problems sequentially: (a) We score the origin of instances (unknown or specifically known) based on the uncertainty of the classifier's prediction; (b) For the first time, we introduce generalized category discovery (GCD) techniques in OWR to assist humans in grouping unlabeled data; (c) For the smooth execution of IL and GCD, we retain an equal number of informative exemplars for each class with diversity as the goal. Moreover, we present a new performance evaluation metric for GCD called harmonic clustering accuracy. Experiments on two standard classification benchmarks and a challenging dataset demonstrate that OpenGCD not only offers excellent compatibility but also substantially outperforms other baselines. Code: https://github.com/Fulin-Gao/OpenGCD.
- Published
- 2023
44. ClassEval: A Manually-Crafted Benchmark for Evaluating LLMs on Class-level Code Generation
- Author
-
Du, Xueying, Liu, Mingwei, Wang, Kaixin, Wang, Hanlin, Liu, Junwei, Chen, Yixuan, Feng, Jiayi, Sha, Chaofeng, Peng, Xin, and Lou, Yiling
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence - Abstract
In this work, we make the first attempt to evaluate LLMs in a more challenging code generation scenario, i.e. class-level code generation. We first manually construct the first class-level code generation benchmark ClassEval of 100 class-level Python code generation tasks with approximately 500 person-hours. Based on it, we then perform the first study of 11 state-of-the-art LLMs on class-level code generation. Based on our results, we have the following main findings. First, we find that all existing LLMs show much worse performance on class-level code generation compared to on standalone method-level code generation benchmarks like HumanEval; and the method-level coding ability cannot equivalently reflect the class-level coding ability among LLMs. Second, we find that GPT-4 and GPT-3.5 still exhibit dominate superior than other LLMs on class-level code generation, and the second-tier models includes Instruct-Starcoder, Instruct-Codegen, and Wizardcoder with very similar performance. Third, we find that generating the entire class all at once (i.e. holistic generation strategy) is the best generation strategy only for GPT-4 and GPT-3.5, while method-by-method generation (i.e. incremental and compositional) is better strategies for the other models with limited ability of understanding long instructions and utilizing the middle information. Lastly, we find the limited model ability of generating method-dependent code and discuss the frequent error types in generated classes. Our benchmark is available at https://github.com/FudanSELab/ClassEval.
- Published
- 2023
45. Evaluating Instruction-Tuned Large Language Models on Code Comprehension and Generation
- Author
-
Yuan, Zhiqiang, Liu, Junwei, Zi, Qiancheng, Liu, Mingwei, Peng, Xin, and Lou, Yiling
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence - Abstract
In this work, we evaluate 10 open-source instructed LLMs on four representative code comprehension and generation tasks. We have the following main findings. First, for the zero-shot setting, instructed LLMs are very competitive on code comprehension and generation tasks and sometimes even better than small SOTA models specifically fine-tuned on each downstream task. We also find that larger instructed LLMs are not always better on code-related tasks. Second, for the few-shot setting, we find that adding demonstration examples substantially helps instructed LLMs perform better on most code comprehension and generation tasks; however, the examples would sometimes induce unstable or even worse performance. Furthermore, we find widely-used BM25-based shot selection strategy significantly outperforms the basic random selection or fixed selection only on generation problems. Third, for the fine-tuning setting, we find that fine-tuning could further improve the model performance on downstream code comprehension and generation tasks compared to the zero-shot/one-shot performance. In addition, after being fine-tuned on the same downstream task dataset, instructed LLMs outperform both the small SOTA models and similar-scaled LLMs without instruction tuning. Based on our findings, we further present practical implications on model and usage recommendation, performance and cost trade-offs, and future direction.
- Published
- 2023
46. Spatio-Temporal Calibration for Omni-Directional Vehicle-Mounted Event Cameras
- Author
-
Li, Xiao, Zhou, Yi, Guo, Ruibin, Peng, Xin, Zhou, Zongtan, and Lu, Huimin
- Subjects
Computer Science - Robotics - Abstract
We present a solution to the problem of spatio-temporal calibration for event cameras mounted on an onmi-directional vehicle. Different from traditional methods that typically determine the camera's pose with respect to the vehicle's body frame using alignment of trajectories, our approach leverages the kinematic correlation of two sets of linear velocity estimates from event data and wheel odometers, respectively. The overall calibration task consists of estimating the underlying temporal offset between the two heterogeneous sensors, and furthermore, recovering the extrinsic rotation that defines the linear relationship between the two sets of velocity estimates. The first sub-problem is formulated as an optimization one, which looks for the optimal temporal offset that maximizes a correlation measurement invariant to arbitrary linear transformation. Once the temporal offset is compensated, the extrinsic rotation can be worked out with an iterative closed-form solver that incrementally registers associated linear velocity estimates. The proposed algorithm is proved effective on both synthetic data and real data, outperforming traditional methods based on alignment of trajectories.
- Published
- 2023
47. CloneRipples: predicting change propagation between code clone instances by graph-based deep learning
- Author
-
Wu, Yijian, Chen, Yuan, Peng, Xin, Hu, Bin, Wang, Xiaochen, Fu, Baiqiang, and Zhao, Wenyun
- Published
- 2025
- Full Text
- View/download PDF
48. Proteomic profile of the antibody diversity in circulating extracellular vesicles of lung adenocarcinoma
- Author
-
Huang, Xinfu, Xiong, Lijuan, Zhang, Yang, Peng, Xin, Ba, Hongping, and Yang, Peng
- Published
- 2024
- Full Text
- View/download PDF
49. Annonaceous acetogenins mimic AA005 targets mitochondrial trifunctional enzyme alpha subunit to treat obesity in male mice
- Author
-
Han, Bing, Li, Zhan-Ming, Zhao, Xu-Yun, Liang, Kai, Mao, Yu-Qin, Zhang, Shi-Long, Huang, Li-Ying, Kong, Chao-Yue, Peng, Xin, Chen, Hui-Ling, Huang, Jia-Ting, Wu, Zhao-Xia, Yao, Jin-Qing, Cai, Pei-Ran, Zhang, Zheng-Yan, Zhang, Xu-Min, Yao, Zhu-Jun, Chen, Guo-Qiang, and Wang, Li-Shun
- Published
- 2024
- Full Text
- View/download PDF
50. Performance-constrained multi-objective optimization of antennas for miniaturization design
- Author
-
Yang, Qi, Wang, Hongqiang, and Peng, Xin
- Published
- 2024
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.