18,918 results on '"Huang, Jin"'
Search Results
2. Asset Building and Child Development: A Policy Model for Inclusive Child Development Accounts
- Author
-
Huang, Jin, Sherraden, Michael, Clancy, Margaret M., Beverly, Sondra G., Shanks, Trina R., and Kim, Youngmi
- Published
- 2021
3. Automatic programming via large language models with population self-evolution for dynamic job shop scheduling problem
- Author
-
Huang, Jin, Li, Xinyu, Gao, Liang, Liu, Qihao, and Teng, Yue
- Subjects
Computer Science - Neural and Evolutionary Computing - Abstract
Heuristic dispatching rules (HDRs) are widely regarded as effective methods for solving dynamic job shop scheduling problems (DJSSP) in real-world production environments. However, their performance is highly scenario-dependent, often requiring expert customization. To address this, genetic programming (GP) and gene expression programming (GEP) have been extensively used for automatic algorithm design. Nevertheless, these approaches often face challenges due to high randomness in the search process and limited generalization ability, hindering the application of trained dispatching rules to new scenarios or dynamic environments. Recently, the integration of large language models (LLMs) with evolutionary algorithms has opened new avenues for prompt engineering and automatic algorithm design. To enhance the capabilities of LLMs in automatic HDRs design, this paper proposes a novel population self-evolutionary (SeEvo) method, a general search framework inspired by the self-reflective design strategies of human experts. The SeEvo method accelerates the search process and enhances exploration capabilities. Experimental results show that the proposed SeEvo method outperforms GP, GEP, end-to-end deep reinforcement learning methods, and more than 10 common HDRs from the literature, particularly in unseen and dynamic scenarios.
- Published
- 2024
4. Training via quantum superposition circumventing local minima and vanishing gradient of sinusoidal neural network
- Author
-
Wen, Zujin, Huang, Jin-Long, and Dahlsten, Oscar
- Subjects
Quantum Physics - Abstract
Deep neural networks have been very successful in applications ranging from computer vision and natural language processing to strategy optimization in games. Recently neural networks with sinusoidal activation functions (SinNN) were found to be ideally suited for representing complex natural signals and their fine spatial and temporal details, which makes them effective representations of images, sound, and video, and good solvers of differential equations. However, training SinNN via gradient descent often results in bad local minima, posing a significant challenge when optimizing their weights. Furthermore, when the weights are discretized for better memory and inference efficiency on small devices, we find that a vanishing gradient problem appears on the resulting discrete SinNN (DSinNN). Brute force search provides an alternative way to find the best weights for DSinNN but is intractable for a large number of parameters. We here provide a qualitatively different training method: an algorithm for quantum training of DSinNNs. The quantum training evolves an initially uniform superposition over weight values to one that is guaranteed to peak on the best weights. We demonstrate the algorithm on toy examples and show that it indeed outperforms gradient descent in optimizing the loss function and outperforms brute force search in the time required.
- Published
- 2024
5. An Improved ESO-Based Line-of-Sight Guidance Law for Path Following of Underactuated Autonomous Underwater Helicopter With Nonlinear Tracking Differentiator and Anti-saturation Controller
- Author
-
Li, Haoda, Liu, Zichen, Huang, Jin, An, Xinyu, and Chen, Ying
- Subjects
Electrical Engineering and Systems Science - Systems and Control - Abstract
This paper presents an Improved Extended-state-observer based Line-of-Sight (IELOS) guidance law for path following of underactuated Autonomous Underwater helicopter (AUH) utilizing a nonlinear tracking differentiator and anti-saturation controller. Due to the high mobility of the AUH, the classical reduced-order Extended-State-Observer (ESO) struggles to accurately track the sideslip angle, especially when rapid variation occurs. By incorporating the nonlinear tracking differentiator and anti-saturation controller, the IELOS guidance law can precisely track sideslip angle and mitigate propeller thrust buffet compared to the classical Extended-state-observer based Line-of-Sight (ELOS) guidance law. The performance of ESO is significantly influenced by the bandwidth, with the Improved Extended-State-Observer (IESO) proving effective at low bandwidths where the classical ESO falls short. The paper establishes the input-to-state stability of the closed-loop system. Subsequently, simulation and pool experimental results are showcased to validate the effectiveness of the IELOS guidance law, which outperforms both the Line-of-Sight (LOS) and Adaptive Line-of-Sight (ALOS) guidance laws in terms of performance.
- Published
- 2024
6. Tuning Fast Memory Size based on Modeling of Page Migration for Tiered Memory
- Author
-
Chen, Shangye, Huang, Jin, Yang, Shuangyan, Liu, Jie, Li, Huaicheng, Nikolopoulos, Dimitrios, Ryu, Junhee, Baek, Jinho, Shin, Kwangsik, and Li, Dong
- Subjects
Computer Science - Performance - Abstract
Tiered memory, built upon a combination of fast memory and slow memory, provides a cost-effective solution to meet ever-increasing requirements from emerging applications for large memory capacity. Reducing the size of fast memory is valuable to improve memory utilization in production and reduce production costs because fast memory tends to be expensive. However, deciding the fast memory size is challenging because there is a complex interplay between application characterization and the overhead of page migration used to mitigate the impact of limited fast memory capacity. In this paper, we introduce a system, Tuna, to decide fast memory size based on modeling of page migration. Tuna uses micro-benchmarking to model the impact of page migration on application performance using three metrics. Tuna decides the fast memory size based on offline modeling results and limited information on workload telemetry. Evaluating with common big-memory applications and using 5% as the performance loss target, we show that Tuna in combination with a page management system (TPP) saves fast memory by 8.5% on average (up to 16%). This is in contrast to the 5% saving in fast memory reported by Microsoft Pond for the same workloads (BFS and SSSP) and the same performance loss target.
- Published
- 2024
7. Analysis of Human Perception in Distinguishing Real and AI-Generated Faces: An Eye-Tracking Based Study
- Author
-
Huang, Jin, Gopalakrishnan, Subhadra, Mittal, Trisha, Zuena, Jake, and Pytlarz, Jaclyn
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Recent advancements in Artificial Intelligence have led to remarkable improvements in generating realistic human faces. While these advancements demonstrate significant progress in generative models, they also raise concerns about the potential misuse of these generated images. In this study, we investigate how humans perceive and distinguish between real and fake images. We designed a perceptual experiment using eye-tracking technology to analyze how individuals differentiate real faces from those generated by AI. Our analysis of StyleGAN-3 generated images reveals that participants can distinguish real from fake faces with an average accuracy of 76.80%. Additionally, we found that participants scrutinize images more closely when they suspect an image to be fake. We believe this study offers valuable insights into human perception of AI-generated media.
- Published
- 2024
8. In-Situ Fine-Tuning of Wildlife Models in IoT-Enabled Camera Traps for Efficient Adaptation
- Author
-
Rastikerdar, Mohammad Mehdi, Huang, Jin, Guan, Hui, and Ganesan, Deepak
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning - Abstract
Wildlife monitoring via camera traps has become an essential tool in ecology, but the deployment of machine learning models for on-device animal classification faces significant challenges due to domain shifts and resource constraints. This paper introduces WildFit, a novel approach that reconciles the conflicting goals of achieving high domain generalization performance and ensuring efficient inference for camera trap applications. WildFit leverages continuous background-aware model fine-tuning to deploy ML models tailored to the current location and time window, allowing it to maintain robust classification accuracy in the new environment without requiring significant computational resources. This is achieved by background-aware data synthesis, which generates training images representing the new domain by blending background images with animal images from the source domain. We further enhance fine-tuning effectiveness through background drift detection and class distribution drift detection, which optimize the quality of synthesized data and improve generalization performance. Our extensive evaluation across multiple camera trap datasets demonstrates that WildFit achieves significant improvements in classification accuracy and computational efficiency compared to traditional approaches.
- Published
- 2024
9. Surface molecular pump enables ultrahigh catalyst activity.
- Author
-
Huang, Jin, Peng, Bosi, Zhu, Cheng, Xu, Mingjie, Liu, Yang, Liu, Zeyan, Zhou, Jingxuan, Wang, Sibo, Duan, Xiangfeng, Heinz, Hendrik, and Huang, Yu
- Abstract
The performance of electrocatalysts is critical for renewable energy technologies. While the electrocatalytic activity can be modulated through structural and compositional engineering following the Sabatier principle, the insufficiently explored catalyst-electrolyte interface is promising to promote microkinetic processes such as physisorption and desorption. By combining experimental designs and molecular dynamics simulations with explicit solvent in high accuracy, we demonstrated that dimethylformamide can work as an effective surface molecular pump to facilitate the entrapment of oxygen and outflux of water. Dimethylformamide disrupts the interfacial network of hydrogen bonds, leading to enhanced activity of the oxygen reduction reaction by a factor of 2 to 3. This strategy works generally for platinum-alloy catalysts, and we introduce an optimal model PtCuNi catalyst with an unprecedented specific activity of 21.8 ± 2.1 mA/cm2 at 0.9 V versus the reversible hydrogen electrode, nearly double the previous record, and an ultrahigh mass activity of 10.7 ± 1.1 A/mgPt.
- Published
- 2024
10. SciLitLLM: How to Adapt LLMs for Scientific Literature Understanding
- Author
-
Li, Sihang, Huang, Jin, Zhuang, Jiaxi, Shi, Yaorui, Cai, Xiaochen, Xu, Mingjun, Wang, Xiang, Zhang, Linfeng, Ke, Guolin, and Cai, Hengxing
- Subjects
Computer Science - Machine Learning ,Computer Science - Computation and Language - Abstract
Scientific literature understanding is crucial for extracting targeted information and garnering insights, thereby significantly advancing scientific discovery. Despite the remarkable success of Large Language Models (LLMs), they face challenges in scientific literature understanding, primarily due to (1) a lack of scientific knowledge and (2) unfamiliarity with specialized scientific tasks. To develop an LLM specialized in scientific literature understanding, we propose a hybrid strategy that integrates continual pre-training (CPT) and supervised fine-tuning (SFT), to simultaneously infuse scientific domain knowledge and enhance instruction-following capabilities for domain-specific tasks.cIn this process, we identify two key challenges: (1) constructing high-quality CPT corpora, and (2) generating diverse SFT instructions. We address these challenges through a meticulous pipeline, including PDF text extraction, parsing content error correction, quality filtering, and synthetic instruction creation. Applying this strategy, we present a suite of LLMs: SciLitLLM, specialized in scientific literature understanding. These models demonstrate promising performance on scientific literature understanding benchmarks. Our contributions are threefold: (1) We present an effective framework that integrates CPT and SFT to adapt LLMs to scientific literature understanding, which can also be easily adapted to other domains. (2) We propose an LLM-based synthesis method to generate diverse and high-quality scientific instructions, resulting in a new instruction set -- SciLitIns -- for supervised fine-tuning in less-represented scientific domains. (3) SciLitLLM achieves promising performance improvements on scientific literature understanding benchmarks.
- Published
- 2024
11. Phase behavior of symmetric diblock copolymers under 3D soft confinement
- Author
-
He, Zhijuan, Huang, Jin, Jiang, Kai, and Shi, An-Chang
- Subjects
Condensed Matter - Soft Condensed Matter - Abstract
The phase behavior of symmetric diblock copolymers under three-dimensional (3D) soft confinement is investigated using the self-consistent field theory. The soft confinement is realized in binary blends composed AB diblock copolymers and C homopolymers, where the copolymers self-assemble to form a droplet embedded in the homopolymer matrix. The phase behavior of the confined block copolymers is regulated by the degree of confinement and the selectivity of the homopolymers, resulting in a rich variety of novel structures. When the C homopolymers are neutral to the A- and B-blocks, stacked lamellae (SL) are formed where the number of layers increases with the droplet volume, resulting in a morphological transition sequence from Janus particle to square SL. When the C homopolymers are strongly selective to the B-blocks, a series of non-lamellar morphologies, including onion-, hamburger-, cross-, ring-, and cookie-like structures, are observed. A detailed free energy analysis reveals a first-order reversible transformation between SL and onion-like (OL) structures when the selectivity of the homopolymers is changed. Our results provide a comprehensive understanding of how various factors, such as the copolymer concentration, homopolymer chain length, degree of confinement, homopolymer selectivity, affect the self-assembled structures of diblock copolymers under soft 3D confinement., Comment: 9pages,7figures
- Published
- 2024
12. Risk Occupancy: A New and Efficient Paradigm through Vehicle-Road-Cloud Collaboration
- Author
-
Chen, Jiaxing, Zhong, Wei, Gao, Bolin, Liu, Yifei, Zou, Hengduo, Liu, Jiaxi, Lu, Yanbo, Huang, Jin, and Zhong, Zhihua
- Subjects
Computer Science - Robotics - Abstract
This study introduces the 4D Risk Occupancy within a vehicle-road-cloud architecture, integrating the road surface spatial, risk, and temporal dimensions, and endowing the algorithm with beyond-line-of-sight, all-angles, and efficient abilities. The algorithm simplifies risk modeling by focusing on directly observable information and key factors, drawing on the concept of Occupancy Grid Maps (OGM), and incorporating temporal prediction to effectively map current and future risk occupancy. Compared to conventional driving risk fields and grid occupancy maps, this algorithm can map global risks more efficiently, simply, and reliably. It can integrate future risk information, adapting to dynamic traffic environments. The 4D Risk Occupancy also unifies the expression of BEV detection and lane line detection results, enhancing the intuitiveness and unity of environmental perception. Using DAIR-V2X data, this paper validates the 4D Risk Occupancy algorithm and develops a local path planning model based on it. Qualitative experiments under various road conditions demonstrate the practicality and robustness of this local path planning model. Quantitative analysis shows that the path planning based on risk occupation significantly improves trajectory planning performance, increasing safety redundancy by 12.5% and reducing average deceleration by 5.41% at an initial braking speed of 8 m/s, thereby improving safety and comfort. This work provides a new global perception method and local path planning method through Vehicle-Road-Cloud architecture, offering a new perceptual paradigm for achieving safer and more efficient autonomous driving., Comment: 13 pages,9 figures
- Published
- 2024
13. ICSFuzz: Collision Detector Bug Discovery in Autonomous Driving Simulators
- Author
-
Fu, Weiwei, Huang, Heqing, Zhang, Yifan, Zhang, Ke, Huang, Jin, Lee, Wei-Bin, and Wang, Jianping
- Subjects
Computer Science - Cryptography and Security - Abstract
With the increasing adoption of autonomous vehicles, ensuring the reliability of autonomous driving systems (ADSs) deployed on autonomous vehicles has become a significant concern. Driving simulators have emerged as crucial platforms for testing autonomous driving systems, offering realistic, dynamic, and configurable environments. However, existing simulation-based ADS testers have largely overlooked the reliability of the simulators, potentially leading to overlooked violation scenarios and subsequent safety security risks during real-world deployment. In our investigations, we identified that collision detectors in simulators could fail to detect and report collisions in certain collision scenarios, referred to as ignored collision scenarios. This paper aims to systematically discover ignored collision scenarios to improve the reliability of autonomous driving simulators. To this end, we present ICSFuzz, a black-box fuzzing approach to discover ignored collision scenarios efficiently. Drawing upon the fact that the ignored collision scenarios are a sub-type of collision scenarios, our approach starts with the determined collision scenarios. Following the guidance provided by empirically studied factors contributing to collisions, we selectively mutate arbitrary collision scenarios in a step-wise manner toward the ignored collision scenarios and effectively discover them. We compare ICSFuzz with DriveFuzz, a state-of-the-art simulation-based ADS testing method, by replacing its oracle with our ignored-collision-aware oracle. The evaluation demonstrates that ICSFuzz outperforms DriveFuzz by finding 10-20x more ignored collision scenarios with a 20-70x speedup. All the discovered ignored collisions have been confirmed by developers with one CVE ID assigned.
- Published
- 2024
14. Controlling superradiant phase transition in quantum Rabi model
- Author
-
Xie, Xuan, Liu, Cheng, Jiang, Lin-Lin, and Huang, Jin-Feng
- Subjects
Quantum Physics - Abstract
In the ultrastrong-coupling regime, the quantum Rabi model can exhibit quantum phase transition (QPT) when the ratio of the qubit transition frequency to the frequency of the cavity field approaches infinity. However, it is challenging to control the QPT in few-body systems because of the limited coupling strength and the A^2 terms. Here, we propose a practical scheme to manipulate the QPT of quantum Rabi model in the strong-coupling regime. By applying a periodic frequency modulation to the two-level system in a standard quantum Rabi model in the strong-coupling regime, an anisotropic quantum Rabi model with ultrastrong and tunable coupling strengths for rotating and counter-rotating terms is obtained. The ground-state and excitation energy of this model in terms of the modulation parameters are studied. We find that the QPT of quantum Rabi model can be observed in the strong-coupling regime and externally controlled by the modulation., Comment: 10 pages, 5 figures
- Published
- 2024
15. Enhancing Ecological Monitoring with Multi-Objective Optimization: A Novel Dataset and Methodology for Segmentation Algorithms
- Author
-
Abraham, Sophia J., Huang, Jin, RichardWebster, Brandon, Milford, Michael, Hauenstein, Jonathan D., and Scheirer, Walter
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
We introduce a unique semantic segmentation dataset of 6,096 high-resolution aerial images capturing indigenous and invasive grass species in Bega Valley, New South Wales, Australia, designed to address the underrepresented domain of ecological data in the computer vision community. This dataset presents a challenging task due to the overlap and distribution of grass species, which is critical for advancing models in ecological and agronomical applications. Our study features a homotopy-based multi-objective fine-tuning approach that balances segmentation accuracy and contextual consistency, applicable to various models. By integrating DiceCELoss for pixel-wise classification and a smoothness loss for spatial coherence, this method evolves during training to enhance robustness against noisy data. Performance baselines are established through a case study on the Segment Anything Model (SAM), demonstrating its effectiveness. Our annotation methodology, emphasizing pen size, zoom control, and memory management, ensures high-quality dataset creation. The dataset and code will be made publicly available, aiming to drive research in computer vision, machine learning, and ecological studies, advancing environmental monitoring and sustainable development.
- Published
- 2024
16. DCA-Bench: A Benchmark for Dataset Curation Agents
- Author
-
Huang, Benhao, Yu, Yingzhuo, Huang, Jin, Zhang, Xingjian, and Ma, Jiaqi
- Subjects
Computer Science - Artificial Intelligence - Abstract
The quality of datasets plays an increasingly crucial role in the research and development of modern artificial intelligence (AI). Despite the proliferation of open dataset platforms nowadays, data quality issues, such as insufficient documentation, inaccurate annotations, and ethical concerns, remain common in datasets widely used in AI. Furthermore, these issues are often subtle and difficult to be detected by rule-based scripts, requiring expensive manual identification and verification by dataset users or maintainers. With the increasing capability of large language models (LLMs), it is promising to streamline the curation of datasets with LLM agents. In this work, as the initial step towards this goal, we propose a dataset curation agent benchmark, DCA-Bench, to measure LLM agents' capability of detecting hidden dataset quality issues. Specifically, we collect diverse real-world dataset quality issues from eight open dataset platforms as a testbed. Additionally, to establish an automatic pipeline for evaluating the success of LLM agents, which requires a nuanced understanding of the agent outputs, we implement a dedicated Evaluator using another LLM agent. We demonstrate that the LLM-based Evaluator empirically aligns well with human evaluation, allowing reliable automatic evaluation on the proposed benchmark. We further conduct experiments on several baseline LLM agents on the proposed benchmark and demonstrate the complexity of the task, indicating that applying LLMs to real-world dataset curation still requires further in-depth exploration and innovation. Finally, the proposed benchmark can also serve as a testbed for measuring the capability of LLMs in problem discovery rather than just problem-solving. The benchmark suite is available at \url{https://github.com/TRAIS-Lab/dca-bench}.
- Published
- 2024
17. MASSW: A New Dataset and Benchmark Tasks for AI-Assisted Scientific Workflows
- Author
-
Zhang, Xingjian, Xie, Yutong, Huang, Jin, Ma, Jinge, Pan, Zhaoying, Liu, Qijia, Xiong, Ziyang, Ergen, Tolga, Shim, Dongsub, Lee, Honglak, and Mei, Qiaozhu
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence - Abstract
Scientific innovation relies on detailed workflows, which include critical steps such as analyzing literature, generating ideas, validating these ideas, interpreting results, and inspiring follow-up research. However, scientific publications that document these workflows are extensive and unstructured. This makes it difficult for both human researchers and AI systems to effectively navigate and explore the space of scientific innovation. To address this issue, we introduce MASSW, a comprehensive text dataset on Multi-Aspect Summarization of Scientific Workflows. MASSW includes more than 152,000 peer-reviewed publications from 17 leading computer science conferences spanning the past 50 years. Using Large Language Models (LLMs), we automatically extract five core aspects from these publications -- context, key idea, method, outcome, and projected impact -- which correspond to five key steps in the research workflow. These structured summaries facilitate a variety of downstream tasks and analyses. The quality of the LLM-extracted summaries is validated by comparing them with human annotations. We demonstrate the utility of MASSW through multiple novel machine-learning tasks that can be benchmarked using this new dataset, which make various types of predictions and recommendations along the scientific workflow. MASSW holds significant potential for researchers to create and benchmark new AI methods for optimizing scientific workflows and fostering scientific innovation in the field. Our dataset is openly available at \url{https://github.com/xingjian-zhang/massw}., Comment: arXiv admin note: text overlap with arXiv:1706.03762 by other authors
- Published
- 2024
18. KB4Rec: A Data Set for Linking Knowledge Bases with Recommender Systems
- Author
-
Zhao, Wayne Xin, He, Gaole, Yang, Kunlin, Dou, Hongjian, Huang, Jin, Ouyang, Siqi, and Wen, Ji-Rong
- Subjects
Information technology ,T58.5-58.64 - Abstract
To develop a knowledge-aware recommender system, a key issue is how to obtain rich and structured knowledge base (KB) information for recommender system (RS) items. Existing data sets or methods either use side information from original RSs (containing very few kinds of useful information) or utilize a private KB. In this paper, we present KB4Rec v1.0, a data set linking KB information for RSs. It has linked three widely used RS data sets with two popular KBs, namely Freebase and YAGO. Based on our linked data set, we first preform qualitative analysis experiments, and then we discuss the effect of two important factors (i.e., popularity and recency) on whether a RS item can be linked to a KB entity. Finally, we compare several knowledge-aware recommendation algorithms on our linked data set.
- Published
- 2019
- Full Text
- View/download PDF
19. Effectiveness of denoising diffusion probabilistic models for fast and high-fidelity whole-event simulation in high-energy heavy-ion experiments
- Author
-
Go, Yeonju, Torbunov, Dmitrii, Rinn, Timothy, Huang, Yi, Yu, Haiwang, Viren, Brett, Lin, Meifeng, Ren, Yihui, and Huang, Jin
- Subjects
Physics - Data Analysis, Statistics and Probability ,High Energy Physics - Experiment ,Nuclear Experiment - Abstract
Artificial intelligence (AI) generative models, such as generative adversarial networks (GANs), variational auto-encoders, and normalizing flows, have been widely used and studied as efficient alternatives for traditional scientific simulations. However, they have several drawbacks, including training instability and inability to cover the entire data distribution, especially for regions where data are rare. This is particularly challenging for whole-event, full-detector simulations in high-energy heavy-ion experiments, such as sPHENIX at the Relativistic Heavy Ion Collider and Large Hadron Collider experiments, where thousands of particles are produced per event and interact with the detector. This work investigates the effectiveness of Denoising Diffusion Probabilistic Models (DDPMs) as an AI-based generative surrogate model for the sPHENIX experiment that includes the heavy-ion event generation and response of the entire calorimeter stack. DDPM performance in sPHENIX simulation data is compared with a popular rival, GANs. Results show that both DDPMs and GANs can reproduce the data distribution where the examples are abundant (low-to-medium calorimeter energies). Nonetheless, DDPMs significantly outperform GANs, especially in high-energy regions where data are rare. Additionally, DDPMs exhibit superior stability compared to GANs. The results are consistent between both central and peripheral centrality heavy-ion collision events. Moreover, DDPMs offer a substantial speedup of approximately a factor of 100 compared to the traditional Geant4 simulation method., Comment: 11 pages, 7 figures
- Published
- 2024
20. Salient Object Detection From Arbitrary Modalities
- Author
-
Huang, Nianchang, Yang, Yang, Xi, Ruida, Zhang, Qiang, Han, Jungong, and Huang, Jin
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Toward desirable saliency prediction, the types and numbers of inputs for a salient object detection (SOD) algorithm may dynamically change in many real-life applications. However, existing SOD algorithms are mainly designed or trained for one particular type of inputs, failing to be generalized to other types of inputs. Consequentially, more types of SOD algorithms need to be prepared in advance for handling different types of inputs, raising huge hardware and research costs. Differently, in this paper, we propose a new type of SOD task, termed Arbitrary Modality SOD (AM SOD). The most prominent characteristics of AM SOD are that the modality types and modality numbers will be arbitrary or dynamically changed. The former means that the inputs to the AM SOD algorithm may be arbitrary modalities such as RGB, depths, or even any combination of them. While, the latter indicates that the inputs may have arbitrary modality numbers as the input type is changed, e.g. single-modality RGB image, dual-modality RGB-Depth (RGB-D) images or triple-modality RGB-Depth-Thermal (RGB-D-T) images. Accordingly, a preliminary solution to the above challenges, \i.e. a modality switch network (MSN), is proposed in this paper. In particular, a modality switch feature extractor (MSFE) is first designed to extract discriminative features from each modality effectively by introducing some modality indicators, which will generate some weights for modality switching. Subsequently, a dynamic fusion module (DFM) is proposed to adaptively fuse features from a variable number of modalities based on a novel Transformer structure. Finally, a new dataset, named AM-XD, is constructed to facilitate research on AM SOD. Extensive experiments demonstrate that our AM SOD method can effectively cope with changes in the type and number of input modalities for robust salient object detection., Comment: 15 Pages, 7 Figures, 8 Tables
- Published
- 2024
21. Modality Prompts for Arbitrary Modality Salient Object Detection
- Author
-
Huang, Nianchang, Yang, Yang, Zhang, Qiang, Han, Jungong, and Huang, Jin
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
This paper delves into the task of arbitrary modality salient object detection (AM SOD), aiming to detect salient objects from arbitrary modalities, eg RGB images, RGB-D images, and RGB-D-T images. A novel modality-adaptive Transformer (MAT) will be proposed to investigate two fundamental challenges of AM SOD, ie more diverse modality discrepancies caused by varying modality types that need to be processed, and dynamic fusion design caused by an uncertain number of modalities present in the inputs of multimodal fusion strategy. Specifically, inspired by prompt learning's ability of aligning the distributions of pre-trained models to the characteristic of downstream tasks by learning some prompts, MAT will first present a modality-adaptive feature extractor (MAFE) to tackle the diverse modality discrepancies by introducing a modality prompt for each modality. In the training stage, a new modality translation contractive (MTC) loss will be further designed to assist MAFE in learning those modality-distinguishable modality prompts. Accordingly, in the testing stage, MAFE can employ those learned modality prompts to adaptively adjust its feature space according to the characteristics of the input modalities, thus being able to extract discriminative unimodal features. Then, MAFE will present a channel-wise and spatial-wise fusion hybrid (CSFH) strategy to meet the demand for dynamic fusion. For that, CSFH dedicates a channel-wise dynamic fusion module (CDFM) and a novel spatial-wise dynamic fusion module (SDFM) to fuse the unimodal features from varying numbers of modalities and meanwhile effectively capture cross-modal complementary semantic and detail information, respectively. Moreover, CSFH will carefully align CDFM and SDFM to different levels of unimodal features based on their characteristics for more effective complementary information exploitation., Comment: 13 pages, 7 Figures, 3 Tables
- Published
- 2024
22. Unleashing the Power of Multi-Task Learning: A Comprehensive Survey Spanning Traditional, Deep, and Pretrained Foundation Model Eras
- Author
-
Yu, Jun, Dai, Yutong, Liu, Xiaokang, Huang, Jin, Shen, Yishan, Zhang, Ke, Zhou, Rong, Adhikarla, Eashan, Ye, Wenxuan, Liu, Yixin, Kong, Zhaoming, Zhang, Kai, Yin, Yilong, Namboodiri, Vinod, Davison, Brian D., Moore, Jason H., and Chen, Yong
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Computer Vision and Pattern Recognition - Abstract
MTL is a learning paradigm that effectively leverages both task-specific and shared information to address multiple related tasks simultaneously. In contrast to STL, MTL offers a suite of benefits that enhance both the training process and the inference efficiency. MTL's key advantages encompass streamlined model architecture, performance enhancement, and cross-domain generalizability. Over the past twenty years, MTL has become widely recognized as a flexible and effective approach in various fields, including CV, NLP, recommendation systems, disease prognosis and diagnosis, and robotics. This survey provides a comprehensive overview of the evolution of MTL, encompassing the technical aspects of cutting-edge methods from traditional approaches to deep learning and the latest trend of pretrained foundation models. Our survey methodically categorizes MTL techniques into five key areas: regularization, relationship learning, feature propagation, optimization, and pre-training. This categorization not only chronologically outlines the development of MTL but also dives into various specialized strategies within each category. Furthermore, the survey reveals how the MTL evolves from handling a fixed set of tasks to embracing a more flexible approach free from task or modality constraints. It explores the concepts of task-promptable and -agnostic training, along with the capacity for ZSL, which unleashes the untapped potential of this historically coveted learning paradigm. Overall, we hope this survey provides the research community with a comprehensive overview of the advancements in MTL from its inception in 1997 to the present in 2023. We address present challenges and look ahead to future possibilities, shedding light on the opportunities and potential avenues for MTL research in a broad manner. This project is publicly available at https://github.com/junfish/Awesome-Multitask-Learning., Comment: 60 figures, 116 pages, 500+ references
- Published
- 2024
23. Going Beyond Popularity and Positivity Bias: Correcting for Multifactorial Bias in Recommender Systems
- Author
-
Huang, Jin, Oosterhuis, Harrie, Mansoury, Masoud, van Hoof, Herke, and de Rijke, Maarten
- Subjects
Computer Science - Information Retrieval - Abstract
Two typical forms of bias in user interaction data with recommender systems (RSs) are popularity bias and positivity bias, which manifest themselves as the over-representation of interactions with popular items or items that users prefer, respectively. Debiasing methods aim to mitigate the effect of selection bias on the evaluation and optimization of RSs. However, existing debiasing methods only consider single-factor forms of bias, e.g., only the item (popularity) or only the rating value (positivity). This is in stark contrast with the real world where user selections are generally affected by multiple factors at once. In this work, we consider multifactorial selection bias in RSs. Our focus is on selection bias affected by both item and rating value factors, which is a generalization and combination of popularity and positivity bias. While the concept of multifactorial bias is intuitive, it brings a severe practical challenge as it requires substantially more data for accurate bias estimation. As a solution, we propose smoothing and alternating gradient descent techniques to reduce variance and improve the robustness of its optimization. Our experimental results reveal that, with our proposed techniques, multifactorial bias corrections are more effective and robust than single-factor counterparts on real-world and synthetic datasets., Comment: SIGIR 2024
- Published
- 2024
- Full Text
- View/download PDF
24. Debiasing Machine Unlearning with Counterfactual Examples
- Author
-
Chen, Ziheng, Wang, Jia, Zhuang, Jun, Reddy, Abbavaram Gowtham, Silvestri, Fabrizio, Huang, Jin, Nag, Kaushiki, Kuang, Kun, Ning, Xin, and Tolomei, Gabriele
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Statistics - Machine Learning - Abstract
The right to be forgotten (RTBF) seeks to safeguard individuals from the enduring effects of their historical actions by implementing machine-learning techniques. These techniques facilitate the deletion of previously acquired knowledge without requiring extensive model retraining. However, they often overlook a critical issue: unlearning processes bias. This bias emerges from two main sources: (1) data-level bias, characterized by uneven data removal, and (2) algorithm-level bias, which leads to the contamination of the remaining dataset, thereby degrading model accuracy. In this work, we analyze the causal factors behind the unlearning process and mitigate biases at both data and algorithmic levels. Typically, we introduce an intervention-based approach, where knowledge to forget is erased with a debiased dataset. Besides, we guide the forgetting procedure by leveraging counterfactual examples, as they maintain semantic data consistency without hurting performance on the remaining dataset. Experimental results demonstrate that our method outperforms existing machine unlearning baselines on evaluation metrics.
- Published
- 2024
25. Inner fission barriers of uranium isotopes in the deformed relativistic Hartree-Bogoliubov theory in continuum
- Author
-
Zhang, Wei, Huang, Jin-Ke, Sun, Ting-Ting, Peng, Jing, and Zhang, Shuang Quan
- Subjects
Nuclear Theory - Abstract
The inner fission barriers of the even-even uranium isotopes from the proton to the neutron drip line are studied with the deformed relativistic Hartree-Bogoliubov theory in continuum. A periodic evolution for the ground state shapes is shown with the neutron number, i.e., spherical shapes at shell closures $N=$126, 184, 258, and prolate dominated shapes between them. In analogy to the shape evolution, the inner fission barriers also exhibit a periodic behavior: peaks at the shell closures and valleys in the mid-shells. The triaxial effect to the inner fission barrier is evaluated using the triaxial relativistic mean field calculations plus a simple BCS method for pairing. With the triaxial correction included, good consistency in the inner barrier heights is found with the available empirical data. Besides, the evolution from the proton to the neutron drip line is in accord with the results by the multi-dimensionally constrained relativistic mean field theory. A flat valley in the fission barrier height is predicted around the neutron-rich nucleus $^{318}$U which may play a role of fission recycling in the astrophysical $r$-process nucleosynthesis., Comment: 7 pages, 6 figures
- Published
- 2024
26. Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context
- Author
-
Gemini Team, Georgiev, Petko, Lei, Ving Ian, Burnell, Ryan, Bai, Libin, Gulati, Anmol, Tanzer, Garrett, Vincent, Damien, Pan, Zhufeng, Wang, Shibo, Mariooryad, Soroosh, Ding, Yifan, Geng, Xinyang, Alcober, Fred, Frostig, Roy, Omernick, Mark, Walker, Lexi, Paduraru, Cosmin, Sorokin, Christina, Tacchetti, Andrea, Gaffney, Colin, Daruki, Samira, Sercinoglu, Olcan, Gleicher, Zach, Love, Juliette, Voigtlaender, Paul, Jain, Rohan, Surita, Gabriela, Mohamed, Kareem, Blevins, Rory, Ahn, Junwhan, Zhu, Tao, Kawintiranon, Kornraphop, Firat, Orhan, Gu, Yiming, Zhang, Yujing, Rahtz, Matthew, Faruqui, Manaal, Clay, Natalie, Gilmer, Justin, Co-Reyes, JD, Penchev, Ivo, Zhu, Rui, Morioka, Nobuyuki, Hui, Kevin, Haridasan, Krishna, Campos, Victor, Mahdieh, Mahdis, Guo, Mandy, Hassan, Samer, Kilgour, Kevin, Vezer, Arpi, Cheng, Heng-Tze, de Liedekerke, Raoul, Goyal, Siddharth, Barham, Paul, Strouse, DJ, Noury, Seb, Adler, Jonas, Sundararajan, Mukund, Vikram, Sharad, Lepikhin, Dmitry, Paganini, Michela, Garcia, Xavier, Yang, Fan, Valter, Dasha, Trebacz, Maja, Vodrahalli, Kiran, Asawaroengchai, Chulayuth, Ring, Roman, Kalb, Norbert, Soares, Livio Baldini, Brahma, Siddhartha, Steiner, David, Yu, Tianhe, Mentzer, Fabian, He, Antoine, Gonzalez, Lucas, Xu, Bibo, Kaufman, Raphael Lopez, Shafey, Laurent El, Oh, Junhyuk, Hennigan, Tom, Driessche, George van den, Odoom, Seth, Lucic, Mario, Roelofs, Becca, Lall, Sid, Marathe, Amit, Chan, Betty, Ontanon, Santiago, He, Luheng, Teplyashin, Denis, Lai, Jonathan, Crone, Phil, Damoc, Bogdan, Ho, Lewis, Riedel, Sebastian, Lenc, Karel, Yeh, Chih-Kuan, Chowdhery, Aakanksha, Xu, Yang, Kazemi, Mehran, Amid, Ehsan, Petrushkina, Anastasia, Swersky, Kevin, Khodaei, Ali, Chen, Gowoon, Larkin, Chris, Pinto, Mario, Yan, Geng, Badia, Adria Puigdomenech, Patil, Piyush, Hansen, Steven, Orr, Dave, Arnold, Sebastien M. R., Grimstad, Jordan, Dai, Andrew, Douglas, Sholto, Sinha, Rishika, Yadav, Vikas, Chen, Xi, Gribovskaya, Elena, Austin, Jacob, Zhao, Jeffrey, Patel, Kaushal, Komarek, Paul, Austin, Sophia, Borgeaud, Sebastian, Friso, Linda, Goyal, Abhimanyu, Caine, Ben, Cao, Kris, Chung, Da-Woon, Lamm, Matthew, Barth-Maron, Gabe, Kagohara, Thais, Olszewska, Kate, Chen, Mia, Shivakumar, Kaushik, Agarwal, Rishabh, Godhia, Harshal, Rajwar, Ravi, Snaider, Javier, Dotiwalla, Xerxes, Liu, Yuan, Barua, Aditya, Ungureanu, Victor, Zhang, Yuan, Batsaikhan, Bat-Orgil, Wirth, Mateo, Qin, James, Danihelka, Ivo, Doshi, Tulsee, Chadwick, Martin, Chen, Jilin, Jain, Sanil, Le, Quoc, Kar, Arjun, Gurumurthy, Madhu, Li, Cheng, Sang, Ruoxin, Liu, Fangyu, Lamprou, Lampros, Munoz, Rich, Lintz, Nathan, Mehta, Harsh, Howard, Heidi, Reynolds, Malcolm, Aroyo, Lora, Wang, Quan, Blanco, Lorenzo, Cassirer, Albin, Griffith, Jordan, Das, Dipanjan, Lee, Stephan, Sygnowski, Jakub, Fisher, Zach, Besley, James, Powell, Richard, Ahmed, Zafarali, Paulus, Dominik, Reitter, David, Borsos, Zalan, Joshi, Rishabh, Pope, Aedan, Hand, Steven, Selo, Vittorio, Jain, Vihan, Sethi, Nikhil, Goel, Megha, Makino, Takaki, May, Rhys, Yang, Zhen, Schalkwyk, Johan, Butterfield, Christina, Hauth, Anja, Goldin, Alex, Hawkins, Will, Senter, Evan, Brin, Sergey, Woodman, Oliver, Ritter, Marvin, Noland, Eric, Giang, Minh, Bolina, Vijay, Lee, Lisa, Blyth, Tim, Mackinnon, Ian, Reid, Machel, Sarvana, Obaid, Silver, David, Chen, Alexander, Wang, Lily, Maggiore, Loren, Chang, Oscar, Attaluri, Nithya, Thornton, Gregory, Chiu, Chung-Cheng, Bunyan, Oskar, Levine, Nir, Chung, Timothy, Eltyshev, Evgenii, Si, Xiance, Lillicrap, Timothy, Brady, Demetra, Aggarwal, Vaibhav, Wu, Boxi, Xu, Yuanzhong, McIlroy, Ross, Badola, Kartikeya, Sandhu, Paramjit, Moreira, Erica, Stokowiec, Wojciech, Hemsley, Ross, Li, Dong, Tudor, Alex, Shyam, Pranav, Rahimtoroghi, Elahe, Haykal, Salem, Sprechmann, Pablo, Zhou, Xiang, Mincu, Diana, Li, Yujia, Addanki, Ravi, Krishna, Kalpesh, Wu, Xiao, Frechette, Alexandre, Eyal, Matan, Dafoe, Allan, Lacey, Dave, Whang, Jay, Avrahami, Thi, Zhang, Ye, Taropa, Emanuel, Lin, Hanzhao, Toyama, Daniel, Rutherford, Eliza, Sano, Motoki, Choe, HyunJeong, Tomala, Alex, Safranek-Shrader, Chalence, Kassner, Nora, Pajarskas, Mantas, Harvey, Matt, Sechrist, Sean, Fortunato, Meire, Lyu, Christina, Elsayed, Gamaleldin, Kuang, Chenkai, Lottes, James, Chu, Eric, Jia, Chao, Chen, Chih-Wei, Humphreys, Peter, Baumli, Kate, Tao, Connie, Samuel, Rajkumar, Santos, Cicero Nogueira dos, Andreassen, Anders, Rakićević, Nemanja, Grewe, Dominik, Kumar, Aviral, Winkler, Stephanie, Caton, Jonathan, Brock, Andrew, Dalmia, Sid, Sheahan, Hannah, Barr, Iain, Miao, Yingjie, Natsev, Paul, Devlin, Jacob, Behbahani, Feryal, Prost, Flavien, Sun, Yanhua, Myaskovsky, Artiom, Pillai, Thanumalayan Sankaranarayana, Hurt, Dan, Lazaridou, Angeliki, Xiong, Xi, Zheng, Ce, Pardo, Fabio, Li, Xiaowei, Horgan, Dan, Stanton, Joe, Ambar, Moran, Xia, Fei, Lince, Alejandro, Wang, Mingqiu, Mustafa, Basil, Webson, Albert, Lee, Hyo, Anil, Rohan, Wicke, Martin, Dozat, Timothy, Sinha, Abhishek, Piqueras, Enrique, Dabir, Elahe, Upadhyay, Shyam, Boral, Anudhyan, Hendricks, Lisa Anne, Fry, Corey, Djolonga, Josip, Su, Yi, Walker, Jake, Labanowski, Jane, Huang, Ronny, Misra, Vedant, Chen, Jeremy, Skerry-Ryan, RJ, Singh, Avi, Rijhwani, Shruti, Yu, Dian, Castro-Ros, Alex, Changpinyo, Beer, Datta, Romina, Bagri, Sumit, Hrafnkelsson, Arnar Mar, Maggioni, Marcello, Zheng, Daniel, Sulsky, Yury, Hou, Shaobo, Paine, Tom Le, Yang, Antoine, Riesa, Jason, Rogozinska, Dominika, Marcus, Dror, Badawy, Dalia El, Zhang, Qiao, Wang, Luyu, Miller, Helen, Greer, Jeremy, Sjos, Lars Lowe, Nova, Azade, Zen, Heiga, Chaabouni, Rahma, Rosca, Mihaela, Jiang, Jiepu, Chen, Charlie, Liu, Ruibo, Sainath, Tara, Krikun, Maxim, Polozov, Alex, Lespiau, Jean-Baptiste, Newlan, Josh, Cankara, Zeyncep, Kwak, Soo, Xu, Yunhan, Chen, Phil, Coenen, Andy, Meyer, Clemens, Tsihlas, Katerina, Ma, Ada, Gottweis, Juraj, Xing, Jinwei, Gu, Chenjie, Miao, Jin, Frank, Christian, Cankara, Zeynep, Ganapathy, Sanjay, Dasgupta, Ishita, Hughes-Fitt, Steph, Chen, Heng, Reid, David, Rong, Keran, Fan, Hongmin, van Amersfoort, Joost, Zhuang, Vincent, Cohen, Aaron, Gu, Shixiang Shane, Mohananey, Anhad, Ilic, Anastasija, Tobin, Taylor, Wieting, John, Bortsova, Anna, Thacker, Phoebe, Wang, Emma, Caveness, Emily, Chiu, Justin, Sezener, Eren, Kaskasoli, Alex, Baker, Steven, Millican, Katie, Elhawaty, Mohamed, Aisopos, Kostas, Lebsack, Carl, Byrd, Nathan, Dai, Hanjun, Jia, Wenhao, Wiethoff, Matthew, Davoodi, Elnaz, Weston, Albert, Yagati, Lakshman, Ahuja, Arun, Gao, Isabel, Pundak, Golan, Zhang, Susan, Azzam, Michael, Sim, Khe Chai, Caelles, Sergi, Keeling, James, Sharma, Abhanshu, Swing, Andy, Li, YaGuang, Liu, Chenxi, Bostock, Carrie Grimes, Bansal, Yamini, Nado, Zachary, Anand, Ankesh, Lipschultz, Josh, Karmarkar, Abhijit, Proleev, Lev, Ittycheriah, Abe, Yeganeh, Soheil Hassas, Polovets, George, Faust, Aleksandra, Sun, Jiao, Rrustemi, Alban, Li, Pen, Shivanna, Rakesh, Liu, Jeremiah, Welty, Chris, Lebron, Federico, Baddepudi, Anirudh, Krause, Sebastian, Parisotto, Emilio, Soricut, Radu, Xu, Zheng, Bloxwich, Dawn, Johnson, Melvin, Neyshabur, Behnam, Mao-Jones, Justin, Wang, Renshen, Ramasesh, Vinay, Abbas, Zaheer, Guez, Arthur, Segal, Constant, Nguyen, Duc Dung, Svensson, James, Hou, Le, York, Sarah, Milan, Kieran, Bridgers, Sophie, Gworek, Wiktor, Tagliasacchi, Marco, Lee-Thorp, James, Chang, Michael, Guseynov, Alexey, Hartman, Ale Jakse, Kwong, Michael, Zhao, Ruizhe, Kashem, Sheleem, Cole, Elizabeth, Miech, Antoine, Tanburn, Richard, Phuong, Mary, Pavetic, Filip, Cevey, Sebastien, Comanescu, Ramona, Ives, Richard, Yang, Sherry, Du, Cosmo, Li, Bo, Zhang, Zizhao, Iinuma, Mariko, Hu, Clara Huiyi, Roy, Aurko, Bijwadia, Shaan, Zhu, Zhenkai, Martins, Danilo, Saputro, Rachel, Gergely, Anita, Zheng, Steven, Jia, Dawei, Antonoglou, Ioannis, Sadovsky, Adam, Gu, Shane, Bi, Yingying, Andreev, Alek, Samangooei, Sina, Khan, Mina, Kocisky, Tomas, Filos, Angelos, Kumar, Chintu, Bishop, Colton, Yu, Adams, Hodkinson, Sarah, Mittal, Sid, Shah, Premal, Moufarek, Alexandre, Cheng, Yong, Bloniarz, Adam, Lee, Jaehoon, Pejman, Pedram, Michel, Paul, Spencer, Stephen, Feinberg, Vladimir, Xiong, Xuehan, Savinov, Nikolay, Smith, Charlotte, Shakeri, Siamak, Tran, Dustin, Chesus, Mary, Bohnet, Bernd, Tucker, George, von Glehn, Tamara, Muir, Carrie, Mao, Yiran, Kazawa, Hideto, Slone, Ambrose, Soparkar, Kedar, Shrivastava, Disha, Cobon-Kerr, James, Sharman, Michael, Pavagadhi, Jay, Araya, Carlos, Misiunas, Karolis, Ghelani, Nimesh, Laskin, Michael, Barker, David, Li, Qiujia, Briukhov, Anton, Houlsby, Neil, Glaese, Mia, Lakshminarayanan, Balaji, Schucher, Nathan, Tang, Yunhao, Collins, Eli, Lim, Hyeontaek, Feng, Fangxiaoyu, Recasens, Adria, Lai, Guangda, Magni, Alberto, De Cao, Nicola, Siddhant, Aditya, Ashwood, Zoe, Orbay, Jordi, Dehghani, Mostafa, Brennan, Jenny, He, Yifan, Xu, Kelvin, Gao, Yang, Saroufim, Carl, Molloy, James, Wu, Xinyi, Arnold, Seb, Chang, Solomon, Schrittwieser, Julian, Buchatskaya, Elena, Radpour, Soroush, Polacek, Martin, Giordano, Skye, Bapna, Ankur, Tokumine, Simon, Hellendoorn, Vincent, Sottiaux, Thibault, Cogan, Sarah, Severyn, Aliaksei, Saleh, Mohammad, Thakoor, Shantanu, Shefey, Laurent, Qiao, Siyuan, Gaba, Meenu, Chang, Shuo-yiin, Swanson, Craig, Zhang, Biao, Lee, Benjamin, Rubenstein, Paul Kishan, Song, Gan, Kwiatkowski, Tom, Koop, Anna, Kannan, Ajay, Kao, David, Schuh, Parker, Stjerngren, Axel, Ghiasi, Golnaz, Gibson, Gena, Vilnis, Luke, Yuan, Ye, Ferreira, Felipe Tiengo, Kamath, Aishwarya, Klimenko, Ted, Franko, Ken, Xiao, Kefan, Bhattacharya, Indro, Patel, Miteyan, Wang, Rui, Morris, Alex, Strudel, Robin, Sharma, Vivek, Choy, Peter, Hashemi, Sayed Hadi, Landon, Jessica, Finkelstein, Mara, Jhakra, Priya, Frye, Justin, Barnes, Megan, Mauger, Matthew, Daun, Dennis, Baatarsukh, Khuslen, Tung, Matthew, Farhan, Wael, Michalewski, Henryk, Viola, Fabio, Quitry, Felix de Chaumont, Lan, Charline Le, Hudson, Tom, Wang, Qingze, Fischer, Felix, Zheng, Ivy, White, Elspeth, Dragan, Anca, Alayrac, Jean-baptiste, Ni, Eric, Pritzel, Alexander, Iwanicki, Adam, Isard, Michael, Bulanova, Anna, Zilka, Lukas, Dyer, Ethan, Sachan, Devendra, Srinivasan, Srivatsan, Muckenhirn, Hannah, Cai, Honglong, Mandhane, Amol, Tariq, Mukarram, Rae, Jack W., Wang, Gary, Ayoub, Kareem, FitzGerald, Nicholas, Zhao, Yao, Han, Woohyun, Alberti, Chris, Garrette, Dan, Krishnakumar, Kashyap, Gimenez, Mai, Levskaya, Anselm, Sohn, Daniel, Matak, Josip, Iturrate, Inaki, Chang, Michael B., Xiang, Jackie, Cao, Yuan, Ranka, Nishant, Brown, Geoff, Hutter, Adrian, Mirrokni, Vahab, Chen, Nanxin, Yao, Kaisheng, Egyed, Zoltan, Galilee, Francois, Liechty, Tyler, Kallakuri, Praveen, Palmer, Evan, Ghemawat, Sanjay, Liu, Jasmine, Tao, David, Thornton, Chloe, Green, Tim, Jasarevic, Mimi, Lin, Sharon, Cotruta, Victor, Tan, Yi-Xuan, Fiedel, Noah, Yu, Hongkun, Chi, Ed, Neitz, Alexander, Heitkaemper, Jens, Sinha, Anu, Zhou, Denny, Sun, Yi, Kaed, Charbel, Hulse, Brice, Mishra, Swaroop, Georgaki, Maria, Kudugunta, Sneha, Farabet, Clement, Shafran, Izhak, Vlasic, Daniel, Tsitsulin, Anton, Ananthanarayanan, Rajagopal, Carin, Alen, Su, Guolong, Sun, Pei, V, Shashank, Carvajal, Gabriel, Broder, Josef, Comsa, Iulia, Repina, Alena, Wong, William, Chen, Warren Weilun, Hawkins, Peter, Filonov, Egor, Loher, Lucia, Hirnschall, Christoph, Wang, Weiyi, Ye, Jingchen, Burns, Andrea, Cate, Hardie, Wright, Diana Gage, Piccinini, Federico, Zhang, Lei, Lin, Chu-Cheng, Gog, Ionel, Kulizhskaya, Yana, Sreevatsa, Ashwin, Song, Shuang, Cobo, Luis C., Iyer, Anand, Tekur, Chetan, Garrido, Guillermo, Xiao, Zhuyun, Kemp, Rupert, Zheng, Huaixiu Steven, Li, Hui, Agarwal, Ananth, Ngani, Christel, Goshvadi, Kati, Santamaria-Fernandez, Rebeca, Fica, Wojciech, Chen, Xinyun, Gorgolewski, Chris, Sun, Sean, Garg, Roopal, Ye, Xinyu, Eslami, S. M. Ali, Hua, Nan, Simon, Jon, Joshi, Pratik, Kim, Yelin, Tenney, Ian, Potluri, Sahitya, Thiet, Lam Nguyen, Yuan, Quan, Luisier, Florian, Chronopoulou, Alexandra, Scellato, Salvatore, Srinivasan, Praveen, Chen, Minmin, Koverkathu, Vinod, Dalibard, Valentin, Xu, Yaming, Saeta, Brennan, Anderson, Keith, Sellam, Thibault, Fernando, Nick, Huot, Fantine, Jung, Junehyuk, Varadarajan, Mani, Quinn, Michael, Raul, Amit, Le, Maigo, Habalov, Ruslan, Clark, Jon, Jalan, Komal, Bullard, Kalesha, Singhal, Achintya, Luong, Thang, Wang, Boyu, Rajayogam, Sujeevan, Eisenschlos, Julian, Jia, Johnson, Finchelstein, Daniel, Yakubovich, Alex, Balle, Daniel, Fink, Michael, Agarwal, Sameer, Li, Jing, Dvijotham, Dj, Pal, Shalini, Kang, Kai, Konzelmann, Jaclyn, Beattie, Jennifer, Dousse, Olivier, Wu, Diane, Crocker, Remi, Elkind, Chen, Jonnalagadda, Siddhartha Reddy, Lee, Jong, Holtmann-Rice, Dan, Kallarackal, Krystal, Liu, Rosanne, Vnukov, Denis, Vats, Neera, Invernizzi, Luca, Jafari, Mohsen, Zhou, Huanjie, Taylor, Lilly, Prendki, Jennifer, Wu, Marcus, Eccles, Tom, Liu, Tianqi, Kopparapu, Kavya, Beaufays, Francoise, Angermueller, Christof, Marzoca, Andreea, Sarcar, Shourya, Dib, Hilal, Stanway, Jeff, Perbet, Frank, Trdin, Nejc, Sterneck, Rachel, Khorlin, Andrey, Li, Dinghua, Wu, Xihui, Goenka, Sonam, Madras, David, Goldshtein, Sasha, Gierke, Willi, Zhou, Tong, Liu, Yaxin, Liang, Yannie, White, Anais, Li, Yunjie, Singh, Shreya, Bahargam, Sanaz, Epstein, Mark, Basu, Sujoy, Lao, Li, Ozturel, Adnan, Crous, Carl, Zhai, Alex, Lu, Han, Tung, Zora, Gaur, Neeraj, Walton, Alanna, Dixon, Lucas, Zhang, Ming, Globerson, Amir, Uy, Grant, Bolt, Andrew, Wiles, Olivia, Nasr, Milad, Shumailov, Ilia, Selvi, Marco, Piccinno, Francesco, Aguilar, Ricardo, McCarthy, Sara, Khalman, Misha, Shukla, Mrinal, Galic, Vlado, Carpenter, John, Villela, Kevin, Zhang, Haibin, Richardson, Harry, Martens, James, Bosnjak, Matko, Belle, Shreyas Rammohan, Seibert, Jeff, Alnahlawi, Mahmoud, McWilliams, Brian, Singh, Sankalp, Louis, Annie, Ding, Wen, Popovici, Dan, Simicich, Lenin, Knight, Laura, Mehta, Pulkit, Gupta, Nishesh, Shi, Chongyang, Fatehi, Saaber, Mitrovic, Jovana, Grills, Alex, Pagadora, Joseph, Petrova, Dessie, Eisenbud, Danielle, Zhang, Zhishuai, Yates, Damion, Mittal, Bhavishya, Tripuraneni, Nilesh, Assael, Yannis, Brovelli, Thomas, Jain, Prateek, Velimirovic, Mihajlo, Akbulut, Canfer, Mu, Jiaqi, Macherey, Wolfgang, Kumar, Ravin, Xu, Jun, Qureshi, Haroon, Comanici, Gheorghe, Wiesner, Jeremy, Gong, Zhitao, Ruddock, Anton, Bauer, Matthias, Felt, Nick, GP, Anirudh, Arnab, Anurag, Zelle, Dustin, Rothfuss, Jonas, Rosgen, Bill, Shenoy, Ashish, Seybold, Bryan, Li, Xinjian, Mudigonda, Jayaram, Erdogan, Goker, Xia, Jiawei, Simsa, Jiri, Michi, Andrea, Yao, Yi, Yew, Christopher, Kan, Steven, Caswell, Isaac, Radebaugh, Carey, Elisseeff, Andre, Valenzuela, Pedro, McKinney, Kay, Paterson, Kim, Cui, Albert, Latorre-Chimoto, Eri, Kim, Solomon, Zeng, William, Durden, Ken, Ponnapalli, Priya, Sosea, Tiberiu, Choquette-Choo, Christopher A., Manyika, James, Robenek, Brona, Vashisht, Harsha, Pereira, Sebastien, Lam, Hoi, Velic, Marko, Owusu-Afriyie, Denese, Lee, Katherine, Bolukbasi, Tolga, Parrish, Alicia, Lu, Shawn, Park, Jane, Venkatraman, Balaji, Talbert, Alice, Rosique, Lambert, Cheng, Yuchung, Sozanschi, Andrei, Paszke, Adam, Kumar, Praveen, Austin, Jessica, Li, Lu, Salama, Khalid, Kim, Wooyeol, Dukkipati, Nandita, Baryshnikov, Anthony, Kaplanis, Christos, Sheng, XiangHai, Chervonyi, Yuri, Unlu, Caglar, Casas, Diego de Las, Askham, Harry, Tunyasuvunakool, Kathryn, Gimeno, Felix, Poder, Siim, Kwak, Chester, Miecnikowski, Matt, Dimitriev, Alek, Parisi, Aaron, Liu, Dangyi, Tsai, Tomy, Shevlane, Toby, Kouridi, Christina, Garmon, Drew, Goedeckemeyer, Adrian, Brown, Adam R., Vijayakumar, Anitha, Elqursh, Ali, Jazayeri, Sadegh, Huang, Jin, Carthy, Sara Mc, Hoover, Jay, Kim, Lucy, Kumar, Sandeep, Chen, Wei, Biles, Courtney, Bingham, Garrett, Rosen, Evan, Wang, Lisa, Tan, Qijun, Engel, David, Pongetti, Francesco, de Cesare, Dario, Hwang, Dongseong, Yu, Lily, Pullman, Jennifer, Narayanan, Srini, Levin, Kyle, Gopal, Siddharth, Li, Megan, Aharoni, Asaf, Trinh, Trieu, Lo, Jessica, Casagrande, Norman, Vij, Roopali, Matthey, Loic, Ramadhana, Bramandia, Matthews, Austin, Carey, CJ, Johnson, Matthew, Goranova, Kremena, Shah, Rohin, Ashraf, Shereen, Dasgupta, Kingshuk, Larsen, Rasmus, Wang, Yicheng, Vuyyuru, Manish Reddy, Jiang, Chong, Ijazi, Joana, Osawa, Kazuki, Smith, Celine, Boppana, Ramya Sree, Bilal, Taylan, Koizumi, Yuma, Xu, Ying, Altun, Yasemin, Shabat, Nir, Bariach, Ben, Korchemniy, Alex, Choo, Kiam, Ronneberger, Olaf, Iwuanyanwu, Chimezie, Zhao, Shubin, Soergel, David, Hsieh, Cho-Jui, Cai, Irene, Iqbal, Shariq, Sundermeyer, Martin, Chen, Zhe, Bursztein, Elie, Malaviya, Chaitanya, Biadsy, Fadi, Shroff, Prakash, Dhillon, Inderjit, Latkar, Tejasi, Dyer, Chris, Forbes, Hannah, Nicosia, Massimo, Nikolaev, Vitaly, Greene, Somer, Georgiev, Marin, Wang, Pidong, Martin, Nina, Sedghi, Hanie, Zhang, John, Banzal, Praseem, Fritz, Doug, Rao, Vikram, Wang, Xuezhi, Zhang, Jiageng, Patraucean, Viorica, Du, Dayou, Mordatch, Igor, Jurin, Ivan, Liu, Lewis, Dubey, Ayush, Mohan, Abhi, Nowakowski, Janek, Ion, Vlad-Doru, Wei, Nan, Tojo, Reiko, Raad, Maria Abi, Hudson, Drew A., Keshava, Vaishakh, Agrawal, Shubham, Ramirez, Kevin, Wu, Zhichun, Nguyen, Hoang, Liu, Ji, Sewak, Madhavi, Petrini, Bryce, Choi, DongHyun, Philips, Ivan, Wang, Ziyue, Bica, Ioana, Garg, Ankush, Wilkiewicz, Jarek, Agrawal, Priyanka, Guo, Danhao, Xue, Emily, Shaik, Naseer, Leach, Andrew, Khan, Sadh MNM, Wiesinger, Julia, Jerome, Sammy, Chakladar, Abhishek, Wang, Alek Wenjiao, Ornduff, Tina, Abu, Folake, Ghaffarkhah, Alireza, Wainwright, Marcus, Cortes, Mario, Liu, Frederick, Maynez, Joshua, Terzis, Andreas, Samangouei, Pouya, Mansour, Riham, Kępa, Tomasz, Aubet, François-Xavier, Algymr, Anton, Banica, Dan, Weisz, Agoston, Orban, Andras, Senges, Alexandre, Andrejczuk, Ewa, Geller, Mark, Santo, Niccolo Dal, Anklin, Valentin, Merey, Majd Al, Baeuml, Martin, Strohman, Trevor, Bai, Junwen, Petrov, Slav, Wu, Yonghui, Hassabis, Demis, Kavukcuoglu, Koray, Dean, Jeffrey, and Vinyals, Oriol
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence - Abstract
In this report, we introduce the Gemini 1.5 family of models, representing the next generation of highly compute-efficient multimodal models capable of recalling and reasoning over fine-grained information from millions of tokens of context, including multiple long documents and hours of video and audio. The family includes two new models: (1) an updated Gemini 1.5 Pro, which exceeds the February version on the great majority of capabilities and benchmarks; (2) Gemini 1.5 Flash, a more lightweight variant designed for efficiency with minimal regression in quality. Gemini 1.5 models achieve near-perfect recall on long-context retrieval tasks across modalities, improve the state-of-the-art in long-document QA, long-video QA and long-context ASR, and match or surpass Gemini 1.0 Ultra's state-of-the-art performance across a broad set of benchmarks. Studying the limits of Gemini 1.5's long-context ability, we find continued improvement in next-token prediction and near-perfect retrieval (>99%) up to at least 10M tokens, a generational leap over existing models such as Claude 3.0 (200k) and GPT-4 Turbo (128k). Finally, we highlight real-world use cases, such as Gemini 1.5 collaborating with professionals on completing their tasks achieving 26 to 75% time savings across 10 different job categories, as well as surprising new capabilities of large language models at the frontier; when given a grammar manual for Kalamang, a language with fewer than 200 speakers worldwide, the model learns to translate English to Kalamang at a similar level to a person who learned from the same content.
- Published
- 2024
27. SciAssess: Benchmarking LLM Proficiency in Scientific Literature Analysis
- Author
-
Cai, Hengxing, Cai, Xiaochen, Chang, Junhan, Li, Sihang, Yao, Lin, Wang, Changxin, Gao, Zhifeng, Wang, Hongshuai, Li, Yongge, Lin, Mujie, Yang, Shuwen, Wang, Jiankun, Xu, Mingjun, Huang, Jin, Fang, Xi, Zhuang, Jiaxi, Yin, Yuqi, Li, Yaqi, Chen, Changhong, Cheng, Zheng, Zhao, Zifeng, Zhang, Linfeng, and Ke, Guolin
- Subjects
Computer Science - Computation and Language - Abstract
Recent breakthroughs in Large Language Models (LLMs) have revolutionized scientific literature analysis. However, existing benchmarks fail to adequately evaluate the proficiency of LLMs in this domain, particularly in scenarios requiring higher-level abilities beyond mere memorization and the handling of multimodal data. In response to this gap, we introduce SciAssess, a benchmark specifically designed for the comprehensive evaluation of LLMs in scientific literature analysis. It aims to thoroughly assess the efficacy of LLMs by evaluating their capabilities in Memorization (L1), Comprehension (L2), and Analysis \& Reasoning (L3). It encompasses a variety of tasks drawn from diverse scientific fields, including biology, chemistry, material, and medicine. To ensure the reliability of SciAssess, rigorous quality control measures have been implemented, ensuring accuracy, anonymization, and compliance with copyright standards. SciAssess evaluates 11 LLMs, highlighting their strengths and areas for improvement. We hope this evaluation supports the ongoing development of LLM applications in scientific literature analysis. SciAssess and its resources are available at \url{https://github.com/sci-assess/SciAssess}.
- Published
- 2024
28. Process optimization of 1-cyanocyclohexaneacetic acid hydrogenation using response surface methodology
- Author
-
Xiong, Neng, Jiang, Lin-Li, Chen, Jia-Yu, Lin, Lei, Huang, Jin-Rong, Shen, Qi, Xue, Ya-Ping, and Zheng, Yu-Guo
- Published
- 2024
- Full Text
- View/download PDF
29. TSA-Net: a temporal knowledge graph completion method with temporal-structural adaptation
- Author
-
Xie, Ruzhong, Ruan, Ke, Huang, Bosong, Yu, Weihao, Xiao, Jing, and Huang, Jin
- Published
- 2024
- Full Text
- View/download PDF
30. Age and size dependent diagnostic reference levels and achievable doses for 48 types of CT examinations: a multicenter cross-sectional study of one million CT examinations in China
- Author
-
Wu, Shiyao, Zhou, Changsheng, Xu, Yikai, Qiao, Wenjun, Xia, Liming, Li, Yang, Huang, Chao, He, Haoqiang, Deng, Dele, Dai, Wei, Huang, Jin, Zhong, Nengzhi, Yang, Guifen, Zhang, Longjiang, Xie, Chuanmiao, and Lu, Guangming
- Published
- 2024
- Full Text
- View/download PDF
31. Validating a measure of computational thinking skills in Chinese kindergartners
- Author
-
Geng, Zuofei, Zeng, Bei, Islam, A. Y. M. Atiquil, Zhang, Xuanyi, and Huang, Jin
- Published
- 2024
- Full Text
- View/download PDF
32. SACANet: end-to-end self-attention-based network for 3D clothing animation
- Author
-
Chen, Yunxi, Cao, Yuanjie, Fang, Fei, Huang, Jin, Hu, Xinrong, He, Ruhan, and Zhang, Junjie
- Published
- 2024
- Full Text
- View/download PDF
33. The fast Euler-Maruyama method for solving multiterm Caputo fractional stochastic delay integro-differential equations
- Author
-
Guo, Huijiao, Huang, Jin, Yang, Yi, and Zhang, Xueli
- Published
- 2024
- Full Text
- View/download PDF
34. Effectiveness of Lilly Connected Care Program (LCCP) App-Based Diabetes Education for Patients With Type 2 Diabetes Treated With Insulin: Retrospective Real-World Study
- Author
-
Zhang, Yiyu, Liu, Chaoyuan, Luo, Shuoming, Huang, Jin, Li, Xia, and Zhou, Zhiguang
- Subjects
Information technology ,T58.5-58.64 ,Public aspects of medicine ,RA1-1270 - Abstract
BackgroundDiabetes poses heavy economic and social burdens worldwide. Mobile apps show great potential for diabetes self-management education. However, there is limited evidence for the effectiveness of providing general diabetes education through mobile apps. ObjectiveThe aim of this study was to clarify the effectiveness of Lilly Connected Care Program (LCCP) app-based diabetes education for glycemic control. MethodsThis retrospective cohort study included patients with diabetes recruited to the LCCP platform from September 1, 2018, to May 31, 2019. Each patient was followed for 12 weeks. According to the number of diabetes education courses they had completed, the patients were divided into the following three groups: group A (0-4 courses), group B (5-29 courses), and group C (≥30 courses). The main outcomes were the change in blood glucose at the 12th week compared with baseline and the differences in blood glucose at the 12th week among the three groups. The associations of the number of diabetes education courses completed with the average blood glucose and frequency of self-monitoring of blood glucose (SMBG) at the 12th week were assessed by multivariate linear regression analyses controlling for other confounding covariates. Univariate and multivariate linear regression analyses were used to assess factors influencing patients’ engagement in the diabetes education courses. ResultsA total of 5011 participants were enrolled. Their mean fasting blood glucose (FBG) and postprandial blood glucose (PBG) were significantly lower at the 12th week than at baseline (FBG, 7.46 [standard deviation (SD) 1.95] vs 7.79 [SD 2.18] mmol/L, P
- Published
- 2020
- Full Text
- View/download PDF
35. Insomnia impairs muscle function via regulating protein degradation and muscle clock
- Author
-
Ouyang, Hui, Jiang, Hong, Huang, Jin, and Liu, Zunjing
- Subjects
Quantitative Biology - Neurons and Cognition - Abstract
Background: Insomnia makes people more physically unable of doing daily duties, which results in a lack of strength, leads to lacking in strength. However, the effects of insomnia on muscle function have not yet been thoroughly investigated. So, the objectives of this study were to clarify how insomnia contributes to the decrease of muscular function and to investigate the mechanisms behind this phenomenon. Methods: To understand how insomnia influence muscle function, we analyzed the expression level of factors associated with muscle protein degradation, muscle protein synthesis , protein synthesis and degradation pathways and muscle clock. Results: The results showed that lower BMI and grip strength were observed in insomnia patients. The mice in the sleep deprivation(SD) group saw a 7.01 g loss in body mass. The SD group's tibialis anterior and gastrocnemius muscle mass decreased after 96 h of SD). The grip strength reduced in SD group. Using the RT-PCR approaches, we found a significant increase in muscle degradation factors expression in SD group versus normal control group. Conclusions: Insomnia can impair muscle function. The mechanism may be associated with the increased expression of muscle degradation related factors , as well as the abnormal expression of Clock gene.
- Published
- 2023
36. Nutrients uptake and removal characteristics by high-yielding sugarcane grown in Guangxi, China
- Author
-
Ou, Hui-Ping, Xie, Ru-Lin, Huang, Jin-Sheng, and Zeng, Yan
- Published
- 2017
- Full Text
- View/download PDF
37. Prediction of the Climatically Suitable Areas of Rice in China Based on Optimized MaxEnt Model
- Author
-
Zhao, Chenyu, Zhang, Fangmin, Huang, Jin, Zhang, Qian, Lu, Yanyu, and Cao, Wen
- Published
- 2024
- Full Text
- View/download PDF
38. Vagus nerve stimulation (VNS) preventing postoperative cognitive dysfunction (POCD): two potential mechanisms in cognitive function
- Author
-
Xie, Zi-Feng, Wang, Sheng-Yu, Gao, Yuan, Zhang, Yi-Dan, Han, Ya-Nan, Huang, Jin, Gao, Mei-Na, and Wang, Chun-Guang
- Published
- 2024
- Full Text
- View/download PDF
39. Effect of process parameters on microstructure and mechanical properties of a nickel-aluminum-bronze alloy fabricated by laser powder bed fusion
- Author
-
Han, Chang-jun, Zou, Yu-jin, Hu, Gao-ling, Dong, Zhi, Li, Kai, Huang, Jin-miao, Li, Bo-yuan, Zhou, Kun, Yang, Yong-qiang, and Wang, Di
- Published
- 2024
- Full Text
- View/download PDF
40. Effect of temperature and reaction path interaction on fluidization reduction kinetics of iron ore powder
- Author
-
Zhu, Guo-min, Hu, Ming-wei, Dou, An-nan, Huang, Jin-yu, Ding, Jing, and Xu, Qi-yan
- Published
- 2024
- Full Text
- View/download PDF
41. Immobilized Cellulase on NH2-MIL-88(Fe) and Its Performance as a Biocatalyst
- Author
-
Jiang, Jing, Gong, Xiaowu, Li, Tiantian, Huang, Jin, Zhou, Na, and Jia, Xin
- Published
- 2024
- Full Text
- View/download PDF
42. Two-step continuous alkali pretreatment for fractionation of the carbohydrates from bamboo and cotton stalk
- Author
-
Mou, Hongyan, Tang, Lv, Liu, Yibei, Huang, Jin, Feng, Lu, Wu, Xiao, and Wang, Zhiwei
- Published
- 2024
- Full Text
- View/download PDF
43. Mitochondrial Treatment Improves Cognitive Impairment Induced by Lipopolysaccharide in Mice
- Author
-
Yan, Qiu-Qing, Liu, Tian-Long, Liu, Ling-Ling, Wei, Yan-Su, Zhao, Yuan-Dan, Yu, Chao, Zhong, Zhen-Guo, Huang, Jin-Lan, and Wu, Deng-Pan
- Published
- 2024
- Full Text
- View/download PDF
44. Dual-branch dilated context convolutional for table detection transformer in the document images
- Author
-
Ni, Ying, Wang, Xiaoli, Peng, Hanghang, Li, Yonzhi, Wang, Jinyang, Li, Haoxuan, and Huang, Jin
- Published
- 2024
- Full Text
- View/download PDF
45. Embedded oxide clusters stabilize sub-2 nm Pt nanoparticles for highly durable fuel cells
- Author
-
Peng, Bosi, Liu, Zeyan, Sementa, Luca, Jia, Qingying, Sun, Qiang, Segre, Carlo U., Liu, Ershuai, Xu, Mingjie, Tsai, Yu-Han (Joseph), Yan, Xingxu, Zhao, Zipeng, Huang, Jin, Pan, Xiaoqing, Duan, Xiangfeng, Fortunelli, Alessandro, and Huang, Yu
- Published
- 2024
- Full Text
- View/download PDF
46. A data-driven approach for the guided regulation of exposed facets in nanoparticles
- Author
-
Ye, Zihao, Shen, Bo, Kang, Dohun, Shen, Jiahong, Huang, Jin, Wang, Zhe, Huang, Liliang, Wolverton, Christopher M., and Mirkin, Chad A.
- Published
- 2024
- Full Text
- View/download PDF
47. Engineering self-catabolic DNAzyme nanospheres for synergistic anticancer therapy
- Author
-
Chen, Yu, Guo, Yu, Wang, Jiaoli, Liu, Ruiting, Yang, Xiaohai, Wang, Kemin, Pu, Ying, Shi, Hui, and Huang, Jin
- Published
- 2024
- Full Text
- View/download PDF
48. A Predictive Model for Graft Failure in Femtosecond Laser-Assisted Penetrating Keratoplasty Among Chinese Patients: A 2-Year Study
- Author
-
Ma, Junxin, Cao, Xueqian, Liu, Yang, Huang, Jin, Gong, Yuting, Pan, Xinyu, Li, Zhongguo, and Wang, Linnong
- Published
- 2024
- Full Text
- View/download PDF
49. Obstetrical and neonatal outcomes after vitrified-warmed blastocyst transfer in day 1 rescue intracytoplasmic sperm injection cycles: a retrospective cohort study
- Author
-
Li, Ming, Zhang, Nan, Huang, Jin, Li, Qin, Li, JunSheng, Li, Rong, Liu, Ping, and Qiao, Jie
- Published
- 2024
- Full Text
- View/download PDF
50. Bright InP quantum dots by Ga-doping for red emitters
- Author
-
Song, Kai-Zheng, He, Xiao-Hang, Chen, Zhe-Yong, Tang, Ge, Huang, Jin-Zhao, and Jiang, Feng-Lei
- Published
- 2024
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.