13,387 results on '"HUANG, LEI"'
Search Results
2. ENZYME CATALYSIS AND DECOLOURISATION OF BRILLIANT REACTIVE RED X-3B BY AZOREDUCTASE FROM A NEWLY ISOLATED PSEUDOMONAS PUTIDA WLY
- Author
-
Cai, Zhiqiang, Huang, Lei, He, Yucai, Shi, Sai, Zhao, Xiyue, Wang, Liqun, and Wang, Li
- Published
- 2022
3. A Metal Production Center on the Southwest Frontier of the Han Empire: An Archaeometallurgical Study of the Heimajing Cemetery Site in Gejiu, Yunnan Province, China
- Author
-
LI, Yingfu, HAN, Dong, YANG, Sheng, HUANG, Lei, YANG, Ge, and LI, Yuniu
- Published
- 2022
- Full Text
- View/download PDF
4. Improving Contextual Faithfulness of Large Language Models via Retrieval Heads-Induced Optimization
- Author
-
Huang, Lei, Feng, Xiaocheng, Ma, Weitao, Fan, Yuchun, Feng, Xiachong, Ye, Yangfan, Zhong, Weihong, Gu, Yuxuan, Wang, Baoxin, Wu, Dayong, Hu, Guoping, and Qin, Bing
- Subjects
Computer Science - Computation and Language - Abstract
Ensuring contextual faithfulness in retrieval-augmented large language models (LLMs) is crucial for building trustworthy information-seeking systems, particularly in long-form question-answering (LFQA) scenarios. In this work, we identify a salient correlation between LFQA faithfulness and retrieval heads, a set of attention heads responsible for retrieving contextual information. Leveraging this insight, we propose RHIO, a framework designed to teach LLMs to explicitly discriminate between faithful and unfaithful generations. RHIO first augments unfaithful samples that simulate realistic model-intrinsic errors by selectively masking retrieval heads. Then, these samples are incorporated into joint training, enabling the model to distinguish unfaithful outputs from faithful ones conditioned on control tokens. Furthermore, these control tokens are leveraged to self-induce contrastive outputs, amplifying their difference through contrastive decoding. Additionally, to facilitate the evaluation of contextual faithfulness, we also introduce GroundBench, a comprehensive benchmark compiled from five existing LFQA datasets. Extensive experimental results on GroundBench demonstrate that RHIO significantly improves faithfulness, even outperforming GPT-4o., Comment: Submitted to ARR October 2024
- Published
- 2025
5. Joint Beamforming and Position Optimization for Fluid RIS-aided ISAC Systems
- Author
-
Ye, Junjie, Zhang, Peichang, Li, Xiao-Peng, Huang, Lei, and Liu, Yuanwei
- Subjects
Electrical Engineering and Systems Science - Signal Processing - Abstract
A fluid reconfigurable intelligent surface (fRIS)-aided integrated sensing and communications (ISAC) system is proposed to enhance multi-target sensing and multi-user communication. Unlike the conventional RIS, the fRIS incorporates movable elements whose positions can be flexibly adjusted to provide extra spatial degrees of freedom. In this system, a joint optimization problem is formulated to minimize sensing beampattern mismatch and communication symbol estimation error by optimizing the symbol estimator, transmit beamformer, fRIS phase shifts, and element positions. To solve this problem, an algorithm based on alternating minimization is devised, where subproblems are solved leveraging augmented Lagrangian method, quadratic programming, semidefinite-relaxation, and majorization-minimization techniques. A key challenge exists that the fRIS element positions affect both the incident and reflective channels, leading to the high-order composite functions regarding the positions. As a remedy, it is proved that the high-order terms can be transformed to linear and linear-difference forms using the characteristics of fRIS and structural channels, which facilitates the position optimization. Numerical results validate the effectiveness of the proposed scheme as compared to the conventional RIS-aided ISAC systems and other benchmarks., Comment: 13 pages, 10 figures, has submitted to an IEEE journal for possible publication
- Published
- 2025
6. Joint Antenna Selection and Beamforming Design for Active RIS-aided ISAC Systems
- Author
-
Ma, Wei, Zhang, Peichang, Ye, Junjie, Guan, Rouyang, Li, Xiao-Peng, and Huang, Lei
- Subjects
Computer Science - Information Theory ,Electrical Engineering and Systems Science - Signal Processing - Abstract
Active reconfigurable intelligent surface (A-RIS) aided integrated sensing and communications (ISAC) system has been considered as a promising paradigm to improve spectrum efficiency. However, massive energy-hungry radio frequency (RF) chains hinder its large-scale deployment. To address this issue, an A-RIS-aided ISAC system with antenna selection (AS) is proposed in this work, where a target is sensed while multiple communication users are served with specifically selected antennas. Specifically, a cuckoo search-based scheme is first utilized to select the antennas associated with high-gain channels. Subsequently, with the properly selected antennas, the weighted sum-rate (WSR) of the system is optimized under the condition of radar probing power level, power budget for the A-RIS and transmitter. To solve the highly non-convex optimization problem, we develop an efficient algorithm based on weighted minimum mean square error (WMMSE) and fractional programming (FP). Simulation results show that the proposed AS scheme and the algorithm are effective, which reduce the number of RF chains without significant performance degradation.
- Published
- 2025
7. The putative center in NGC 1052
- Author
-
Baczko, Anne-Kathrin, Kadler, Matthias, Ros, Eduardo, Fromm, Christian M., Wielgus, Maciek, Perucho, Manel, Krichbaum, Thomas P., Baloković, Mislav, Blackburn, Lindy, Chan, Chi-kwan, Issaoun, Sara, Janssen, Michael, Ricci, Luca, Akiyama, Kazunori, Albentosa-Ruíz, Ezequiel, Alberdi, Antxon, Alef, Walter, Algaba, Juan Carlos, Anantua, Richard, Asada, Keiichi, Azulay, Rebecca, Bach, Uwe, Ball, David, Bandyopadhyay, Bidisha, Barrett, John, Bauböck, Michi, Benson, Bradford A., Bintley, Dan, Blundell, Raymond, Bouman, Katherine L., Bower, Geoffrey C., Boyce, Hope, Bremer, Michael, Brinkerink, Christiaan D., Brissenden, Roger, Britzen, Silke, Broderick, Avery E., Broguiere, Dominique, Bronzwaer, Thomas, Bustamante, Sandra, Byun, Do-Young, Carlstrom, John E., Ceccobello, Chiara, Chael, Andrew, Chang, Dominic O., Chatterjee, Koushik, Chatterjee, Shami, Chen, Ming-Tang, Chen, Yongjun, Cheng, Xiaopeng, Cho, Ilje, Christian, Pierre, Conroy, Nicholas S., Conway, John E., Cordes, James M., Crawford, Thomas M., Crew, Geoffrey B., Cruz-Osorio, Alejandro, Cui, Yuzhu, Dahale, Rohan, Davelaar, Jordy, De Laurentis, Mariafelicia, Deane, Roger, Dempsey, Jessica, Desvignes, Gregory, Dexter, Jason, Dhruv, Vedant, Dihingia, Indu K., Doeleman, Sheperd S., Dougall, Sean Taylor, Dzib, Sergio A., Eatough, Ralph P., Emami, Razieh, Falcke, Heino, Farah, Joseph, Fish, Vincent L., Fomalont, Edward, Ford, H. Alyson, Foschi, Marianna, Fraga-Encinas, Raquel, Freeman, William T., Friberg, Per, Fuentes, Antonio, Galison, Peter, Gammie, Charles F., García, Roberto, Gentaz, Olivier, Georgiev, Boris, Goddi, Ciriaco, Gold, Roman, Gómez-Ruiz, Arturo I., Gómez, José L., Gu, Minfeng, Gurwell, Mark, Hada, Kazuhiro, Haggard, Daryl, Haworth, Kari, Hecht, Michael H., Hesper, Ronald, Heumann, Dirk, Ho, Luis C., Ho, Paul, Honma, Mareki, Huang, Chih-Wei L., Huang, Lei, Hughes, David H., Impellizzeri, C. M. Violette, Inoue, Makoto, James, David J., Jannuzi, Buell T., Jeter, Britton, Jiang, Wu, Jiménez-Rosales, Alejandra, Johnson, Michael D., Jorstad, Svetlana, Joshi, Abhishek V., Jung, Taehyun, Karami, Mansour, Karuppusamy, Ramesh, Kawashima, Tomohisa, Keating, Garrett K., Kettenis, Mark, Kim, Dong-Jin, Kim, Jae-Young, Kim, Jongsoo, Kim, Junhan, Kino, Motoki, Koay, Jun Yi, Kocherlakota, Prashant, Kofuji, Yutaro, Koyama, Shoko, Kramer, Carsten, Kramer, Joana A., Kramer, Michael, Kuo, Cheng-Yu, La Bella, Noemi, Lauer, Tod R., Lee, Daeyoung, Lee, Sang-Sung, Leung, Po Kin, Levis, Aviad, Li, Zhiyuan, Lico, Rocco, Lindahl, Greg, Lindqvist, Michael, Lisakov, Mikhail, Liu, Jun, Liu, Kuo, Liuzzo, Elisabetta, Lo, Wen-Ping, Lobanov, Andrei P., Loinard, Laurent, Lonsdale, Colin J., Lowitz, Amy E., Lu, Ru-Sen, MacDonald, Nicholas R., Mao, Jirong, Marchili, Nicola, Markoff, Sera, Marrone, Daniel P., Marscher, Alan P., Martí-Vidal, Iván, Matsushita, Satoki, Matthews, Lynn D., Medeiros, Lia, Menten, Karl M., Michalik, Daniel, Mizuno, Izumi, Mizuno, Yosuke, Moran, James M., Moriyama, Kotaro, Moscibrodzka, Monika, Mulaudzi, Wanga, Müller, Cornelia, Müller, Hendrik, Mus, Alejandro, Musoke, Gibwa, Myserlis, Ioannis, Nadolski, Andrew, Nagai, Hiroshi, Nagar, Neil M., Nair, Dhanya G., Nakamura, Masanori, Narayanan, Gopal, Natarajan, Iniyan, Nathanail, Antonios, Fuentes, Santiago Navarro, Neilsen, Joey, Neri, Roberto, Ni, Chunchong, Noutsos, Aristeidis, Nowak, Michael A., Oh, Junghwan, Okino, Hiroki, Sánchez, Héctor Raúl Olivares, Ortiz-León, Gisela N., Oyama, Tomoaki, Özel, Feryal, Palumbo, Daniel C. M., Paraschos, Georgios Filippos, Park, Jongho, Parsons, Harriet, Patel, Nimesh, Pen, Ue-Li, Pesce, Dominic W., Piétu, Vincent, Plambeck, Richard, PopStefanija, Aleksandar, Porth, Oliver, Pötzl, Felix M., Prather, Ben, Preciado-López, Jorge A., Principe, Giacomo, Psaltis, Dimitrios, Pu, Hung-Yi, Ramakrishnan, Venkatessh, Rao, Ramprasad, Rawlings, Mark G., Raymond, Alexander W., Ricarte, Angelo, Ripperda, Bart, Roelofs, Freek, Rogers, Alan, Romero-Cañizales, Cristina, Roshanineshat, Arash, Rottmann, Helge, Roy, Alan L., Ruiz, Ignacio, Ruszczyk, Chet, Rygl, Kazi L. J., Sánchez, Salvador, Sánchez-Argüelles, David, Sánchez-Portal, Miguel, Sasada, Mahito, Satapathy, Kaushik, Savolainen, Tuomas, Schloerb, F. Peter, Schonfeld, Jonathan, Schuster, Karl-Friedrich, Shao, Lijing, Shen, Zhiqiang, Small, Des, Sohn, Bong Won, SooHoo, Jason, Salas, León David Sosapanta, Souccar, Kamal, Stanway, Joshua S., Sun, He, Tazaki, Fumie, Tetarenko, Alexandra J., Tiede, Paul, Tilanus, Remo P. J., Titus, Michael, Torne, Pablo, Toscano, Teresa, Traianou, Efthalia, Trent, Tyler, Trippe, Sascha, Turk, Matthew, van Bemmel, Ilse, van Langevelde, Huib Jan, van Rossum, Daniel R., Vos, Jesse, Wagner, Jan, Ward-Thompson, Derek, Wardle, John, Washington, Jasmin E., Weintroub, Jonathan, Wharton, Robert, Wiik, Kaj, Witzel, Gunther, Wondrak, Michael F., Wong, George N., Wu, Qingwen, Yadlapalli, Nitika, Yamaguchi, Paul, Yfantis, Aristomenis, Yoon, Doosoo, Young, André, Young, Ken, Younsi, Ziri, Yu, Wei, Yuan, Feng, Yuan, Ye-Fei, Zensus, J. Anton, Zhang, Shuo, and Zhao, Guang-Yao
- Subjects
Astrophysics - High Energy Astrophysical Phenomena ,Astrophysics - Astrophysics of Galaxies - Abstract
Many active galaxies harbor powerful relativistic jets, however, the detailed mechanisms of their formation and acceleration remain poorly understood. To investigate the area of jet acceleration and collimation with the highest available angular resolution, we study the innermost region of the bipolar jet in the nearby low-ionization nuclear emission-line region (LINER) galaxy NGC 1052. We combined observations of NGC 1052 taken with VLBA, GMVA, and EHT over one week in the spring of 2017. For the first time, NGC 1052 was detected with the EHT, providing a size of the central region in-between both jet bases of 250 RS (Schwarzschild radii) perpendicular to the jet axes. This size estimate supports previous studies of the jets expansion profile which suggest two breaks of the profile at around 300 RS and 10000 RS distances to the core. Furthermore, we estimated the magnetic field to be 1.25 Gauss at a distance of 22 {\mu}as from the central engine by fitting a synchrotron-self absorption spectrum to the innermost emission feature, which shows a spectral turn-over at about 130 GHz. Assuming a purely poloidal magnetic field, this implies an upper limit on the magnetic field strength at the event horizon of 26000 Gauss, which is consistent with previous measurements. The complex, low-brightness, double-sided jet structure in NGC 1052 makes it a challenge to detect the source at millimeter (mm) wavelengths. However, our first EHT observations have demonstrated that detection is possible up to at least 230 GHz. This study offers a glimpse through the dense surrounding torus and into the innermost central region, where the jets are formed. This has enabled us to finally resolve this region and provide improved constraints on its expansion and magnetic field strength., Comment: 22 pages, 10 figures, published in A&A
- Published
- 2025
- Full Text
- View/download PDF
8. A multi-frequency study of sub-parsec jets with the Event Horizon Telescope
- Author
-
Röder, Jan, Wielgus, Maciek, Lobanov, Andrei P., Krichbaum, Thomas P., Nair, Dhanya G., Lee, Sang-Sung, Ros, Eduardo, Fish, Vincent L., Blackburn, Lindy, Chan, Chi-kwan, Issaoun, Sara, Janssen, Michael, Johnson, Michael D., Doeleman, Sheperd S., Bower, Geoffrey C., Crew, Geoffrey B., Tilanus, Remo P. J., Savolainen, Tuomas, Impellizzeri, C. M. Violette, Alberdi, Antxon, Baczko, Anne-Kathrin, Gómez, José L., Lu, Ru-Sen, Paraschos, Georgios F., Traianou, Efthalia, Goddi, Ciriaco, Kim, Daewon, Lisakov, Mikhail, Kovalev, Yuri Y., Voitsik, Petr A., Sokolovsky, Kirill V., Akiyama, Kazunori, Albentosa-Ruíz, Ezequiel, Alef, Walter, Algaba, Juan Carlos, Anantua, Richard, Asada, Keiichi, Azulay, Rebecca, Bach, Uwe, Ball, David, Baloković, Mislav, Bandyopadhyay, Bidisha, Barrett, John, Bauböck, Michi, Benson, Bradford A., Bintley, Dan, Blundell, Raymond, Bouman, Katherine L., Bremer, Michael, Brinkerink, Christiaan D., Brissenden, Roger, Britzen, Silke, Broderick, Avery E., Broguiere, Dominique, Bronzwaer, Thomas, Bustamante, Sandra, Byun, Do-Young, Carlstrom, John E., Ceccobello, Chiara, Chael, Andrew, Chang, Dominic O., Chatterjee, Koushik, Chatterjee, Shami, Chen, Ming-Tang, Chen, Yongjun, Cheng, Xiaopeng, Cho, Ilje, Christian, Pierre, Conroy, Nicholas S., Conway, John E., Cordes, James M., Crawford, Thomas M., Cruz-Osorio, Alejandro, Cui, Yuzhu, Curd, Brandon, Dahale, Rohan, Davelaar, Jordy, De Laurentis, Mariafelicia, Deane, Roger, Dempsey, Jessica, Desvignes, Gregory, Dexter, Jason, Dhruv, Vedant, Dihingia, Indu K., Dougall, Sean Taylor, Dzib, Sergio A., Eatough, Ralph P., Emami, Razieh, Falcke, Heino, Farah, Joseph, Fomalont, Edward, Ford, H. Alyson, Foschi, Marianna, Fraga-Encinas, Raquel, Freeman, William T., Friberg, Per, Fromm, Christian M., Fuentes, Antonio, Galison, Peter, Gammie, Charles F., García, Roberto, Gentaz, Olivier, Georgiev, Boris, Gold, Roman, Gómez-Ruiz, Arturo I., Gu, Minfeng, Gurwell, Mark, Hada, Kazuhiro, Haggard, Daryl, Haworth, Kari, Hecht, Michael H., Hesper, Ronald, Heumann, Dirk, Ho, Luis C., Ho, Paul, Honma, Mareki, Huang, Chih-Wei L., Huang, Lei, Hughes, David H., Ikeda, Shiro, Inoue, Makoto, James, David J., Jannuzi, Buell T., Jeter, Britton, Jiang, Wu, Jiménez-Rosales, Alejandra, Jorstad, Svetlana, Joshi, Abhishek V., Jung, Taehyun, Karami, Mansour, Karuppusamy, Ramesh, Kawashima, Tomohisa, Keating, Garrett K., Kettenis, Mark, Kim, Dong-Jin, Kim, Jae-Young, Kim, Jongsoo, Kim, Junhan, Kino, Motoki, Koay, Jun Yi, Kocherlakota, Prashant, Kofuji, Yutaro, Koyama, Shoko, Kramer, Carsten, Kramer, Joana A., Kramer, Michael, Kuo, Cheng-Yu, La Bella, Noemi, Lauer, Tod R., Lee, Daeyoung, Leung, Po Kin, Levis, Aviad, Li, Zhiyuan, Lico, Rocco, Lindahl, Greg, Lindqvist, Michael, Liu, Jun, Liu, Kuo, Liuzzo, Elisabetta, Lo, Wen-Ping, Loinard, Laurent, Lonsdale, Colin J., Lowitz, Amy E., MacDonald, Nicholas R., Mao, Jirong, Marchili, Nicola, Markoff, Sera, Marrone, Daniel P., Marscher, Alan P., Martí-Vidal, Iván, Matsushita, Satoki, Matthews, Lynn D., Medeiros, Lia, Menten, Karl M., Michalik, Daniel, Mizuno, Izumi, Mizuno, Yosuke, Moran, James M., Moriyama, Kotaro, Moscibrodzka, Monika, Mulaudzi, Wanga, Müller, Cornelia, Müller, Hendrik, Mus, Alejandro, Musoke, Gibwa, Myserlis, Ioannis, Nadolski, Andrew, Nagai, Hiroshi, Nagar, Neil M., Nakamura, Masanori, Narayanan, Gopal, Natarajan, Iniyan, Nathanail, Antonios, Fuentes, Santiago Navarro, Neilsen, Joey, Neri, Roberto, Ni, Chunchong, Noutsos, Aristeidis, Nowak, Michael A., Oh, Junghwan, Okino, Hiroki, Sánchez, Héctor R. Olivares, Ortiz-León, Gisela N., Oyama, Tomoaki, özel, Feryal, Palumbo, Daniel C. M., Park, Jongho, Parsons, Harriet, Patel, Nimesh, Pen, Ue-Li, Pesce, Dominic W., Piétu, Vincent, Plambeck, Richard, PopStefanija, Aleksandar, Porth, Oliver, Pötzl, Felix M., Prather, Ben, Preciado-López, Jorge A., Principe, Giacomo, Psaltis, Dimitrios, Pu, Hung-Yi, Ramakrishnan, Venkatessh, Rao, Ramprasad, Rawlings, Mark G., Ricarte, Angelo, Ripperda, Bart, Roelofs, Freek, Rogers, Alan, Romero-Cañizales, Cristina, Roshanineshat, Arash, Rottmann, Helge, Roy, Alan L., Ruiz, Ignacio, Ruszczyk, Chet, Rygl, Kazi L. J., Sánchez, Salvador, Sánchez-Argüelles, David, Sánchez-Portal, Miguel, Sasada, Mahito, Satapathy, Kaushik, Schloerb, F. Peter, Schonfeld, Jonathan, Schuster, Karl-Friedrich, Shao, Lijing, Shen, Zhiqiang, Small, Des, Sohn, Bong Won, SooHoo, Jason, Salas, León David Sosapanta, Souccar, Kamal, Stanway, Joshua S., Sun, He, Tazaki, Fumie, Tetarenko, Alexandra J., Tiede, Paul, Titus, Michael, Torne, Pablo, Toscano, Teresa, Trent, Tyler, Trippe, Sascha, Turk, Matthew, van Bemmel, Ilse, van Langevelde, Huib J., van Rossum, Daniel R., Vos, Jesse, Wagner, Jan, Ward-Thompson, Derek, Wardle, John, Washington, Jasmin E., Weintroub, Jonathan, Wharton, Robert, Wiik, Kaj, Witzel, Gunther, Wondrak, Michael F., Wong, George N., Wu, Qingwen, Yadlapalli, Nitika, Yamaguchi, Paul, Yfantis, Aristomenis, Yoon, Doosoo, Young, André, Young, Ken, Younsi, Ziri, Yu, Wei, Yuan, Feng, Yuan, Ye-Fei, Zensus, J. Anton, Zhang, Shuo, Zhao, Guang-Yao, and Zhao, Shan-Shan
- Subjects
Astrophysics - High Energy Astrophysical Phenomena ,Astrophysics - Astrophysics of Galaxies ,Astrophysics - Instrumentation and Methods for Astrophysics - Abstract
The 2017 observing campaign of the Event Horizon Telescope (EHT) delivered the first very long baseline interferometry (VLBI) images at the observing frequency of 230 GHz, leading to a number of unique studies on black holes and relativistic jets from active galactic nuclei (AGN). In total, eighteen sources were observed: the main science targets, Sgr A* and M87 along with various calibrators. We investigated the morphology of the sixteen AGN in the EHT 2017 data set, focusing on the properties of the VLBI cores: size, flux density, and brightness temperature. We studied their dependence on the observing frequency in order to compare it with the Blandford-K\"onigl (BK) jet model. We modeled the source structure of seven AGN in the EHT 2017 data set using linearly polarized circular Gaussian components and collected results for the other nine AGN from dedicated EHT publications, complemented by lower frequency data in the 2-86 GHz range. Then, we studied the dependences of the VLBI core flux density, size, and brightness temperature on the frequency measured in the AGN host frame. We compared the observations with the BK jet model and estimated the magnetic field strength dependence on the distance from the central black hole. Our results indicate a deviation from the standard BK model, particularly in the decrease of the brightness temperature with the observing frequency. Either bulk acceleration of the jet material, energy transfer from the magnetic field to the particles, or both are required to explain the observations.
- Published
- 2025
9. SLAM: Towards Efficient Multilingual Reasoning via Selective Language Alignment
- Author
-
Fan, Yuchun, Mu, Yongyu, Wang, Yilin, Huang, Lei, Ruan, Junhao, Li, Bei, Xiao, Tong, Huang, Shujian, Feng, Xiaocheng, and Zhu, Jingbo
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence - Abstract
Despite the significant improvements achieved by large language models (LLMs) in English reasoning tasks, these models continue to struggle with multilingual reasoning. Recent studies leverage a full-parameter and two-stage training paradigm to teach models to first understand non-English questions and then reason. However, this method suffers from both substantial computational resource computing and catastrophic forgetting. The fundamental cause is that, with the primary goal of enhancing multilingual comprehension, an excessive number of irrelevant layers and parameters are tuned during the first stage. Given our findings that the representation learning of languages is merely conducted in lower-level layers, we propose an efficient multilingual reasoning alignment approach that precisely identifies and fine-tunes the layers responsible for handling multilingualism. Experimental results show that our method, SLAM, only tunes 6 layers' feed-forward sub-layers including 6.5-8% of all parameters within 7B and 13B LLMs, achieving superior average performance than all strong baselines across 10 languages. Meanwhile, SLAM only involves one training stage, reducing training time by 4.1-11.9 compared to the two-stage method., Comment: Accepted by COLING 2025 (Oral)
- Published
- 2025
10. Length Controlled Generation for Black-box LLMs
- Author
-
Gu, Yuxuan, Wang, Wenjie, Feng, Xiaocheng, Zhong, Weihong, Zhu, Kun, Huang, Lei, Chua, Tat-Seng, and Qin, Bing
- Subjects
Computer Science - Computation and Language - Abstract
Large language models (LLMs) have demonstrated impressive instruction following capabilities, while still struggling to accurately manage the length of the generated text, which is a fundamental requirement in many real-world applications. Existing length control methods involve fine-tuning the parameters of LLMs, which is inefficient and suboptimal for practical use. In this paper, we propose a novel iterative sampling framework for text length control, integrating the Metropolis-Hastings algorithm with an importance sampling acceleration strategy. This framework efficiently and reliably regulates LLMs to generate length-constrained text without modifying the underlying parameters, thereby preserving the original capabilities of LLMs. Experimental results demonstrate that our framework achieves almost 100\% success rates of length control on Llama3.1 for tasks such as length-controlled abstractive summarization and length-constrained instruction following, with minimal additional computational overhead. This also highlights the significant potential of our method for precise length control across a broader range of applications, without compromising the versatility of LLMs., Comment: Preprint
- Published
- 2024
11. XTransplant: A Probe into the Upper Bound Performance of Multilingual Capability and Culture Adaptability in LLMs via Mutual Cross-lingual Feed-forward Transplantation
- Author
-
Ye, Yangfan, Feng, Xiaocheng, Feng, Xiachong, Qin, Libo, Huang, Yichong, Huang, Lei, Ma, Weitao, Zhang, Zhirui, Lu, Yunfei, Yan, Xiaohui, Tang, Duyu, Tu, Dandan, and Qin, Bing
- Subjects
Computer Science - Computation and Language - Abstract
Current large language models (LLMs) often exhibit imbalances in multilingual capabilities and cultural adaptability, largely due to their English-centric pretraining data. To address this imbalance, we propose a probing method named XTransplant that explores cross-lingual latent interactions via cross-lingual feed-forward transplantation during inference stage, with the hope of enabling the model to leverage the strengths of both English and non-English languages. Through extensive pilot experiments, we empirically prove that both the multilingual capabilities and cultural adaptability of LLMs hold the potential to be significantly improved by XTransplant, respectively from En -> non-En and non-En -> En, highlighting the underutilization of current LLMs' multilingual potential. And the patterns observed in these pilot experiments further motivate an offline scaling inference strategy, which demonstrates consistent performance improvements in multilingual and culture-aware tasks, sometimes even surpassing multilingual supervised fine-tuning. And we do hope our further analysis and discussion could help gain deeper insights into XTransplant mechanism.
- Published
- 2024
12. Bench-CoE: a Framework for Collaboration of Experts from Benchmark
- Author
-
Wang, Yuanshuai, Zhang, Xingjian, Zhao, Jinkun, Wen, Siwei, Feng, Peilin, Liao, Shuhao, Huang, Lei, and Wu, Wenjun
- Subjects
Computer Science - Artificial Intelligence - Abstract
Large Language Models (LLMs) are key technologies driving intelligent systems to handle multiple tasks. To meet the demands of various tasks, an increasing number of LLMs-driven experts with diverse capabilities have been developed, accompanied by corresponding benchmarks to evaluate their performance. This paper proposes the Bench-CoE framework, which enables Collaboration of Experts (CoE) by effectively leveraging benchmark evaluations to achieve optimal performance across various tasks. Bench-CoE includes a set of expert models, a router for assigning tasks to corresponding experts, and a benchmark dataset for training the router. Moreover, we formulate Query-Level and Subject-Level approaches based on our framework, and analyze the merits and drawbacks of these two approaches. Finally, we conduct a series of experiments with vary data distributions on both language and multimodal tasks to validate that our proposed Bench-CoE outperforms any single model in terms of overall performance. We hope this method serves as a baseline for further research in this area. The code is available at \url{https://github.com/ZhangXJ199/Bench-CoE}., Comment: The code is available at \url{https://github.com/ZhangXJ199/Bench-CoE}
- Published
- 2024
13. Copy-Move Forgery Detection and Question Answering for Remote Sensing Image
- Author
-
Zhang, Ze, Zhao, Enyuan, Wan, Ziyi, Nie, Jie, Liang, Xinyue, and Huang, Lei
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Multimedia - Abstract
This paper introduces the task of Remote Sensing Copy-Move Question Answering (RSCMQA). Unlike traditional Remote Sensing Visual Question Answering (RSVQA), RSCMQA focuses on interpreting complex tampering scenarios and inferring relationships between objects. Based on the practical needs of national defense security and land resource monitoring, we have developed an accurate and comprehensive global dataset for remote sensing image copy-move question answering, named RS-CMQA-2.1M. These images were collected from 29 different regions across 14 countries. Additionally, we have refined a balanced dataset, RS-CMQA-B, to address the long-standing issue of long-tail data in the remote sensing field. Furthermore, we propose a region-discriminative guided multimodal CMQA model, which enhances the accuracy of answering questions about tampered images by leveraging prompt about the differences and connections between the source and tampered domains. Extensive experiments demonstrate that our method provides a stronger benchmark for RS-CMQA compared to general VQA and RSVQA models. Our dataset and code are available at https://github.com/shenyedepisa/RSCMQA., Comment: 7 figs, 7 tables
- Published
- 2024
14. Correlated Rydberg Electromagnetically Induced Transparencys
- Author
-
Huang, Lei, Wang, Peng-fei, Zhang, Han-xiao, Zhu, Yu, Yang, Hong, and Yan, Dong
- Subjects
Quantum Physics ,Physics - Atomic Physics - Abstract
In the regime of Rydberg electromagnetically induced transparency, we study the correlated behaviors between the transmission spectra of a pair of probe fields passing through respective parallel one-dimensional cold Rydberg ensembles. Due to the van der Waals (vdW) interactions between Rydberg atoms, each ensemble exhibits a local optical nonlinearity, where the output EIT spectra are sensitive to both the input probe intensity and the photonic statistics. More interestingly, a nonlocal optical nonlinearity emerges between two spatially separated ensembles, as the probe transmissivity and probe correlation at the exit of one Rydberg ensemble can be manipulated by the probe field at the input of the other Rydberg ensemble. Realizing correlated Rydberg EITs holds great potential for applications in quantum control, quantum network, quantum walk and so on., Comment: 7 pages, 6 figures
- Published
- 2024
15. MUltiplexed Survey Telescope: Perspectives for Large-Scale Structure Cosmology in the Era of Stage-V Spectroscopic Survey
- Author
-
Zhao, Cheng, Huang, Song, He, Mengfan, Montero-Camacho, Paulo, Liu, Yu, Renard, Pablo, Tang, Yunyi, Verdier, Aurelien, Xu, Wenshuo, Yang, Xiaorui, Yu, Jiaxi, Zhang, Yao, Zhao, Siyi, Zhou, Xingchen, He, Shengyu, Kneib, Jean-Paul, Li, Jiayi, Li, Zhuoyang, Wang, Wen-Ting, Xianyu, Zhong-Zhi, Zhang, Yidian, Gsponer, Rafaela, Li, Xiao-Dong, Rocher, Antoine, Zou, Siwei, Tan, Ting, Huang, Zhiqi, Wang, Zhuoxiao, Li, Pei, Rombach, Maxime, Dong, Chenxing, Forero-Sanchez, Daniel, Shan, Huanyuan, Wang, Tao, Li, Yin, Zhai, Zhongxu, Wang, Yuting, Zhao, Gong-Bo, Shi, Yong, Mao, Shude, Huang, Lei, Guo, Liquan, and Cai, Zheng
- Subjects
Astrophysics - Cosmology and Nongalactic Astrophysics ,Astrophysics - Instrumentation and Methods for Astrophysics - Abstract
The MUltiplexed Survey Telescope (MUST) is a 6.5-meter telescope under development. Dedicated to highly-multiplexed, wide-field spectroscopic surveys, MUST observes over 20,000 targets simultaneously using 6.2-mm pitch positioning robots within a ~5 deg2 field of view. MUST aims to carry out the first Stage-V spectroscopic survey in the 2030s to map the 3D Universe with over 100 million galaxies and quasars, spanning from the nearby Universe to redshift z~5.5, corresponding to around 1 billion years after the Big Bang. To cover this extensive redshift range, we present an initial conceptual target selection algorithm for different types of galaxies, from local bright galaxies, luminous red galaxies, and emission line galaxies to high-redshift (2 < z < 5.5) Lyman-break galaxies. Using Fisher forecasts, we demonstrate that MUST can address fundamental questions in cosmology, including the nature of dark energy, test of gravity theories, and investigations into primordial physics. This is the first paper in the series of science white papers for MUST, with subsequent developments focusing on additional scientific cases such as galaxy and quasar evolution, Milky Way physics, and dynamic phenomena in the time-domain Universe., Comment: To be submitted to SCPMA
- Published
- 2024
16. $\textit{Dirigo}$: A Method to Extract Event Logs for Object-Centric Processes
- Author
-
Wei, Jia, Ouyang, Chun, ter Hofstede, Arthur, Wang, Ying, and Huang, Lei
- Subjects
Computer Science - Databases - Abstract
Real-world processes involve multiple object types with intricate interrelationships. Traditional event logs (in XES format), which record process execution centred around the case notion, are restricted to a single-object perspective, making it difficult to capture the behaviour of multiple objects and their interactions. To address this limitation, object-centric event logs (OCEL) have been introduced to capture both the objects involved in a process and their interactions with events. The object-centric event data (OCED) metamodel extends the OCEL format by further capturing dynamic object attributes and object-to-object relations. Recently OCEL 2.0 has been proposed based on OCED metamodel. Current research on generating OCEL logs requires specific input data sources, and resulting log data often fails to fully conform to OCEL 2.0. Moreover, the generated OCEL logs vary across different representational formats and their quality remains unevaluated. To address these challenges, a set of quality criteria for evaluating OCEL log representations is established. Guided by these criteria, $\textit{Dirigo}$ is proposed -- a method for extracting event logs that not only conforms to OCEL 2.0 but also extends it by capturing the temporal aspect of dynamic object-to-object relations. Object-role Modelling (ORM), a conceptual data modelling technique, is employed to describe the artifact produced at each step of $\textit{Dirigo}$. To validate the applicability of $\textit{Dirigo}$, it is applied to a real-life use case, extracting an event log via simulation. The quality of the log representation of the extracted event log is compared to those of existing OCEL logs using the established quality criteria.
- Published
- 2024
17. Discrete Modeling via Boundary Conditional Diffusion Processes
- Author
-
Gu, Yuxuan, Feng, Xiaocheng, Huang, Lei, Wu, Yingsheng, Zhou, Zekun, Zhong, Weihong, Zhu, Kun, and Qin, Bing
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence - Abstract
We present an novel framework for efficiently and effectively extending the powerful continuous diffusion processes to discrete modeling. Previous approaches have suffered from the discrepancy between discrete data and continuous modeling. Our study reveals that the absence of guidance from discrete boundaries in learning probability contours is one of the main reasons. To address this issue, we propose a two-step forward process that first estimates the boundary as a prior distribution and then rescales the forward trajectory to construct a boundary conditional diffusion model. The reverse process is proportionally adjusted to guarantee that the learned contours yield more precise discrete data. Experimental results indicate that our approach achieves strong performance in both language modeling and discrete image generation tasks. In language modeling, our approach surpasses previous state-of-the-art continuous diffusion language models in three translation tasks and a summarization task, while also demonstrating competitive performance compared to auto-regressive transformers. Moreover, our method achieves comparable results to continuous diffusion models when using discrete ordinal pixels and establishes a new state-of-the-art for categorical image generation on the Cifar-10 dataset., Comment: NeuraIPS 2024 poster
- Published
- 2024
18. SafeBench: A Safety Evaluation Framework for Multimodal Large Language Models
- Author
-
Ying, Zonghao, Liu, Aishan, Liang, Siyuan, Huang, Lei, Guo, Jinyang, Zhou, Wenbo, Liu, Xianglong, and Tao, Dacheng
- Subjects
Computer Science - Cryptography and Security - Abstract
Multimodal Large Language Models (MLLMs) are showing strong safety concerns (e.g., generating harmful outputs for users), which motivates the development of safety evaluation benchmarks. However, we observe that existing safety benchmarks for MLLMs show limitations in query quality and evaluation reliability limiting the detection of model safety implications as MLLMs continue to evolve. In this paper, we propose \toolns, a comprehensive framework designed for conducting safety evaluations of MLLMs. Our framework consists of a comprehensive harmful query dataset and an automated evaluation protocol that aims to address the above limitations, respectively. We first design an automatic safety dataset generation pipeline, where we employ a set of LLM judges to recognize and categorize the risk scenarios that are most harmful and diverse for MLLMs; based on the taxonomy, we further ask these judges to generate high-quality harmful queries accordingly resulting in 23 risk scenarios with 2,300 multi-modal harmful query pairs. During safety evaluation, we draw inspiration from the jury system in judicial proceedings and pioneer the jury deliberation evaluation protocol that adopts collaborative LLMs to evaluate whether target models exhibit specific harmful behaviors, providing a reliable and unbiased assessment of content security risks. In addition, our benchmark can also be extended to the audio modality showing high scalability and potential. Based on our framework, we conducted large-scale experiments on 15 widely-used open-source MLLMs and 6 commercial MLLMs (e.g., GPT-4o, Gemini), where we revealed widespread safety issues in existing MLLMs and instantiated several insights on MLLM safety performance such as image quality and parameter size.
- Published
- 2024
19. Advancing Large Language Model Attribution through Self-Improving
- Author
-
Huang, Lei, Feng, Xiaocheng, Ma, Weitao, Zhao, Liang, Fan, Yuchun, Zhong, Weihong, Xu, Dongliang, Yang, Qing, Liu, Hongtao, and Qin, Bing
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence - Abstract
Teaching large language models (LLMs) to generate text with citations to evidence sources can mitigate hallucinations and enhance verifiability in information-seeking systems. However, improving this capability requires high-quality attribution data, which is costly and labor-intensive. Inspired by recent advances in self-improvement that enhance LLMs without manual annotation, we present START, a Self-Taught AttRibuTion framework for iteratively improving the attribution capability of LLMs. First, to prevent models from stagnating due to initially insufficient supervision signals, START leverages the model to self-construct synthetic training data for warming up. To further self-improve the model's attribution ability, START iteratively utilizes fine-grained preference supervision signals constructed from its sampled responses to encourage robust, comprehensive, and attributable generation. Experiments on three open-domain question-answering datasets, covering long-form QA and multi-step reasoning, demonstrate significant performance gains of 25.13% on average without relying on human annotations and more advanced models. Further analysis reveals that START excels in aggregating information across multiple sources., Comment: Accepted by EMNLP 2024 Main Conference
- Published
- 2024
20. First Very Long Baseline Interferometry Detections at 870{\mu}m
- Author
-
Raymond, Alexander W., Doeleman, Sheperd S., Asada, Keiichi, Blackburn, Lindy, Bower, Geoffrey C., Bremer, Michael, Broguiere, Dominique, Chen, Ming-Tang, Crew, Geoffrey B., Dornbusch, Sven, Fish, Vincent L., García, Roberto, Gentaz, Olivier, Goddi, Ciriaco, Han, Chih-Chiang, Hecht, Michael H., Huang, Yau-De, Janssen, Michael, Keating, Garrett K., Koay, Jun Yi, Krichbaum, Thomas P., Lo, Wen-Ping, Matsushita, Satoki, Matthews, Lynn D., Moran, James M., Norton, Timothy J., Patel, Nimesh, Pesce, Dominic W., Ramakrishnan, Venkatessh, Rottmann, Helge, Roy, Alan L., Sánchez, Salvador, Tilanus, Remo P. J., Titus, Michael, Torne, Pablo, Wagner, Jan, Weintroub, Jonathan, Wielgus, Maciek, Young, André, Akiyama, Kazunori, Albentosa-Ruíz, Ezequiel, Alberdi, Antxon, Alef, Walter, Algaba, Juan Carlos, Anantua, Richard, Azulay, Rebecca, Bach, Uwe, Baczko, Anne-Kathrin, Ball, David, Baloković, Mislav, Bandyopadhyay, Bidisha, Barrett, John, Bauböck, Michi, Benson, Bradford A., Bintley, Dan, Blundell, Raymond, Bouman, Katherine L., Boyce, Hope, Brissenden, Roger, Britzen, Silke, Broderick, Avery E., Bronzwaer, Thomas, Bustamante, Sandra, Carlstrom, John E., Chael, Andrew, Chan, Chi-kwan, Chang, Dominic O., Chatterjee, Koushik, Chatterjee, Shami, Chen, Yongjun, Cheng, Xiaopeng, Cho, Ilje, Christian, Pierre, Conroy, Nicholas S., Conway, John E., Crawford, Thomas M., Cruz-Osorio, Alejandro, Cui, Yuzhu, Dahale, Rohan, Davelaar, Jordy, De Laurentis, Mariafelicia, Deane, Roger, Dempsey, Jessica, Desvignes, Gregory, Dexter, Jason, Dhruv, Vedant, Dihingia, Indu K., Dzib, Sergio A., Eatough, Ralph P., Emami, Razieh, Falcke, Heino, Farah, Joseph, Fomalont, Edward, Fontana, Anne-Laure, Ford, H. Alyson, Foschi, Marianna, Fraga-Encinas, Raquel, Freeman, William T., Friberg, Per, Fromm, Christian M., Fuentes, Antonio, Galison, Peter, Gammie, Charles F., Georgiev, Boris, Gold, Roman, Gómez-Ruiz, Arturo I., Gómez, José L., Gu, Minfeng, Gurwell, Mark, Hada, Kazuhiro, Haggard, Daryl, Hesper, Ronald, Heumann, Dirk, Ho, Luis C., Ho, Paul, Honma, Mareki, Huang, Chih-Wei L., Huang, Lei, Hughes, David H., Ikeda, Shiro, Impellizzeri, C. M. Violette, Inoue, Makoto, Issaoun, Sara, James, David J., Jannuzi, Buell T., Jeter, Britton, Jiang, Wu, Jiménez-Rosales, Alejandra, Johnson, Michael D., Jorstad, Svetlana, Jones, Adam C., Joshi, Abhishek V., Jung, Taehyun, Karuppusamy, Ramesh, Kawashima, Tomohisa, Kettenis, Mark, Kim, Dong-Jin, Kim, Jae-Young, Kim, Jongsoo, Kim, Junhan, Kino, Motoki, Kocherlakota, Prashant, Kofuji, Yutaro, Koch, Patrick M., Koyama, Shoko, Kramer, Carsten, Kramer, Joana A., Kramer, Michael, Kubo, Derek, Kuo, Cheng-Yu, La Bella, Noemi, Lee, Sang-Sung, Levis, Aviad, Li, Zhiyuan, Lico, Rocco, Lindahl, Greg, Lindqvist, Michael, Lisakov, Mikhail, Liu, Jun, Liu, Kuo, Liuzzo, Elisabetta, Lobanov, Andrei P., Loinard, Laurent, Lonsdale, Colin J., Lowitz, Amy E., Lu, Ru-Sen, MacDonald, Nicholas R., Mahieu, Sylvain, Maier, Doris, Mao, Jirong, Marchili, Nicola, Markoff, Sera, Marrone, Daniel P., Marscher, Alan P., Martí-Vidal, Iván, Medeiros, Lia, Menten, Karl M., Mizuno, Izumi, Mizuno, Yosuke, Montgomery, Joshua, Moriyama, Kotaro, Moscibrodzka, Monika, Mulaudzi, Wanga, Müller, Cornelia, Müller, Hendrik, Mus, Alejandro, Musoke, Gibwa, Myserlis, Ioannis, Nagai, Hiroshi, Nagar, Neil M., Nakamura, Masanori, Narayanan, Gopal, Natarajan, Iniyan, Nathanail, Antonios, Fuentes, Santiago Navarro, Neilsen, Joey, Ni, Chunchong, Nowak, Michael A., Oh, Junghwan, Okino, Hiroki, Sánchez, Héctor Raúl Olivares, Oyama, Tomoaki, Özel, Feryal, Palumbo, Daniel C. M., Paraschos, Georgios Filippos, Park, Jongho, Parsons, Harriet, Pen, Ue-Li, Piétu, Vincent, PopStefanija, Aleksandar, Porth, Oliver, Prather, Ben, Principe, Giacomo, Psaltis, Dimitrios, Pu, Hung-Yi, Raffin, Philippe A., Rao, Ramprasad, Rawlings, Mark G., Ricarte, Angelo, Ripperda, Bart, Roelofs, Freek, Romero-Cañizales, Cristina, Ros, Eduardo, Roshanineshat, Arash, Ruiz, Ignacio, Ruszczyk, Chet, Rygl, Kazi L. J., Sánchez-Argüelles, David, Sánchez-Portal, Miguel, Sasada, Mahito, Satapathy, Kaushik, Savolainen, Tuomas, Schloerb, F. Peter, Schonfeld, Jonathan, Schuster, Karl-Friedrich, Shao, Lijing, Shen, Zhiqiang, Small, Des, Sohn, Bong Won, SooHoo, Jason, Salas, León David Sosapanta, Souccar, Kamal, Srinivasan, Ranjani, Stanway, Joshua S., Sun, He, Tazaki, Fumie, Tetarenko, Alexandra J., Tiede, Paul, Toma, Kenji, Toscano, Teresa, Traianou, Efthalia, Trent, Tyler, Trippe, Sascha, Turk, Matthew, van Bemmel, Ilse, van Langevelde, Huib Jan, van Rossum, Daniel R., Vos, Jesse, Ward-Thompson, Derek, Wardle, John, Washington, Jasmin E., Wharton, Robert, Wiik, Kaj, Witzel, Gunther, Wondrak, Michael F., Wong, George N., Wu, Qingwen, Yadlapalli, Nitika, Yamaguchi, Paul, Yfantis, Aristomenis, Yoon, Doosoo, Younsi, Ziri, Yu, Wei, Yuan, Feng, Yuan, Ye-Fei, Zensus, J. Anton, Zhang, Shuo, Zhao, Guang-Yao, and Zhao, Shan-Shan
- Subjects
Astrophysics - Instrumentation and Methods for Astrophysics ,Astrophysics - High Energy Astrophysical Phenomena - Abstract
The first very long baseline interferometry (VLBI) detections at 870$\mu$m wavelength (345$\,$GHz frequency) are reported, achieving the highest diffraction-limited angular resolution yet obtained from the surface of the Earth, and the highest-frequency example of the VLBI technique to date. These include strong detections for multiple sources observed on inter-continental baselines between telescopes in Chile, Hawaii, and Spain, obtained during observations in October 2018. The longest-baseline detections approach 11$\,$G$\lambda$ corresponding to an angular resolution, or fringe spacing, of 19$\mu$as. The Allan deviation of the visibility phase at 870$\mu$m is comparable to that at 1.3$\,$mm on the relevant integration time scales between 2 and 100$\,$s. The detections confirm that the sensitivity and signal chain stability of stations in the Event Horizon Telescope (EHT) array are suitable for VLBI observations at 870$\mu$m. Operation at this short wavelength, combined with anticipated enhancements of the EHT, will lead to a unique high angular resolution instrument for black hole studies, capable of resolving the event horizons of supermassive black holes in both space and time., Comment: Corresponding author: S. Doeleman
- Published
- 2024
- Full Text
- View/download PDF
21. ProteinRPN: Towards Accurate Protein Function Prediction with Graph-Based Region Proposals
- Author
-
Mitra, Shania, Huang, Lei, and Kellis, Manolis
- Subjects
Quantitative Biology - Quantitative Methods ,Computer Science - Machine Learning - Abstract
Protein function prediction is a crucial task in bioinformatics, with significant implications for understanding biological processes and disease mechanisms. While the relationship between sequence and function has been extensively explored, translating protein structure to function continues to present substantial challenges. Various models, particularly, CNN and graph-based deep learning approaches that integrate structural and functional data, have been proposed to address these challenges. However, these methods often fall short in elucidating the functional significance of key residues essential for protein functionality, as they predominantly adopt a retrospective perspective, leading to suboptimal performance. Inspired by region proposal networks in computer vision, we introduce the Protein Region Proposal Network (ProteinRPN) for accurate protein function prediction. Specifically, the region proposal module component of ProteinRPN identifies potential functional regions (anchors) which are refined through the hierarchy-aware node drop pooling layer favoring nodes with defined secondary structures and spatial proximity. The representations of the predicted functional nodes are enriched using attention mechanisms and subsequently fed into a Graph Multiset Transformer, which is trained with supervised contrastive (SupCon) and InfoNCE losses on perturbed protein structures. Our model demonstrates significant improvements in predicting Gene Ontology (GO) terms, effectively localizing functional residues within protein structures. The proposed framework provides a robust, scalable solution for protein function annotation, advancing the understanding of protein structure-function relationships in computational biology.
- Published
- 2024
22. A versatile informative diffusion model for single-cell ATAC-seq data generation and analysis
- Author
-
Huang, Lei, Xiong, Lei, Sun, Na, Liu, Zunpeng, Wong, Ka-Chun, and Kellis, Manolis
- Subjects
Quantitative Biology - Genomics ,Quantitative Biology - Biomolecules - Abstract
The rapid advancement of single-cell ATAC sequencing (scATAC-seq) technologies holds great promise for investigating the heterogeneity of epigenetic landscapes at the cellular level. The amplification process in scATAC-seq experiments often introduces noise due to dropout events, which results in extreme sparsity that hinders accurate analysis. Consequently, there is a significant demand for the generation of high-quality scATAC-seq data in silico. Furthermore, current methodologies are typically task-specific, lacking a versatile framework capable of handling multiple tasks within a single model. In this work, we propose ATAC-Diff, a versatile framework, which is based on a latent diffusion model conditioned on the latent auxiliary variables to adapt for various tasks. ATAC-Diff is the first diffusion model for the scATAC-seq data generation and analysis, composed of auxiliary modules encoding the latent high-level variables to enable the model to learn the semantic information to sample high-quality data. Gaussian Mixture Model (GMM) as the latent prior and auxiliary decoder, the yield variables reserve the refined genomic information beneficial for downstream analyses. Another innovation is the incorporation of mutual information between observed and hidden variables as a regularization term to prevent the model from decoupling from latent variables. Through extensive experiments, we demonstrate that ATAC-Diff achieves high performance in both generation and analysis tasks, outperforming state-of-the-art models.
- Published
- 2024
23. Ex3: Automatic Novel Writing by Extracting, Excelsior and Expanding
- Author
-
Huang, Lei, Guo, Jiaming, He, Guanhua, Zhang, Xishan, Zhang, Rui, Peng, Shaohui, Liu, Shaoli, and Chen, Tianshi
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence - Abstract
Generating long-term texts such as novels using artificial intelligence has always been a challenge. A common approach is to use large language models (LLMs) to construct a hierarchical framework that first plans and then writes. Despite the fact that the generated novels reach a sufficient length, they exhibit poor logical coherence and appeal in their plots and deficiencies in character and event depiction, ultimately compromising the overall narrative quality. In this paper, we propose a method named Extracting Excelsior and Expanding. Ex3 initially extracts structure information from raw novel data. By combining this structure information with the novel data, an instruction-following dataset is meticulously crafted. This dataset is then utilized to fine-tune the LLM, aiming for excelsior generation performance. In the final stage, a tree-like expansion method is deployed to facilitate the generation of arbitrarily long novels. Evaluation against previous methods showcases Ex3's ability to produce higher-quality long-form novels.
- Published
- 2024
24. Learning Fine-Grained Grounded Citations for Attributed Large Language Models
- Author
-
Huang, Lei, Feng, Xiaocheng, Ma, Weitao, Gu, Yuxuan, Zhong, Weihong, Feng, Xiachong, Yu, Weijiang, Peng, Weihua, Tang, Duyu, Tu, Dandan, and Qin, Bing
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence - Abstract
Despite the impressive performance on information-seeking tasks, large language models (LLMs) still struggle with hallucinations. Attributed LLMs, which augment generated text with in-line citations, have shown potential in mitigating hallucinations and improving verifiability. However, current approaches suffer from suboptimal citation quality due to their reliance on in-context learning. Furthermore, the practice of citing only coarse document identifiers makes it challenging for users to perform fine-grained verification. In this work, we introduce FRONT, a training framework designed to teach LLMs to generate Fine-Grained Grounded Citations. By grounding model outputs in fine-grained supporting quotes, these quotes guide the generation of grounded and consistent responses, not only improving citation quality but also facilitating fine-grained verification. Experiments on the ALCE benchmark demonstrate the efficacy of FRONT in generating superior grounded responses and highly supportive citations. With LLaMA-2-7B, the framework significantly outperforms all the baselines, achieving an average of 14.21% improvement in citation quality across all datasets, even surpassing ChatGPT., Comment: Accepted by ACL 2024 Findings
- Published
- 2024
25. EXIT: An EXplicit Interest Transfer Framework for Cross-Domain Recommendation
- Author
-
Huang, Lei, Li, Weitao, Zhang, Chenrui, Wang, Jinpeng, Yi, Xianchun, and Chen, Sheng
- Subjects
Computer Science - Information Retrieval ,Computer Science - Artificial Intelligence - Abstract
Cross-domain recommendation has attracted substantial interest in industrial apps such as Meituan, which serves multiple business domains via knowledge transfer and meets the diverse interests of users. However, existing methods typically follow an implicit modeling paradigm that blends the knowledge from both the source and target domains, and design intricate network structures to share learned embeddings or patterns between domains to improve recommendation accuracy. Since the transfer of interest signals is unsupervised, these implicit paradigms often struggle with the negative transfer resulting from differences in service functions and presentation forms across different domains. In this paper, we propose a simple and effective EXplicit Interest Transfer framework named EXIT to address the stated challenge. Specifically, we propose a novel label combination approach that enables the model to directly learn beneficial source domain interests through supervised learning, while excluding inappropriate interest signals. Moreover, we introduce a scene selector network to model the interest transfer intensity under fine-grained scenes. Offline experiments conducted on the industrial production dataset and online A/B tests validate the superiority and effectiveness of our proposed framework. Without complex network structures or training processes, EXIT can be easily deployed in the industrial recommendation system. EXIT has been successfully deployed in the online homepage recommendation system of Meituan App, serving the main traffic., Comment: Accepted at CIKM 2024
- Published
- 2024
26. Practical continuous-variable quantum secret sharing using local local oscillator
- Author
-
Liao, Qin, Fei, Zhuoying, Huang, Lei, and Fu, Xiquan
- Subjects
Quantum Physics - Abstract
Although continuous-variable quantum secret sharing (CVQSS) has been theoretically proven to be secure, it may still be vulnerable to various local oscillator (LO)-aimed attacks. To close this loophole, we propose a practical CVQSS scheme using local LO (LLO), which is called LLO-CVQSS. In this scheme, LO is no longer generated by each user but can be locally generated by the legitimate party, i.e., the dealer. This waives the necessity that all LOs have to be transmitted through an untrusted channel, which makes CVQSS system naturally immune to all LO-aimed attacks, greatly enhancing its practical security. We also develop a specially designed phase compensation method for LLO-CVQSS so that the phase noise of the whole system can be eliminated. We finally construct a noise model for LLO-CVQSS and derive its security bound against both eavesdroppers and dishonest users. Numerical simulation shows that LLO-CVQSS is able to support 30 users at the same time and its maximal transmission distance reaches 112 km, revealing that LLO-CVQSS is not only has the ability to defend itself against all LO-aimed attacks but also has the potential for building large-scale practical quantum communication networks., Comment: 13 pages, 7 figures
- Published
- 2024
27. The white-light superflares from cool stars in GWAC triggers
- Author
-
Li, Guang-Wei, Wang, Liang, Yuan, Hai-Long, Xin, Li-Ping, Wang, Jing, Wu, Chao, Li, Hua-Li, Haerken, Hasitieer, Wang, Wei-Hua, Cai, Hong-Bo, Han, Xu-Hui, Xu, Yang, Huang, Lei, Lu, Xiao-Meng, Bai, Jian-Ying, Wang, Xiang-Yu, Dai, Zi-Gao, Liang, En-Wei, and Wei, Jian-Yan
- Subjects
Astrophysics - Solar and Stellar Astrophysics - Abstract
M-type stars are the ones that flare most frequently, but how big their maximum flare energy can reach is still unknown. We present 163 flares from 162 individual M2 through L1-type stars that triggered the GWAC, with flare energies ranging from $10^{32.2}$ to $10^{36.4}$ erg . The flare amplitudes range from $\triangle G = 0.84$ to $\sim 10$ mag. Flare energy increases with stellar surface temperature ($T_{\rm eff}$) but both $\triangle G$ and equivalent duration $\log_{10}(ED)$ seem to be independent of $T_{\rm eff}$. Combining periods detected from light curves of TESS and K2, spectra from LAMOST, SDSS and the 2.16 m Telescope, and the Gaia DR3 data, we found that these GWAC flare stars are young. For the stars that have spectra, we found that these stars are in or very near to the saturation region, and $\log_{10}(L_{\rm H\alpha}/L_{\rm bol})$ is lower for M7-L1 stars than for M2-M6 stars. We also studied the relation between GWAC flare bolometric energy $E_{\rm bol}$ and stellar hemispherical area $S$, and found that $\log_{10}E_{\rm bol}$ (in erg) increases with increasing $S$ (in cm$^2$), and the maximum flare energy $\log_{10}E_{\rm bol, max} \geqslant \log_{10}S + 14.25$. For M7-L1 stars, there seem to be other factors limiting their maximum flare energies in addition to stellar hemispherical area., Comment: 18 pages, 11 figures, 4 tables
- Published
- 2024
- Full Text
- View/download PDF
28. Intelligent Reflecting Surface-Assisted NLOS Sensing With OFDM Signals
- Author
-
Wang, Jilin, Fang, Jun, Li, Hongbin, and Huang, Lei
- Subjects
Electrical Engineering and Systems Science - Signal Processing - Abstract
This work addresses the problem of intelligent reflecting surface (IRS) assisted target sensing in a non-line-of-sight (NLOS) scenario, where an IRS is employed to facilitate the radar/access point (AP) to sense the targets when the line-of-sight (LOS) path between the AP and the target is blocked by obstacles. To sense the targets, the AP transmits a train of uniformly-spaced orthogonal frequency division multiplexing (OFDM) pulses, and then perceives the targets based on the echoes from the AP-IRS-targets-IRS-AP channel. To resolve an inherent scaling ambiguity associated with IRS-assisted NLOS sensing, we propose a two-phase sensing scheme by exploiting the diversity in the illumination pattern of the IRS across two different phases. Specifically, the received echo signals from the two phases are formulated as third-order tensors. Then a canonical polyadic (CP) decomposition-based method is developed to estimate each target's parameters including the direction of arrival (DOA), Doppler shift and time delay. Our analysis reveals that the proposed method achieves reliable NLOS sensing using a modest quantity of pulse/subcarrier resources. Simulation results are provided to show the effectiveness of the proposed method under the challenging scenario where the degrees-of-freedom provided by the AP-IRS channel are not enough for resolving the scaling ambiguity.
- Published
- 2024
29. Investigating and Mitigating the Multimodal Hallucination Snowballing in Large Vision-Language Models
- Author
-
Zhong, Weihong, Feng, Xiaocheng, Zhao, Liang, Li, Qiming, Huang, Lei, Gu, Yuxuan, Ma, Weitao, Xu, Yuan, and Qin, Bing
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence ,Computer Science - Computation and Language - Abstract
Though advanced in understanding visual information with human languages, Large Vision-Language Models (LVLMs) still suffer from multimodal hallucinations. A natural concern is that during multimodal interaction, the generated hallucinations could influence the LVLMs' subsequent generation. Thus, we raise a question: When presented with a query relevant to the previously generated hallucination, will LVLMs be misled and respond incorrectly, even though the ground visual information exists? To answer this, we propose a framework called MMHalSnowball to evaluate LVLMs' behaviors when encountering generated hallucinations, where LVLMs are required to answer specific visual questions within a curated hallucinatory conversation. Crucially, our experiment shows that the performance of open-source LVLMs drops by at least $31\%$, indicating that LVLMs are prone to accept the generated hallucinations and make false claims that they would not have supported without distractions. We term this phenomenon Multimodal Hallucination Snowballing. To mitigate this, we further propose a training-free method called Residual Visual Decoding, where we revise the output distribution of LVLMs with the one derived from the residual visual input, providing models with direct access to the visual information. Experiments show that our method can mitigate more than $24\%$ of the snowballed multimodal hallucination while maintaining capabilities., Comment: Accepted to ACL 2024 Main Conference. 21 pages, 20 figures
- Published
- 2024
30. Unveiling Entity-Level Unlearning for Large Language Models: A Comprehensive Analysis
- Author
-
Ma, Weitao, Feng, Xiaocheng, Zhong, Weihong, Huang, Lei, Ye, Yangfan, Feng, Xiachong, and Qin, Bing
- Subjects
Computer Science - Computation and Language - Abstract
Large language model unlearning has garnered increasing attention due to its potential to address security and privacy concerns, leading to extensive research in the field. However, much of this research has concentrated on instance-level unlearning, specifically targeting the removal of predefined instances containing sensitive content. This focus has left a significant gap in the exploration of full entity-level unlearning, which is critical in real-world scenarios such as copyright protection. To this end, we propose a novel task of Entity-level unlearning, which aims to erase entity-related knowledge from the target model completely. To thoroughly investigate this task, we systematically evaluate trending unlearning algorithms, revealing that current methods struggle to achieve effective entity-level unlearning. Then, we further explore the factors that influence the performance of the unlearning algorithms, identifying that knowledge coverage and the size of the forget set play pivotal roles. Notably, our analysis also uncovers that entities introduced through fine-tuning are more vulnerable to unlearning than pre-trained entities. These findings collectively offer valuable insights for advancing entity-level unlearning for LLMs., Comment: Accepted by COLING 2025
- Published
- 2024
31. On the complexity of matrix Putinar's Positivstellensatz
- Author
-
Huang, Lei
- Subjects
Mathematics - Optimization and Control - Abstract
This paper studies the complexity of matrix Putinar's Positivstellens{\"a}tz on the semialgebraic set that is given by the polynomial matrix inequality. \rev{When the quadratic module generated by the constrained polynomial matrix is Archimedean}, we prove a polynomial bound on the degrees of terms appearing in the representation of matrix Putinar's Positivstellens{\"a}tz. Estimates on the exponent and constant are given. As a byproduct, a polynomial bound on the convergence rate of matrix sum-of-squares relaxations is obtained, which resolves an open question raised by Dinh and Pham. When the constraining set is unbounded, we also prove a similar bound for the matrix version of Putinar--Vasilescu's Positivstellens{\"a}tz by exploiting homogenization techniques.
- Published
- 2024
32. ADSNet: Cross-Domain LTV Prediction with an Adaptive Siamese Network in Advertising
- Author
-
Wang, Ruize, Xu, Hui, Cheng, Ying, He, Qi, Zhou, Xing, Feng, Rui, Xu, Wei, Huang, Lei, and Jiang, Jie
- Subjects
Computer Science - Information Retrieval ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning - Abstract
Advertising platforms have evolved in estimating Lifetime Value (LTV) to better align with advertisers' true performance metric. However, the sparsity of real-world LTV data presents a significant challenge to LTV predictive model(i.e., pLTV), severely limiting the their capabilities. Therefore, we propose to utilize external data, in addition to the internal data of advertising platform, to expand the size of purchase samples and enhance the LTV prediction model of the advertising platform. To tackle the issue of data distribution shift between internal and external platforms, we introduce an Adaptive Difference Siamese Network (ADSNet), which employs cross-domain transfer learning to prevent negative transfer. Specifically, ADSNet is designed to learn information that is beneficial to the target domain. We introduce a gain evaluation strategy to calculate information gain, aiding the model in learning helpful information for the target domain and providing the ability to reject noisy samples, thus avoiding negative transfer. Additionally, we also design a Domain Adaptation Module as a bridge to connect different domains, reduce the distribution distance between them, and enhance the consistency of representation space distribution. We conduct extensive offline experiments and online A/B tests on a real advertising platform. Our proposed ADSNet method outperforms other methods, improving GINI by 2$\%$. The ablation study highlights the importance of the gain evaluation strategy in negative gain sample rejection and improving model performance. Additionally, ADSNet significantly improves long-tail prediction. The online A/B tests confirm ADSNet's efficacy, increasing online LTV by 3.47$\%$ and GMV by 3.89$\%$., Comment: Accepted to KDD 2024
- Published
- 2024
33. On the Nonlinearity of Layer Normalization
- Author
-
Ni, Yunhao, Guo, Yuxin, Jia, Junlong, and Huang, Lei
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence - Abstract
Layer normalization (LN) is a ubiquitous technique in deep learning but our theoretical understanding to it remains elusive. This paper investigates a new theoretical direction for LN, regarding to its nonlinearity and representation capacity. We investigate the representation capacity of a network with layerwise composition of linear and LN transformations, referred to as LN-Net. We theoretically show that, given $m$ samples with any label assignment, an LN-Net with only 3 neurons in each layer and $O(m)$ LN layers can correctly classify them. We further show the lower bound of the VC dimension of an LN-Net. The nonlinearity of LN can be amplified by group partition, which is also theoretically demonstrated with mild assumption and empirically supported by our experiments. Based on our analyses, we consider to design neural architecture by exploiting and amplifying the nonlinearity of LN, and the effectiveness is supported by our experiments., Comment: 42 pages, accepted to ICML 2024
- Published
- 2024
34. Efficient Data Distribution Estimation for Accelerated Federated Learning
- Author
-
Wang, Yuanli and Huang, Lei
- Subjects
Computer Science - Distributed, Parallel, and Cluster Computing ,Computer Science - Machine Learning - Abstract
Federated Learning(FL) is a privacy-preserving machine learning paradigm where a global model is trained in-situ across a large number of distributed edge devices. These systems are often comprised of millions of user devices and only a subset of available devices can be used for training in each epoch. Designing a device selection strategy is challenging, given that devices are highly heterogeneous in both their system resources and training data. This heterogeneity makes device selection very crucial for timely model convergence and sufficient model accuracy. To tackle the FL client heterogeneity problem, various client selection algorithms have been developed, showing promising performance improvement in terms of model coverage and accuracy. In this work, we study the overhead of client selection algorithms in a large scale FL environment. Then we propose an efficient data distribution summary calculation algorithm to reduce the overhead in a real-world large scale FL environment. The evaluation shows that our proposed solution could achieve up to 30x reduction in data summary time, and up to 360x reduction in clustering time.
- Published
- 2024
35. CoR-GS: Sparse-View 3D Gaussian Splatting via Co-Regularization
- Author
-
Zhang, Jiawei, Li, Jiahe, Yu, Xiaohan, Huang, Lei, Gu, Lin, Zheng, Jin, and Bai, Xiao
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
3D Gaussian Splatting (3DGS) creates a radiance field consisting of 3D Gaussians to represent a scene. With sparse training views, 3DGS easily suffers from overfitting, negatively impacting rendering. This paper introduces a new co-regularization perspective for improving sparse-view 3DGS. When training two 3D Gaussian radiance fields, we observe that the two radiance fields exhibit point disagreement and rendering disagreement that can unsupervisedly predict reconstruction quality, stemming from the randomness of densification implementation. We further quantify the two disagreements and demonstrate the negative correlation between them and accurate reconstruction, which allows us to identify inaccurate reconstruction without accessing ground-truth information. Based on the study, we propose CoR-GS, which identifies and suppresses inaccurate reconstruction based on the two disagreements: (1) Co-pruning considers Gaussians that exhibit high point disagreement in inaccurate positions and prunes them. (2) Pseudo-view co-regularization considers pixels that exhibit high rendering disagreement are inaccurate and suppress the disagreement. Results on LLFF, Mip-NeRF360, DTU, and Blender demonstrate that CoR-GS effectively regularizes the scene geometry, reconstructs the compact representations, and achieves state-of-the-art novel view synthesis quality under sparse training views., Comment: Accepted at ECCV 2024. Project page: https://jiaw-z.github.io/CoR-GS/
- Published
- 2024
36. TinyLLaVA Factory: A Modularized Codebase for Small-scale Large Multimodal Models
- Author
-
Jia, Junlong, Hu, Ying, Weng, Xi, Shi, Yiming, Li, Miao, Zhang, Xingjian, Zhou, Baichuan, Liu, Ziyu, Luo, Jie, Huang, Lei, and Wu, Ji
- Subjects
Computer Science - Machine Learning - Abstract
We present TinyLLaVA Factory, an open-source modular codebase for small-scale large multimodal models (LMMs) with a focus on simplicity of code implementations, extensibility of new features, and reproducibility of training results. Following the design philosophy of the factory pattern in software engineering, TinyLLaVA Factory modularizes the entire system into interchangeable components, with each component integrating a suite of cutting-edge models and methods, meanwhile leaving room for extensions to more features. In addition to allowing users to customize their own LMMs, TinyLLaVA Factory provides popular training recipes to let users pretrain and finetune their models with less coding effort. Empirical experiments validate the effectiveness of our codebase. The goal of TinyLLaVA Factory is to assist researchers and practitioners in exploring the wide landscape of designing and training small-scale LMMs with affordable computational resources., Comment: Our codebase is made public at https://github.com/TinyLLaVA/TinyLLaVA_Factory with documentation available at https://tinyllava-factory.readthedocs.io/en/latest/
- Published
- 2024
37. Probing orbits of stellar mass objects deep in galactic nuclei with quasi-periodic eruptions -- II: population analysis
- Author
-
Zhou, Cong, Zhong, Binyu, Zeng, Yuhe, Huang, Lei, and Pan, Zhen
- Subjects
Astrophysics - High Energy Astrophysical Phenomena ,General Relativity and Quantum Cosmology - Abstract
Quasi-periodic eruptions (QPEs) are intense repeating soft X-ray bursts with recurrence times about a few hours to a few weeks from galactic nuclei. Though the debates on the origin of QPEs have not completely settled down, more and more analyses favor the interpretation that QPEs are the result of collisions between a stellar mass object (a stellar mass black hole or a main sequence star) and an accretion disk around a supermassive black hole (SMBH) in galactic nuclei. If this interpretation is correct, QPEs will be invaluable in probing the orbits of stellar mass objects in the vicinity of SMBHs, and further inferring the formation of extreme mass ratio inspirals (EMRIs), one of the major targets of spaceborne gravitational wave missions. In this work, we extended the EMRI orbital analysis in Paper I arXiv:2401.11190 to all the known QPE sources with more than $6$ flares observed. Among all the analyzed 5 QPE sources, two distinct EMRI populations are identified: 4 EMRIs are of low orbital eccentricity (consistent with 0) which should be born in the wet EMRI formation channel, and 1 mildly eccentric EMRI (with $e= 0.25^{+0.18}_{-0.20}$ at 2-$\sigma$ confidence level) is consistent with the predictions of both the dry loss-cone formation channel and the Hills mechanism., Comment: 23 pages, 16 figures
- Published
- 2024
- Full Text
- View/download PDF
38. Scene inconsistency effect in object judgement: evidence from semantic and syntactic separation
- Author
-
Li, Wenrui, Ma, Xiaofang, Huang, Lei, and Guan, Jian
- Published
- 2025
- Full Text
- View/download PDF
39. Real-world safety of maribavir: a retrospective study based on signal detection in the FDA adverse event reporting system
- Author
-
Gong, Qineng, Zhang, Linlin, Wu, Hui, Miao, Yang, Huang, Lei, Yin, Cunlin, and Hu, Ye
- Published
- 2025
- Full Text
- View/download PDF
40. Senescence-related Genes as Prognostic Markers for STEMI Patients: LASSO Regression-Based Bioinformatics and External Validation
- Author
-
Wang, Xing-jie, Huang, Lei, Hou, Min, and Guo, Jie
- Published
- 2025
- Full Text
- View/download PDF
41. Boundedness and stabilization in a two-species competition system with density-dependent motility and indirect signal production: Boundedness and stabilization in a two-species competition...
- Author
-
Huang, Lei, Zeng, Fugeng, Zhou, Luxu, and Lu, Youjun
- Published
- 2025
- Full Text
- View/download PDF
42. Changes in global fluvial sediment concentrations and fluxes between 1985 and 2020
- Author
-
Sun, Xianghan, Tian, Liqiao, Fang, Hongwei, Walling, Des E., Huang, Lei, Park, Edward, Li, Deren, Zheng, Chunmiao, and Feng, Lian
- Published
- 2025
- Full Text
- View/download PDF
43. Monoallelic expression can govern penetrance of inborn errors of immunity
- Author
-
Stewart, O’Jay, Gruber, Conor, Randolph, Haley E., Patel, Roosheel, Ramba, Meredith, Calzoni, Enrica, Huang, Lei Haley, Levy, Jay, Buta, Sofija, Lee, Angelica, Sazeides, Christos, Prue, Zoe, Hoytema van Konijnenburg, David P., Chinn, Ivan K., Pedroza, Luis A., Lupski, James R., Schmitt, Erica G., Cooper, Megan A., Puel, Anne, Peng, Xiao, Boisson-Dupuis, Stéphanie, Bustamante, Jacinta, Okada, Satoshi, Martin-Fernandez, Marta, Orange, Jordan S., Casanova, Jean-Laurent, Milner, Joshua D., and Bogunovic, Dusan
- Published
- 2025
- Full Text
- View/download PDF
44. Ultrasound-Targeted β-Catenin Gene Therapy Improves the Cardiac Function in Mice After Myocardial Infarction
- Author
-
Yang, Lei, Gao, Tong, Huang, Yu, Wang, Pei-he, Han, Xin-hao, Wu, Jie, Huang, Lei, Da, Qing-en, Ouyang, Kun-fu, Han, Zhen, Tian, Hai, and Sun, Lu
- Published
- 2025
- Full Text
- View/download PDF
45. Numerical simulation of seismic waves in transversely isotropic media based on orthogonal body-fitted grids
- Author
-
Liu, Zhi-qiang, Li, Gang-zhu, Huang, Lei, Niu, Xing-guo, Zhang, Xiao-meng, and Gao, Cheng
- Published
- 2024
- Full Text
- View/download PDF
46. Ant nest-like WO3 films for improving electrochromic and energy-storage dual-functional performance by the surface modification of N-doped carbon
- Author
-
Xie, Wan-Ting, Xiao, Ming-Qing, Huang, Lei, Qiu, Qing-Qing, Li, Huan, Qi, Xiao-Peng, and Zeng, Jin-Ming
- Published
- 2024
- Full Text
- View/download PDF
47. Differential stress responsiveness determines intraspecies virulence heterogeneity and host adaptation in Listeria monocytogenes
- Author
-
Hafner, Lukas, Gadin, Enzo, Huang, Lei, Frouin, Arthur, Laporte, Fabien, Gaultier, Charlotte, Vieira, Afonso, Maudet, Claire, Varet, Hugo, Moura, Alexandra, Bracq-Dieye, Hélène, Tessaud-Rita, Nathalie, Maury, Mylène, Dazas, Melody, Legendre, Rachel, Gastineau, Pauline, Tsai, Yu-Huan, Coppée, Jean-Yves, Charlier, Caroline, Patin, Etienne, Chikhi, Rayan, Rocha, Eduardo P. C., Leclercq, Alexandre, Disson, Olivier, Aschard, Hugues, and Lecuit, Marc
- Published
- 2024
- Full Text
- View/download PDF
48. A modified machine learning algorithm for multi-collinearity environmental data
- Author
-
Tian, Haitao, Huang, Lei, Hu, Shouri, and Wu, Wangqi
- Published
- 2024
- Full Text
- View/download PDF
49. Machine learning force field study of carboxylate ligands on the surface of zinc-blende CdSe quantum dots
- Author
-
Zhang, Haibing, Cao, Bichuan, Huang, Lei, Peng, Xiaogang, and Wang, Linjun
- Published
- 2024
- Full Text
- View/download PDF
50. In-hospital outcomes of older patients with gastric cancer and their risk factors: large comprehensive institution-based study
- Author
-
Huang, Lei, Liu, Yunmei, Wang, Lei, Rong, Lan, and Hu, Weiguo
- Published
- 2024
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.