92,570 results on '"Mahajan A."'
Search Results
2. On the smallness of charm loop effects in $B\to K^{(*)} \ell\ell$ at low $q^2$: light meson Distribution Amplitude analysis
- Author
-
Mahajan, Namit and Mishra, Dayanand
- Subjects
High Energy Physics - Phenomenology - Abstract
The non-local effects originating from the charm quark loops at dilepton invariant masses smaller than the charmonium threshold in $B\to K \ell\ell$ are evaluated with light meson distribution amplitudes. The revised estimates with B-meson distribution amplitude within a Light Cone Sum Rule approach yielded results about three orders smaller than the original computation. In view of the importance of these non-factorizable soft gluon effects, both conceptually and phenomenologically, an independent evaluation is necessary. It is found that to twist-4 accuracy, these soft gluon effects vanish when evaluated employing the kaon distribution amplitude. Similar results hold for $B\to K^* \ell\ell$ to the leading twist. This eliminates one of the major sources of potential uncertainty which usually makes it difficult for a clear case of new physics, should the data show deviations from the standard model.
- Published
- 2024
3. A Prototype Model of Zero-Trust Architecture Blockchain with EigenTrust-Based Practical Byzantine Fault Tolerance Protocol to Manage Decentralized Clinical Trials
- Author
-
Peepliwall, Ashok Kumar, Pandey, Hari Mohan, Prakash, Surya, Mahajan, Anand A, Chowhan, Sudhinder Singh, Kumar, Vinesh, and Sharma, Rahul
- Subjects
Computer Science - Cryptography and Security ,Computer Science - Emerging Technologies ,Computer Science - Information Retrieval - Abstract
The COVID-19 pandemic necessitated the emergence of decentralized Clinical Trials (DCTs) due to patient retention, accelerate trials, improve data accessibility, enable virtual care, and facilitate seamless communication through integrated systems. However, integrating systems in DCTs exposes clinical data to potential security threats, making them susceptible to theft at any stage, a high risk of protocol deviations, and monitoring issues. To mitigate these challenges, blockchain technology serves as a secure framework, acting as a decentralized ledger, creating an immutable environment by establishing a zero-trust architecture, where data are deemed untrusted until verified. In combination with Internet of Things (IoT)-enabled wearable devices, blockchain secures the transfer of clinical trial data on private blockchains during DCT automation and operations. This paper proposes a prototype model of the Zero-Trust Architecture Blockchain (z-TAB) to integrate patient-generated clinical trial data during DCT operation management. The EigenTrust-based Practical Byzantine Fault Tolerance (T-PBFT) algorithm has been incorporated as a consensus protocol, leveraging Hyperledger Fabric. Furthermore, the Internet of Things (IoT) has been integrated to streamline data processing among stakeholders within the blockchain platforms. Rigorous evaluation has been done to evaluate the quality of the system., Comment: NA
- Published
- 2024
4. No Thick Atmosphere on the Terrestrial Exoplanet Gl 486b
- Author
-
Mansfield, Megan Weiner, Xue, Qiao, Zhang, Michael, Mahajan, Alexandra S., Ih, Jegug, Koll, Daniel, Bean, Jacob L., Coy, Brandon Park, Eastman, Jason D., Kempton, Eliza M. -R., Kite, Edwin S., and Lunine, Jonathan
- Subjects
Astrophysics - Earth and Planetary Astrophysics - Abstract
A primary science goal for JWST is to detect and characterize the atmospheres of terrestrial planets orbiting M dwarfs (M-Earths). The existence of atmospheres on M-Earths is highly uncertain because their host stars' extended history of high XUV irradiation may act to completely remove their atmospheres. We present two JWST secondary eclipse observations of the M-Earth Gl 486b (also known as GJ 486b) between 5-12 $\mu$m. We combined these observations with a precise analysis of the host star parameters to derive a planetary dayside temperature of $T_{p}=865 \pm 14$ K. We compared this temperature to the maximum expected temperature for a zero albedo, zero heat redistribution bare rock and derived a temperature ratio of $R=\frac{T_{p,dayside}}{T_{p,max}}=0.97 \pm 0.01$. This value is consistent with an airless body with a slight non-zero albedo or a thin atmosphere with $<1$% H$_{2}$O or $<1$ ppm CO$_{2}$. However, it is inconsistent with an Earth- or Venus-like atmosphere, and the spectrum shows no clear emission or absorption features. Additionally, our observations are inconsistent with the water-rich atmospheric scenario allowed by previous transit observations and suggest the transmission spectrum was instead shaped by stellar contamination (Moran et al. 2023). Given the potential for atmospheric escape throughout the system's $\geq6.6$-Gyr lifetime (Diamond-Lowe et al. 2024), we conclude that the observations are likely best explained by an airless planet. This result is the most precise measurement yet of terrestrial exoplanet thermal emission with JWST, which places a strong constraint on the position of the "Cosmic Shoreline" between airless bodies and those with atmospheres., Comment: 13 pages, 6 figures, submitted to ApJL
- Published
- 2024
5. Efficient FGM optimization with a novel design space and DeepONet
- Author
-
Agrawal, Piyush, Mahajan, Ihina, Choubey, Shivam, and Agrawal, Manish
- Subjects
Computer Science - Computational Engineering, Finance, and Science - Abstract
This manuscript proposes an optimization framework to find the tailor-made functionally graded material (FGM) profiles for thermoelastic applications. This optimization framework consists of (1) a random profile generation scheme, (2) deep learning (DL) based surrogate models for the prediction of thermal and structural quantities, and (3) a genetic algorithm (GA). From the proposed random profile generation scheme, we strive for a generic design space that does not contain impractical designs, i.e., profiles with sharp gradations. We also show that the power law is a strict subset of the proposed design space. We use a dense neural network-based surrogate model for the prediction of maximum stress, while the deep neural operator DeepONet is used for the prediction of the thermal field. The point-wise effective prediction of the thermal field enables us to implement the constraint that the metallic content of the FGM remains within a specified limit. The integration of the profile generation scheme and DL-based surrogate models with GA provides us with an efficient optimization scheme. The efficacy of the proposed framework is demonstrated through various numerical examples.
- Published
- 2024
6. JWST Thermal Emission of the Terrestrial Exoplanet GJ 1132b
- Author
-
Xue, Qiao, Bean, Jacob L., Zhang, Michael, Mahajan, Alexandra S., Ih, Jegug, Eastman, Jason D., Lunine, Jonathan I., Mansfield, Megan Weiner, Coy, Brandon P., Kempton, Eliza M. -R., Koll, Daniel D., and Kite, Edwin S.
- Subjects
Astrophysics - Earth and Planetary Astrophysics - Abstract
We present thermal emission measurements of GJ 1132b spanning 5--12 um obtained with the Mid-Infrared Instrument Low-Resolution Spectrometer (MIRI/LRS) on the James Webb Space Telescope (JWST). GJ 1132b is an M-dwarf rocky planet with Teq=584 K and an orbital period of 1.6 days. We measure a white-light secondary eclipse depth of 140+/-17 ppm, which corresponds to a dayside brightness temperature of Tp,dayside= 709+/-31 K using improved star and planet parameters. This measured temperature is only 1 sigma below the maximum possible dayside temperature of a bare rock (i.e., assuming a zero albedo planet with no heat redistribution, Tmax = 746+14/-11 K). The emission spectrum is consistent with a featureless blackbody, which agrees with a wide range of possible surface compositions. By comparing forward models to the dayside emission spectrum, we rule out Earth-thickness (P ~ 1 bar) atmospheres with at least 1% H2O, atmospheres of any modeled thickness (10^-4 -- 10^2 bar) that contain at least 1% CO2, and thick, Venus-like atmospheres (P>~100 bar) with at least 1 ppm CO2 or H2O. We therefore conclude that GJ 1132b likely does not have a significant atmosphere. This finding supports the concept of a universal 'Cosmic Shoreline' given the high level of bolometric and XUV irradiation received by the planet., Comment: Accepted by ApJL
- Published
- 2024
7. The Llama 3 Herd of Models
- Author
-
Dubey, Abhimanyu, Jauhri, Abhinav, Pandey, Abhinav, Kadian, Abhishek, Al-Dahle, Ahmad, Letman, Aiesha, Mathur, Akhil, Schelten, Alan, Yang, Amy, Fan, Angela, Goyal, Anirudh, Hartshorn, Anthony, Yang, Aobo, Mitra, Archi, Sravankumar, Archie, Korenev, Artem, Hinsvark, Arthur, Rao, Arun, Zhang, Aston, Rodriguez, Aurelien, Gregerson, Austen, Spataru, Ava, Roziere, Baptiste, Biron, Bethany, Tang, Binh, Chern, Bobbie, Caucheteux, Charlotte, Nayak, Chaya, Bi, Chloe, Marra, Chris, McConnell, Chris, Keller, Christian, Touret, Christophe, Wu, Chunyang, Wong, Corinne, Ferrer, Cristian Canton, Nikolaidis, Cyrus, Allonsius, Damien, Song, Daniel, Pintz, Danielle, Livshits, Danny, Esiobu, David, Choudhary, Dhruv, Mahajan, Dhruv, Garcia-Olano, Diego, Perino, Diego, Hupkes, Dieuwke, Lakomkin, Egor, AlBadawy, Ehab, Lobanova, Elina, Dinan, Emily, Smith, Eric Michael, Radenovic, Filip, Zhang, Frank, Synnaeve, Gabriel, Lee, Gabrielle, Anderson, Georgia Lewis, Nail, Graeme, Mialon, Gregoire, Pang, Guan, Cucurell, Guillem, Nguyen, Hailey, Korevaar, Hannah, Xu, Hu, Touvron, Hugo, Zarov, Iliyan, Ibarra, Imanol Arrieta, Kloumann, Isabel, Misra, Ishan, Evtimov, Ivan, Copet, Jade, Lee, Jaewon, Geffert, Jan, Vranes, Jana, Park, Jason, Mahadeokar, Jay, Shah, Jeet, van der Linde, Jelmer, Billock, Jennifer, Hong, Jenny, Lee, Jenya, Fu, Jeremy, Chi, Jianfeng, Huang, Jianyu, Liu, Jiawen, Wang, Jie, Yu, Jiecao, Bitton, Joanna, Spisak, Joe, Park, Jongsoo, Rocca, Joseph, Johnstun, Joshua, Saxe, Joshua, Jia, Junteng, Alwala, Kalyan Vasuden, Upasani, Kartikeya, Plawiak, Kate, Li, Ke, Heafield, Kenneth, Stone, Kevin, El-Arini, Khalid, Iyer, Krithika, Malik, Kshitiz, Chiu, Kuenley, Bhalla, Kunal, Rantala-Yeary, Lauren, van der Maaten, Laurens, Chen, Lawrence, Tan, Liang, Jenkins, Liz, Martin, Louis, Madaan, Lovish, Malo, Lubo, Blecher, Lukas, Landzaat, Lukas, de Oliveira, Luke, Muzzi, Madeline, Pasupuleti, Mahesh, Singh, Mannat, Paluri, Manohar, Kardas, Marcin, Oldham, Mathew, Rita, Mathieu, Pavlova, Maya, Kambadur, Melanie, Lewis, Mike, Si, Min, Singh, Mitesh Kumar, Hassan, Mona, Goyal, Naman, Torabi, Narjes, Bashlykov, Nikolay, Bogoychev, Nikolay, Chatterji, Niladri, Duchenne, Olivier, Çelebi, Onur, Alrassy, Patrick, Zhang, Pengchuan, Li, Pengwei, Vasic, Petar, Weng, Peter, Bhargava, Prajjwal, Dubal, Pratik, Krishnan, Praveen, Koura, Punit Singh, Xu, Puxin, He, Qing, Dong, Qingxiao, Srinivasan, Ragavan, Ganapathy, Raj, Calderer, Ramon, Cabral, Ricardo Silveira, Stojnic, Robert, Raileanu, Roberta, Girdhar, Rohit, Patel, Rohit, Sauvestre, Romain, Polidoro, Ronnie, Sumbaly, Roshan, Taylor, Ross, Silva, Ruan, Hou, Rui, Wang, Rui, Hosseini, Saghar, Chennabasappa, Sahana, Singh, Sanjay, Bell, Sean, Kim, Seohyun Sonia, Edunov, Sergey, Nie, Shaoliang, Narang, Sharan, Raparthy, Sharath, Shen, Sheng, Wan, Shengye, Bhosale, Shruti, Zhang, Shun, Vandenhende, Simon, Batra, Soumya, Whitman, Spencer, Sootla, Sten, Collot, Stephane, Gururangan, Suchin, Borodinsky, Sydney, Herman, Tamar, Fowler, Tara, Sheasha, Tarek, Georgiou, Thomas, Scialom, Thomas, Speckbacher, Tobias, Mihaylov, Todor, Xiao, Tong, Karn, Ujjwal, Goswami, Vedanuj, Gupta, Vibhor, Ramanathan, Vignesh, Kerkez, Viktor, Gonguet, Vincent, Do, Virginie, Vogeti, Vish, Petrovic, Vladan, Chu, Weiwei, Xiong, Wenhan, Fu, Wenyin, Meers, Whitney, Martinet, Xavier, Wang, Xiaodong, Tan, Xiaoqing Ellen, Xie, Xinfeng, Jia, Xuchao, Wang, Xuewei, Goldschlag, Yaelle, Gaur, Yashesh, Babaei, Yasmine, Wen, Yi, Song, Yiwen, Zhang, Yuchen, Li, Yue, Mao, Yuning, Coudert, Zacharie Delpierre, Yan, Zheng, Chen, Zhengxing, Papakipos, Zoe, Singh, Aaditya, Grattafiori, Aaron, Jain, Abha, Kelsey, Adam, Shajnfeld, Adam, Gangidi, Adithya, Victoria, Adolfo, Goldstand, Ahuva, Menon, Ajay, Sharma, Ajay, Boesenberg, Alex, Vaughan, Alex, Baevski, Alexei, Feinstein, Allie, Kallet, Amanda, Sangani, Amit, Yunus, Anam, Lupu, Andrei, Alvarado, Andres, Caples, Andrew, Gu, Andrew, Ho, Andrew, Poulton, Andrew, Ryan, Andrew, Ramchandani, Ankit, Franco, Annie, Saraf, Aparajita, Chowdhury, Arkabandhu, Gabriel, Ashley, Bharambe, Ashwin, Eisenman, Assaf, Yazdan, Azadeh, James, Beau, Maurer, Ben, Leonhardi, Benjamin, Huang, Bernie, Loyd, Beth, De Paola, Beto, Paranjape, Bhargavi, Liu, Bing, Wu, Bo, Ni, Boyu, Hancock, Braden, Wasti, Bram, Spence, Brandon, Stojkovic, Brani, Gamido, Brian, Montalvo, Britt, Parker, Carl, Burton, Carly, Mejia, Catalina, Wang, Changhan, Kim, Changkyu, Zhou, Chao, Hu, Chester, Chu, Ching-Hsiang, Cai, Chris, Tindal, Chris, Feichtenhofer, Christoph, Civin, Damon, Beaty, Dana, Kreymer, Daniel, Li, Daniel, Wyatt, Danny, Adkins, David, Xu, David, Testuggine, Davide, David, Delia, Parikh, Devi, Liskovich, Diana, Foss, Didem, Wang, Dingkang, Le, Duc, Holland, Dustin, Dowling, Edward, Jamil, Eissa, Montgomery, Elaine, Presani, Eleonora, Hahn, Emily, Wood, Emily, Brinkman, Erik, Arcaute, Esteban, Dunbar, Evan, Smothers, Evan, Sun, Fei, Kreuk, Felix, Tian, Feng, Ozgenel, Firat, Caggioni, Francesco, Guzmán, Francisco, Kanayet, Frank, Seide, Frank, Florez, Gabriela Medina, Schwarz, Gabriella, Badeer, Gada, Swee, Georgia, Halpern, Gil, Thattai, Govind, Herman, Grant, Sizov, Grigory, Guangyi, Zhang, Lakshminarayanan, Guna, Shojanazeri, Hamid, Zou, Han, Wang, Hannah, Zha, Hanwen, Habeeb, Haroun, Rudolph, Harrison, Suk, Helen, Aspegren, Henry, Goldman, Hunter, Damlaj, Ibrahim, Molybog, Igor, Tufanov, Igor, Veliche, Irina-Elena, Gat, Itai, Weissman, Jake, Geboski, James, Kohli, James, Asher, Japhet, Gaya, Jean-Baptiste, Marcus, Jeff, Tang, Jeff, Chan, Jennifer, Zhen, Jenny, Reizenstein, Jeremy, Teboul, Jeremy, Zhong, Jessica, Jin, Jian, Yang, Jingyi, Cummings, Joe, Carvill, Jon, Shepard, Jon, McPhie, Jonathan, Torres, Jonathan, Ginsburg, Josh, Wang, Junjie, Wu, Kai, U, Kam Hou, Saxena, Karan, Prasad, Karthik, Khandelwal, Kartikay, Zand, Katayoun, Matosich, Kathy, Veeraraghavan, Kaushik, Michelena, Kelly, Li, Keqian, Huang, Kun, Chawla, Kunal, Lakhotia, Kushal, Huang, Kyle, Chen, Lailin, Garg, Lakshya, A, Lavender, Silva, Leandro, Bell, Lee, Zhang, Lei, Guo, Liangpeng, Yu, Licheng, Moshkovich, Liron, Wehrstedt, Luca, Khabsa, Madian, Avalani, Manav, Bhatt, Manish, Tsimpoukelli, Maria, Mankus, Martynas, Hasson, Matan, Lennie, Matthew, Reso, Matthias, Groshev, Maxim, Naumov, Maxim, Lathi, Maya, Keneally, Meghan, Seltzer, Michael L., Valko, Michal, Restrepo, Michelle, Patel, Mihir, Vyatskov, Mik, Samvelyan, Mikayel, Clark, Mike, Macey, Mike, Wang, Mike, Hermoso, Miquel Jubert, Metanat, Mo, Rastegari, Mohammad, Bansal, Munish, Santhanam, Nandhini, Parks, Natascha, White, Natasha, Bawa, Navyata, Singhal, Nayan, Egebo, Nick, Usunier, Nicolas, Laptev, Nikolay Pavlovich, Dong, Ning, Zhang, Ning, Cheng, Norman, Chernoguz, Oleg, Hart, Olivia, Salpekar, Omkar, Kalinli, Ozlem, Kent, Parkin, Parekh, Parth, Saab, Paul, Balaji, Pavan, Rittner, Pedro, Bontrager, Philip, Roux, Pierre, Dollar, Piotr, Zvyagina, Polina, Ratanchandani, Prashant, Yuvraj, Pritish, Liang, Qian, Alao, Rachad, Rodriguez, Rachel, Ayub, Rafi, Murthy, Raghotham, Nayani, Raghu, Mitra, Rahul, Li, Raymond, Hogan, Rebekkah, Battey, Robin, Wang, Rocky, Maheswari, Rohan, Howes, Russ, Rinott, Ruty, Bondu, Sai Jayesh, Datta, Samyak, Chugh, Sara, Hunt, Sara, Dhillon, Sargun, Sidorov, Sasha, Pan, Satadru, Verma, Saurabh, Yamamoto, Seiji, Ramaswamy, Sharadh, Lindsay, Shaun, Feng, Sheng, Lin, Shenghao, Zha, Shengxin Cindy, Shankar, Shiva, Zhang, Shuqiang, Wang, Sinong, Agarwal, Sneha, Sajuyigbe, Soji, Chintala, Soumith, Max, Stephanie, Chen, Stephen, Kehoe, Steve, Satterfield, Steve, Govindaprasad, Sudarshan, Gupta, Sumit, Cho, Sungmin, Virk, Sunny, Subramanian, Suraj, Choudhury, Sy, Goldman, Sydney, Remez, Tal, Glaser, Tamar, Best, Tamara, Kohler, Thilo, Robinson, Thomas, Li, Tianhe, Zhang, Tianjun, Matthews, Tim, Chou, Timothy, Shaked, Tzook, Vontimitta, Varun, Ajayi, Victoria, Montanez, Victoria, Mohan, Vijai, Kumar, Vinay Satish, Mangla, Vishal, Albiero, Vítor, Ionescu, Vlad, Poenaru, Vlad, Mihailescu, Vlad Tiberiu, Ivanov, Vladimir, Li, Wei, Wang, Wenchen, Jiang, Wenwen, Bouaziz, Wes, Constable, Will, Tang, Xiaocheng, Wang, Xiaofang, Wu, Xiaojian, Wang, Xiaolan, Xia, Xide, Wu, Xilun, Gao, Xinbo, Chen, Yanjun, Hu, Ye, Jia, Ye, Qi, Ye, Li, Yenda, Zhang, Yilin, Zhang, Ying, Adi, Yossi, Nam, Youngjin, Yu, Wang, Hao, Yuchen, Qian, Yundi, He, Yuzi, Rait, Zach, DeVito, Zachary, Rosnbrick, Zef, Wen, Zhaoduo, Yang, Zhenyu, and Zhao, Zhiwei
- Subjects
Computer Science - Artificial Intelligence ,Computer Science - Computation and Language ,Computer Science - Computer Vision and Pattern Recognition - Abstract
Modern artificial intelligence (AI) systems are powered by foundation models. This paper presents a new set of foundation models, called Llama 3. It is a herd of language models that natively support multilinguality, coding, reasoning, and tool usage. Our largest model is a dense Transformer with 405B parameters and a context window of up to 128K tokens. This paper presents an extensive empirical evaluation of Llama 3. We find that Llama 3 delivers comparable quality to leading language models such as GPT-4 on a plethora of tasks. We publicly release Llama 3, including pre-trained and post-trained versions of the 405B parameter language model and our Llama Guard 3 model for input and output safety. The paper also presents the results of experiments in which we integrate image, video, and speech capabilities into Llama 3 via a compositional approach. We observe this approach performs competitively with the state-of-the-art on image, video, and speech recognition tasks. The resulting models are not yet being broadly released as they are still under development.
- Published
- 2024
8. SSPACE Astrobiology Payload-1 (SAP-1)
- Author
-
Lokaveer, A, Anjana, Thomas, Yasir, Maliyekkal, Yogahariharan, S, Dewangan, Akash, Mahajan, Saurabh Kishor, Tembhurne, Sakshi Aravind, Gupta, Gunja Subhash, Bhalla, Devashish, Dhruva, Anantha Datta, Kumar, Aloke, Viswanathan, Koushik, Khaire, Vikram, Narayanan, Anand, and Hari, Priyadarshnam
- Subjects
Astrophysics - Instrumentation and Methods for Astrophysics ,Astrophysics - Earth and Planetary Astrophysics ,Physics - Biological Physics - Abstract
The SSPACE Astrobiology Payload (SAP) series, starting with the SAP-1 project is designed to conduct in-situ microbiology experiments in low earth orbit. This payload series aims to understand the behaviour of microbial organisms in space, particularly those critical for human health, and the corresponding effects due to microgravity and solar/galactic radiation. SAP-1 focuses on studying Bacillus clausii and Bacillus coagulans, bacteria beneficial to humans. It aims to provide a space laboratory for astrobiology experiments under microgravity conditions. The hardware developed for these experiments is indigenous and tailored to meet the unique requirements of autonomous microbiology experiments by controlling pressure, temperature, and nutrition flow to bacteria. A rotating platform, which forms the core design, is innovatively utilised to regulate the flow and mixing of nutrients with dormant bacteria. The technology demonstration models developed at SSPACE have yielded promising results, with ongoing efforts to refine, adapt for space conditions, and prepare for integration with nanosatellites or space modules. The anticipated payload will be compact, approximately 1U in size (10cm x 10cm x 10cm), consume less than 5W power, and offer flexibility for various microbiological studies., Comment: 20 Pages, Submitted to Advances in Space Research
- Published
- 2024
9. Discover-then-Name: Task-Agnostic Concept Bottlenecks via Automated Concept Discovery
- Author
-
Rao, Sukrut, Mahajan, Sweta, Böhle, Moritz, and Schiele, Bernt
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning - Abstract
Concept Bottleneck Models (CBMs) have recently been proposed to address the 'black-box' problem of deep neural networks, by first mapping images to a human-understandable concept space and then linearly combining concepts for classification. Such models typically require first coming up with a set of concepts relevant to the task and then aligning the representations of a feature extractor to map to these concepts. However, even with powerful foundational feature extractors like CLIP, there are no guarantees that the specified concepts are detectable. In this work, we leverage recent advances in mechanistic interpretability and propose a novel CBM approach -- called Discover-then-Name-CBM (DN-CBM) -- that inverts the typical paradigm: instead of pre-selecting concepts based on the downstream classification task, we use sparse autoencoders to first discover concepts learnt by the model, and then name them and train linear probes for classification. Our concept extraction strategy is efficient, since it is agnostic to the downstream task, and uses concepts already known to the model. We perform a comprehensive evaluation across multiple datasets and CLIP architectures and show that our method yields semantically meaningful concepts, assigns appropriate names to them that make them easy to interpret, and yields performant and interpretable CBMs. Code available at https://github.com/neuroexplicit-saar/discover-then-name., Comment: 40 pages, 21 figures, 6 tables, European Conference on Computer Vision (ECCV) 2024
- Published
- 2024
10. MeshFeat: Multi-Resolution Features for Neural Fields on Meshes
- Author
-
Mahajan, Mihir, Hofherr, Florian, and Cremers, Daniel
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Machine Learning - Abstract
Parametric feature grid encodings have gained significant attention as an encoding approach for neural fields since they allow for much smaller MLPs, which significantly decreases the inference time of the models. In this work, we propose MeshFeat, a parametric feature encoding tailored to meshes, for which we adapt the idea of multi-resolution feature grids from Euclidean space. We start from the structure provided by the given vertex topology and use a mesh simplification algorithm to construct a multi-resolution feature representation directly on the mesh. The approach allows the usage of small MLPs for neural fields on meshes, and we show a significant speed-up compared to previous representations while maintaining comparable reconstruction quality for texture reconstruction and BRDF representation. Given its intrinsic coupling to the vertices, the method is particularly well-suited for representations on deforming meshes, making it a good fit for object animation., Comment: To appear at European Conference on Computer Vision (ECCV), 2024
- Published
- 2024
11. Integrated Hardware Architecture and Device Placement Search
- Author
-
Wang, Irene, Tarnawski, Jakub, Phanishayee, Amar, and Mahajan, Divya
- Subjects
Computer Science - Machine Learning ,Computer Science - Hardware Architecture ,Computer Science - Distributed, Parallel, and Cluster Computing - Abstract
Distributed execution of deep learning training involves a dynamic interplay between hardware accelerator architecture and device placement strategy. This is the first work to explore the co-optimization of determining the optimal architecture and device placement strategy through novel algorithms, improving the balance of computational resources, memory usage, and data distribution. Our architecture search leverages tensor and vector units, determining their quantity and dimensionality, and on-chip and off-chip memory configurations. It also determines the microbatch size and decides whether to recompute or stash activations, balancing the memory footprint of training and storage size. For each explored architecture configuration, we use an Integer Linear Program (ILP) to find the optimal schedule for executing operators on the accelerator. The ILP results then integrate with a dynamic programming solution to identify the most effective device placement strategy, combining data, pipeline, and tensor model parallelism across multiple accelerators. Our approach achieves higher throughput on large language models compared to the state-of-the-art TPUv4 and the Spotlight accelerator search framework. The entire source code of PHAZE is available at https://github.com/msr-fiddle/phaze., Comment: Accepted at the 41st International Conference on Machine Learning (ICML), 2024
- Published
- 2024
12. Data-driven Forecasting of Deep Learning Performance on GPUs
- Author
-
Lee, Seonho, Phanishayee, Amar, and Mahajan, Divya
- Subjects
Computer Science - Machine Learning ,Computer Science - Performance - Abstract
Deep learning kernels exhibit predictable memory accesses and compute patterns, making GPUs' parallel architecture well-suited for their execution. Software and runtime systems for GPUs are optimized to better utilize the stream multiprocessors, on-chip cache, and off-chip high-bandwidth memory. As deep learning models and GPUs evolve, access to newer GPUs is often limited, raising questions about the performance of new model architectures on existing GPUs, existing models on new GPUs, and new model architectures on new GPUs. To address these questions, we introduce NeuSight, a framework to predict the performance of various deep learning models, for both training and inference, on unseen GPUs without requiring actual execution. The framework leverages both GPU hardware behavior and software library optimizations to estimate end-to-end performance. Previous work uses regression models that capture linear trends or multilayer perceptrons to predict the overall latency of deep learning kernels on GPUs. These approaches suffer from higher error percentages when forecasting performance on unseen models and new GPUs. Instead, NeuSight decomposes the prediction problem into smaller problems, bounding the prediction through fundamental performance laws. NeuSight decomposes a single deep learning kernel prediction into smaller working sets called tiles, which are executed independently on the GPU. Tile-granularity predictions are determined using a machine learning approach and aggregated to estimate end-to-end latency. NeuSight outperforms prior work across various deep learning workloads and the latest GPUs. It reduces the percentage error from 198% and 19.7% to 3.8% in predicting the latency of GPT3 model for training and inference on H100, compared to state-of-the-art prior works, where both GPT3 and H100 were not used to train the framework.
- Published
- 2024
13. New insights into the internal structure of GJ 1214 b informed by JWST
- Author
-
Nixon, Matthew C., Piette, Anjali A. A., Kempton, Eliza M. -R., Gao, Peter, Bean, Jacob L., Steinrueck, Maria E., Mahajan, Alexandra S., Eastman, Jason D., Zhang, Michael, and Rogers, Leslie A.
- Subjects
Astrophysics - Earth and Planetary Astrophysics - Abstract
Recent JWST observations of the sub-Neptune GJ 1214 b suggest that it hosts a high-metallicity (>100x solar), hazy atmosphere. Emission spectra of the planet show molecular absorption features, most likely due to atmospheric H2O. In light of this new information, we conduct a thorough reevaluation of the planet's internal structure. We consider interior models with mixed H/He/H2O envelopes of varying composition, informed by atmospheric constraints from the JWST phase curve, in order to determine possible bulk compositions and internal structures. Self-consistent atmospheric models consistent with the JWST observations are used to set boundary conditions for the interior. We find that a total envelope mass fraction of at least 8.1% is required to explain the planet's mass and radius. Regardless of H2O content, the maximum H/He mass fraction of the planet is 5.8%. We find that a 1:1 ice-to-rock ratio along with 3.4-4.8% H/He is also a permissible solution. In addition, we consider a pure H2O (steam) envelope and find that such a scenario is possible, albeit with a high ice-to-rock ratio of at least 3.76:1, which may be unrealistic from a planet formation standpoint. We discuss possible formation pathways for the different internal structures that are consistent with observations. Since our results depend strongly on the atmospheric composition and haze properties, more precise observations of the planet's atmosphere would allow for further constraints on its internal structure. This type of analysis can be applied to any sub-Neptune with atmospheric constraints to better understand its interior., Comment: Accepted for publication in ApJ Letters. 13 pages, 6 figures, 2 tables
- Published
- 2024
14. Long-Horizon Planning for Multi-Agent Robots in Partially Observable Environments
- Author
-
Nayak, Siddharth, Orozco, Adelmo Morrison, Have, Marina Ten, Thirumalai, Vittal, Zhang, Jackson, Chen, Darren, Kapoor, Aditya, Robinson, Eric, Gopalakrishnan, Karthik, Harrison, James, Ichter, Brian, Mahajan, Anuj, and Balakrishnan, Hamsa
- Subjects
Computer Science - Robotics ,Computer Science - Multiagent Systems - Abstract
The ability of Language Models (LMs) to understand natural language makes them a powerful tool for parsing human instructions into task plans for autonomous robots. Unlike traditional planning methods that rely on domain-specific knowledge and handcrafted rules, LMs generalize from diverse data and adapt to various tasks with minimal tuning, acting as a compressed knowledge base. However, LMs in their standard form face challenges with long-horizon tasks, particularly in partially observable multi-agent settings. We propose an LM-based Long-Horizon Planner for Multi-Agent Robotics (LLaMAR), a cognitive architecture for planning that achieves state-of-the-art results in long-horizon tasks within partially observable environments. LLaMAR employs a plan-act-correct-verify framework, allowing self-correction from action execution feedback without relying on oracles or simulators. Additionally, we present MAP-THOR, a comprehensive test suite encompassing household tasks of varying complexity within the AI2-THOR environment. Experiments show that LLaMAR achieves a 30% higher success rate compared to other state-of-the-art LM-based multi-agent planners., Comment: 27 pages, 4 figures, 5 tables
- Published
- 2024
15. Giant graviton expansion from eigenvalue instantons
- Author
-
Chen, Yiming, Mahajan, Raghu, and Tang, Haifeng
- Subjects
High Energy Physics - Theory - Abstract
Recently, S. Murthy has proposed a convergent expansion of free partition functions and superconformal indices of finite-$N$ purely adjoint gauge theories based on a Fredholm determinant expansion. This expansion has been dubbed the giant graviton expansion and takes the form of an infinite series of corrections to the $N=\infty$ result, with the $m^\text{th}$ correction being of order $e^{-mN}$. We show that this expansion can be reproduced using eigenvalue instantons in unitary matrix integrals. This perspective allows us to get the giant graviton expansion proposed by S. Murthy without the intermediate step of the Hubbard Stratonovich transformation., Comment: 12 pages, 1 figure
- Published
- 2024
16. Periodic agent-state based Q-learning for POMDPs
- Author
-
Sinha, Amit, Geist, Mathieu, and Mahajan, Aditya
- Subjects
Computer Science - Machine Learning - Abstract
The standard approach for Partially Observable Markov Decision Processes (POMDPs) is to convert them to a fully observed belief-state MDP. However, the belief state depends on the system model and is therefore not viable in reinforcement learning (RL) settings. A widely used alternative is to use an agent state, which is a model-free, recursively updateable function of the observation history. Examples include frame stacking and recurrent neural networks. Since the agent state is model-free, it is used to adapt standard RL algorithms to POMDPs. However, standard RL algorithms like Q-learning learn a stationary policy. Our main thesis that we illustrate via examples is that because the agent state does not satisfy the Markov property, non-stationary agent-state based policies can outperform stationary ones. To leverage this feature, we propose PASQL (periodic agent-state based Q-learning), which is a variant of agent-state-based Q-learning that learns periodic policies. By combining ideas from periodic Markov chains and stochastic approximation, we rigorously establish that PASQL converges to a cyclic limit and characterize the approximation error of the converged periodic policy. Finally, we present a numerical experiment to highlight the salient features of PASQL and demonstrate the benefit of learning periodic policies over stationary policies.
- Published
- 2024
17. Learning Disentangled Representation in Object-Centric Models for Visual Dynamics Prediction via Transformers
- Author
-
Gandhi, Sanket, Atul, Mahajan, Samanyu, Sharma, Vishal, Gupta, Rushil, Mondal, Arnab Kumar, and Singla, Parag
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
Recent work has shown that object-centric representations can greatly help improve the accuracy of learning dynamics while also bringing interpretability. In this work, we take this idea one step further, ask the following question: "can learning disentangled representation further improve the accuracy of visual dynamics prediction in object-centric models?" While there has been some attempt to learn such disentangled representations for the case of static images \citep{nsb}, to the best of our knowledge, ours is the first work which tries to do this in a general setting for video, without making any specific assumptions about the kind of attributes that an object might have. The key building block of our architecture is the notion of a {\em block}, where several blocks together constitute an object. Each block is represented as a linear combination of a given number of learnable concept vectors, which is iteratively refined during the learning process. The blocks in our model are discovered in an unsupervised manner, by attending over object masks, in a style similar to discovery of slots \citep{slot_attention}, for learning a dense object-centric representation. We employ self-attention via transformers over the discovered blocks to predict the next state resulting in discovery of visual dynamics. We perform a series of experiments on several benchmark 2-D, and 3-D datasets demonstrating that our architecture (1) can discover semantically meaningful blocks (2) help improve accuracy of dynamics prediction compared to SOTA object-centric models (3) perform significantly better in OOD setting where the specific attribute combinations are not seen earlier during training. Our experiments highlight the importance discovery of disentangled representation for visual dynamics prediction.
- Published
- 2024
18. AgentInstruct: Toward Generative Teaching with Agentic Flows
- Author
-
Mitra, Arindam, Del Corro, Luciano, Zheng, Guoqing, Mahajan, Shweti, Rouhana, Dany, Codas, Andres, Lu, Yadong, Chen, Wei-ge, Vrousgos, Olga, Rosset, Corby, Silva, Fillipe, Khanpour, Hamed, Lara, Yash, and Awadallah, Ahmed
- Subjects
Computer Science - Artificial Intelligence ,Computer Science - Computation and Language ,Computer Science - Machine Learning - Abstract
Synthetic data is becoming increasingly important for accelerating the development of language models, both large and small. Despite several successful use cases, researchers also raised concerns around model collapse and drawbacks of imitating other models. This discrepancy can be attributed to the fact that synthetic data varies in quality and diversity. Effective use of synthetic data usually requires significant human effort in curating the data. We focus on using synthetic data for post-training, specifically creating data by powerful models to teach a new skill or behavior to another model, we refer to this setting as Generative Teaching. We introduce AgentInstruct, an extensible agentic framework for automatically creating large amounts of diverse and high-quality synthetic data. AgentInstruct can create both the prompts and responses, using only raw data sources like text documents and code files as seeds. We demonstrate the utility of AgentInstruct by creating a post training dataset of 25M pairs to teach language models different skills, such as text editing, creative writing, tool usage, coding, reading comprehension, etc. The dataset can be used for instruction tuning of any base model. We post-train Mistral-7b with the data. When comparing the resulting model Orca-3 to Mistral-7b-Instruct (which uses the same base model), we observe significant improvements across many benchmarks. For example, 40% improvement on AGIEval, 19% improvement on MMLU, 54% improvement on GSM8K, 38% improvement on BBH and 45% improvement on AlpacaEval. Additionally, it consistently outperforms other models such as LLAMA-8B-instruct and GPT-3.5-turbo.
- Published
- 2024
19. NEBULA: Neural Empirical Bayes Under Latent Representations for Efficient and Controllable Design of Molecular Libraries
- Author
-
Nowara, Ewa M., Pinheiro, Pedro O., Mahajan, Sai Pooja, Mahmood, Omar, Watkins, Andrew Martin, Saremi, Saeed, and Maser, Michael
- Subjects
Computer Science - Machine Learning ,Quantitative Biology - Biomolecules - Abstract
We present NEBULA, the first latent 3D generative model for scalable generation of large molecular libraries around a seed compound of interest. Such libraries are crucial for scientific discovery, but it remains challenging to generate large numbers of high quality samples efficiently. 3D-voxel-based methods have recently shown great promise for generating high quality samples de novo from random noise (Pinheiro et al., 2023). However, sampling in 3D-voxel space is computationally expensive and use in library generation is prohibitively slow. Here, we instead perform neural empirical Bayes sampling (Saremi & Hyvarinen, 2019) in the learned latent space of a vector-quantized variational autoencoder. NEBULA generates large molecular libraries nearly an order of magnitude faster than existing methods without sacrificing sample quality. Moreover, NEBULA generalizes better to unseen drug-like molecules, as demonstrated on two public datasets and multiple recently released drugs. We expect the approach herein to be highly enabling for machine learning-based drug discovery. The code is available at https://github.com/prescient-design/nebula
- Published
- 2024
20. Perioperative Changes in Plasma Nitrite and IL-6 Levels Predict Postoperative Atrial Fibrillation (POAF) and Acute Kidney Injury (AKI) after Cardiac Surgery.
- Author
-
Fischer, Matthew, Howard-Quijano, Kimberly, Zong, Nobel, Youn, Ji, Liu, Norika, Scovotti, Jennifer, Grogan, Tristan, Mahajan, Aman, and Cai, Hua
- Subjects
ROS ,acute kidney injury ,atrial fibrillation ,nitric oxide (NO) ,nitrite ,post-operative complications - Abstract
Background: Postoperative atrial fibrillation (POAF) and acute kidney injury (AKI) are common yet significant complications after cardiac surgery, with incidences of up to 40% for each. Here, we assessed plasma nitrite and serum interleukin-6 (IL-6) levels before and after cardiac surgery to quantify the extent to which oxidative stress and inflammation contribute to POAF and AKI occurrence. Methods: We prospectively enrolled 206 cardiac surgical patients. Plasma nitrite and serum IL-6 levels were determined preoperatively and at 24 h, 48 h and 72 h postoperatively. The patients had continuous EKG monitoring for occurrence of POAF, while daily serum creatinine was measured for determination of stage 1 + AKI. Results: Postoperatively, 78 (38%) patients experienced AF, and 47 (23%) patients experienced stage 1 + AKI. POAF analysis: Age, ACE-inhibitor use, valve surgery and percent change in baseline plasma nitrite at 24 h postoperatively were associated with POAF in multiple logistic regression analysis. The inclusion of this new biomarker significantly improved the POAF prediction model (AUC 0.77 for clinical risk factors alone, to AUC 0.81). AKI analysis: A history of diabetes mellitus was associated with AKI in multiple logistic regression analysis, and the addition of preoperative IL-6 levels improved the prediction model for AKI occurrence (AUC 0.69 to AUC 0.74). Conclusions: We previously observed selective upregulation of NADPH oxidase isoform 4 (NOX4) in patients with AF, a critical causal role of NOX4 for AF in zebrafish and a robust inhibitory effect of nitric oxide (NO) on NOX4. Our data innovatively demonstrate that a reduction in circulating nitrite levels, likely implicative of elevated NOX4-mediated oxidative stress, independently associates with POAF and improves POAF prediction, whereas the inclusion of circulating IL-6 levels improves the prediction model for AKI. Therefore, therapeutic strategies to mitigate these pathophysiological sequalae of surgical stress may reduce the incidence of severe postoperative complications of POAF and AKI.
- Published
- 2024
21. First Results of the Magnetometer (MAG) Payload onboard Aditya-L1 Spacecraft
- Author
-
Yadav, Vipin K., Vijaya, Y., Srikar, P. T., Prasad, B. Krishnam, Mahajan, Monika, Mallikarjun, K. V. L. N., Narendra, S., Adoni, Abhijit A., Rai, Vijay S., Veeresha, D. R., and Zamani, Syeeda N.
- Subjects
Physics - Space Physics - Abstract
Aditya-L1 is the first Indian solar mission placed at the first Lagrangian (L1) point to study the Sun. A fluxgate magnetometer (MAG) is one of the seven payloads and one of the three in-situ payloads onboard to measure the interplanetary magnetic field (IMF) coming from the Sun towards the Earth. At present, the Aditya-L1 spacecraft is in a halo-orbit around the L1 point and the MAG payload is ON is continuously measuring the IMF. This paper presents the first measurements of the IMF by MAG.
- Published
- 2024
22. Upshifted frequency of electromagnetic plasma waves due to reflecting gravitational waves acting as almost-luminal mirrors
- Author
-
Asenjo, Felipe A. and Mahajan, Swadesh M.
- Subjects
Astrophysics - High Energy Astrophysical Phenomena - Abstract
We show that dispersive gravitational waves, as a background spacetime, can reflect electromagnetic waves in a plasma. This reflection upshifts the frequency of the reflected wave, being larger for low-frequency incident waves. This effect takes place when the gravitational wave background propagates almost at the speed of light, allowing it to behave similar to a luminal mirror to electromagnetic plasma waves.
- Published
- 2024
23. Normalization of ZZ instanton amplitudes in type 0B minimal superstring theory
- Author
-
Chakrabhavi, Vivek, Eniceicu, Dan Stefan, Mahajan, Raghu, and Murdia, Chitraang
- Subjects
High Energy Physics - Theory - Abstract
We study ZZ instanton corrections in the $(2,4k)$ $N=1$ minimal superstring theory with the type 0B GSO projection, which becomes the type 0B $N=1$ super-JT gravity in the $k \to \infty$ limit. Each member of the $(2,4k)$ family of theories has two phases distinguished by the sign of the Liouville bulk cosmological constant. The worldsheet method for computing the one-loop normalization constant multiplying the instanton corrections gives an ill-defined answer in both phases. We fix these divergences using insights from string field theory and find finite, unambiguous results. Each member of the $(2,4k)$ family of theories is dual to a double-scaled one-matrix integral, where the double-scaling limit can be obtained starting either from a unitary matrix integral with a leading one-cut saddle point, or from a hermitian matrix integral with a leading two-cut saddle point. The matrix integral exhibits a gap-closing transition, which is the same as the double-scaled Gross-Witten-Wadia transition when $k=1$. We also compute instanton corrections in the double-scaled matrix integral for all $k$ and in both phases, and find perfect agreement with the string theory results., Comment: 27 pages of main text + 27 pages of appendices + references
- Published
- 2024
24. M2Lingual: Enhancing Multilingual, Multi-Turn Instruction Alignment in Large Language Models
- Author
-
Maheshwary, Rishabh, Yadav, Vikas, Nguyen, Hoang, Mahajan, Khyati, and Madhusudhan, Sathwik Tejaswi
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning - Abstract
Instruction finetuning (IFT) is critical for aligning Large Language Models (LLMs) to follow instructions. While many effective IFT datasets have been introduced recently, they predominantly focus on high-resource languages like English. To better align LLMs across a broad spectrum of languages and tasks, we propose a fully synthetic, novel taxonomy (Evol) guided Multilingual, Multi-turn instruction finetuning dataset, called M2Lingual. It is constructed by first selecting a diverse set of seed examples and then utilizing the proposed Evol taxonomy to convert these seeds into complex and challenging multi-turn instructions. We demonstrate the effectiveness of M2Lingual by training LLMs of varying sizes and showcasing the enhanced performance across a diverse set of languages. We contribute the 2 step Evol taxonomy with the guided generation code: https://github.com/ServiceNow/M2Lingual, as well as the first fully synthetic, general and task-oriented, multi-turn, multilingual dataset built with Evol - M2Lingual: https://huggingface.co/datasets/ServiceNow-AI/ M2Lingual - containing 182K total IFT pairs, covering 70 languages and 17+ NLP tasks., Comment: 39 pages
- Published
- 2024
25. Improved Modularity and New Features in ipie: Towards Even Larger AFQMC Calculations on CPUs and GPUs at Zero and Finite Temperatures
- Author
-
Jiang, Tong, Baumgarten, Moritz K. A., Loos, Pierre-François, Mahajan, Ankit, Scemama, Anthony, Ung, Shu Fay, Zhang, Jinghong, Malone, Fionn D, and Lee, Joonho
- Subjects
Physics - Chemical Physics - Abstract
ipie is a Python-based auxiliary-field quantum Monte Carlo (AFQMC) package that has undergone substantial improvements since its initial release [J. Chem. Theory Comput., 2022, 19(1): 109-121]. This paper outlines the improved modularity and new capabilities implemented in ipie. We highlight the ease of incorporating different trial and walker types and the seamless integration of ipie with external libraries. We enable distributed Hamiltonian simulations, allowing for multi-GPU simulations of large systems. This development enabled us to compute the interaction energy of a benzene dimer with 84 electrons and 1512 orbitals, which otherwise would not have fit on a single GPU. We also support GPU-accelerated multi-slater determinant trial wavefunctions [arXiv:2406.08314] to enable efficient and highly accurate simulations of large-scale systems. This allows for near-exact ground state energies of multi-reference clusters, [Cu$_2$O$_2$]$^{2+}$ and [Fe$_2$S$_2$(SCH$_3$)]$^{2-}$. We also describe implementations of free projection AFQMC, finite temperature AFQMC, AFQMC for electron-phonon systems, and automatic differentiation in AFQMC for calculating physical properties. These advancements position ipie as a leading platform for AFQMC research in quantum chemistry, facilitating more complex and ambitious computational method development and their applications., Comment: 17 pages, 13 figures
- Published
- 2024
26. The effect of separatrix density and PFC material on H-mode confinement in the ITPA global H-mode database
- Author
-
Kotschenreuther, M., Liu, X., Hatch, D. R., and Mahajan, S. M.
- Subjects
Physics - Plasma Physics - Abstract
Recent data, added to the ITPA global H-mode database [1] for ASDEX-U [2] and JET-ILW, reveals that the separatrix density $n_{sep}$ has a correlation with the H20 factor (Confinement time relative to the ITPA20-IL scaling)[1]. These trends are analyzed in detail. They are not a result of proximity to the density limit. The normalized $n_{sepN} = n_{sep} / \bar{n}$ is introduced, motivated by theory ($\bar{n}$ is the average density). The trends in $n_{sepN}$ can be understood in terms of the two main mechanisms of pedestal characteristics -- MHD stability and recently developed theories of gyrokinetic transport. Careful analysis shows these mechanisms can be distinguished in the data. The most dramatic improvement in confinement time arises primarily from reductions in pedestal transport. A new definition of density peaking that includes core peaking is found to best explain H20 when advanced H-modes are included: $n_{sepN0} = n_{sep}/n(0)$, the inverse of the total density peaking from the separatrix to the axis. The highest H-factors are reached by the confluence of relatively low normalized $n_{sepN0}$ plus high Shafranov shift or poloidal beta. The importance of these two variables is also theoretically predicted from recent analysis of the gyrokinetic system, where a constraint can limit the access of ITG/TEM modes to free energy in equilibrium gradients. The Plasma Facing Component (PFC) material also shows a strong influence in the data. This is likely due to the importance of $n(0)/n_{sep}$ to attaining high H20, in conjunction with the known tendency for tungsten (W) to accumulate with density peaking and low transport. Preliminary results indicate that $n_{sepN0}$ might also be important with core ITBs.
- Published
- 2024
27. An uniform lower bound for classical Kloosterman sums and an application
- Author
-
Mahajan, Jewel, Das, Jishu, and Baier, Stephan
- Subjects
Mathematics - Number Theory ,Primary 11L05, 11L07, 11F72 - Abstract
We present an elementary uniform lower bound for the classical Kloosterman sum $S(a,b;c)$ under the condition of its non-vanishing and $(ab,c)=1$, with $c$ being an odd integer. We then apply this lower bound for Kloosterman sums to derive an explicit lower bound in the Petersson's trace formula, subject to a pertinent condition. Consequently, we achieve a modified version of a theorem by Jung and Sardari, wherein the parameters $k$ and $N$ are permitted to vary independently., Comment: 11 pages
- Published
- 2024
28. Benchmarking the Exponential Ansatz for the Holstein model
- Author
-
Yang, Junjie, Cui, Zhi-Hao, Mahajan, Ankit, Zhai, Huanchen, Reichman, David R., and Chan, Garnet Kin-Lic
- Subjects
Condensed Matter - Materials Science ,Physics - Chemical Physics - Abstract
Polarons are quasiparticles formed as a result of lattice distortions induced by charge carriers. The single-electron Holstein model captures the fundamentals of single polaron physics. We examine the power of the exponential ansatz for the polaron ground-state wavefunction in its coupled cluster, canonical transformation, and (canonically transformed) perturbative variants across the parameter space of the Holstein model. Our benchmark serves to guide future developments of polaron wavefunctions beyond the single-electron Holstein model., Comment: 10 pages, 6 figures
- Published
- 2024
29. Equity Implications of Net-Zero Emissions: A Multi-Model Analysis of Energy Expenditures Across Income Classes Under Economy-Wide Deep Decarbonization Policies
- Author
-
Bistlinea, John, Onda, Chikara, Browning, Morgan, Emmerling, Johannes, Iyer, Gokul, Mahajan, Megan, McFarland, Jim, McJeon, Haewon, Orvis, Robbie, Fonseca, Francisco Ralston, Roney, Christopher, Sandoval, Noah, Sarmiento, Luis, Weyant, John, Woollacott, Jared, and Yuan, Mei
- Subjects
Physics - Physics and Society ,Economics - General Economics - Abstract
With companies, states, and countries targeting net-zero emissions around midcentury, there are questions about how these targets alter household welfare and finances, including distributional effects across income groups. This paper examines the distributional dimensions of technology transitions and net-zero policies with a focus on welfare impacts across household incomes. The analysis uses a model intercomparison with a range of energy-economy models using harmonized policy scenarios reaching economy-wide, net-zero CO2 emissions across the United States in 2050. We employ a novel linking approach that connects output from detailed energy system models with survey microdata on energy expenditures across income classes to provide distributional analysis of net-zero policies. Although there are differences in model structure and input assumptions, we find broad agreement in qualitative trends in policy incidence and energy burdens across income groups. Models generally agree that direct energy expenditures for many households will likely decline over time with reference and net-zero policies. However, there is variation in the extent of changes relative to current levels, energy burdens relative to reference levels, and electricity expenditures. Policy design, primarily how climate policy revenues are used, has first-order impacts on distributional outcomes. Net-zero policy costs, in both absolute and relative terms, are unevenly distributed across households, and relative increases in energy expenditures are higher for lowest-income households. However, we also find that recycled revenues from climate policies have countervailing effects when rebated on a per-capita basis, offsetting higher energy burdens and potentially even leading to net progressive outcomes.
- Published
- 2024
- Full Text
- View/download PDF
30. DePIN: A Framework for Token-Incentivized Participatory Sensing
- Author
-
Chiu, Michael T. C., Mahajan, Sachit, Ballandies, Mark C., and Kalabić, Uroš V.
- Subjects
Computer Science - Computer Science and Game Theory - Abstract
There is always demand for integrating data into microeconomic decision making. Participatory sensing deals with how real-world data may be extracted with stakeholder participation and resolves a problem of Big Data, which is concerned with monetizing data extracted from individuals without their participation. We present how Decentralized Physical Infrastructure Networks (DePINs) extend participatory sensing. We discuss the threat models of these networks and how DePIN cryptoeconomics can advance participatory sensing.
- Published
- 2024
31. Improved classical shadows from local symmetries in the Schur basis
- Author
-
Grier, Daniel, Liu, Sihan, and Mahajan, Gaurav
- Subjects
Quantum Physics ,Computer Science - Data Structures and Algorithms ,Computer Science - Information Theory ,Computer Science - Machine Learning - Abstract
We study the sample complexity of the classical shadows task: what is the fewest number of copies of an unknown state you need to measure to predict expected values with respect to some class of observables? Large joint measurements are likely required in order to minimize sample complexity, but previous joint measurement protocols only work when the unknown state is pure. We present the first joint measurement protocol for classical shadows whose sample complexity scales with the rank of the unknown state. In particular we prove $\mathcal O(\sqrt{rB}/\epsilon^2)$ samples suffice, where $r$ is the rank of the state, $B$ is a bound on the squared Frobenius norm of the observables, and $\epsilon$ is the target accuracy. In the low-rank regime, this is a nearly quadratic advantage over traditional approaches that use single-copy measurements. We present several intermediate results that may be of independent interest: a solution to a new formulation of classical shadows that captures functions of non-identical input states; a generalization of a ``nice'' Schur basis used for optimal qubit purification and quantum majority vote; and a measurement strategy that allows us to use local symmetries in the Schur basis to avoid intractable Weingarten calculations in the analysis.
- Published
- 2024
32. Ab Initio Polaron Wave Functions
- Author
-
Robinson, Paul J., Lee, Joonho, Mahajan, Ankit, and Reichman, David R.
- Subjects
Condensed Matter - Strongly Correlated Electrons ,Condensed Matter - Materials Science - Abstract
In this work we demonstrate that accurate ground state wave functions may be constructed for polarons in a fully ab initio setting across the wide range of couplings associated with both the large and small polaron limits. We present a single general unitary transformation approach which encompasses an ab initio version of the Lee-Low-Pines theory at weak coupling and the coherent state Landau-Pekar framework at strong coupling while interpolating between these limits in general cases. We show that perturbation theory around these limits may be performed in a facile manner to assess the accuracy of the approach, as well as provide an independent route to the ab initio properties of polarons. We test these ideas on the case of LiF, where the electron-polaron is expected to be large and relatively weakly coupled, while the hole-polaron is expected to be a strongly coupled small polaron., Comment: 12 pages, 3 figures
- Published
- 2024
33. Structure and dynamics of electron-phonon coupled systems using neural quantum states
- Author
-
Mahajan, Ankit, Robinson, Paul J., Lee, Joonho, and Reichman, David R.
- Subjects
Condensed Matter - Strongly Correlated Electrons ,Condensed Matter - Materials Science - Abstract
In this work, we use neural quantum states (NQS) to describe the high-dimensional wave functions of electron-phonon coupled systems. We demonstrate that NQS can accurately and systematically learn the underlying physics of such problems through a variational Monte Carlo optimization of the energy with minimal incorporation of physical information even in highly challenging cases. We assess the ability of our approach across various lattice model examples featuring different types of couplings. The flexibility of our NQS formulation is demonstrated via application to ab initio models parametrized by density functional perturbation theory consisting of electron or hole bands coupled linearly to dispersive phonons. We compute accurate real-frequency spectral properties of electron-phonon systems via a novel formalism based on NQS. Our work establishes a general framework for exploring diverse ground state and dynamical phenomena arising in electron-phonon systems, including the non-perturbative interplay of correlated electronic and electron-phonon effects in systems ranging from simple lattice models to realistic models of materials parametrized by ab initio calculations., Comment: 17 pages, 13 figures
- Published
- 2024
34. (H,Li)$_{6}$Ru$_{2}$O$_{6}$ : a zero-field Ru$^{3+}$-based Kitaev Quantum Spin Liquid
- Author
-
Bachhar, Sanjay, Baenitz, M., Luetkens, Hubertus, Pujari, Sumiran, and Mahajan, A. V.
- Subjects
Condensed Matter - Strongly Correlated Electrons - Abstract
We report the synthesis and properties of (H,Li)$_{6}$Ru$_{2}$O$_{6}$, which is shown to be a $J_{\text{eff}}=\frac{1}{2}$ system made out of Ru$^{3+}$ moments in a honeycomb geometry. Bulk magnetization, heat capacity, nuclear magnetic resonance (NMR), and muon spin relaxation ($\mu$SR) rule out the presence of static moments or any spin glass phase down to 500 mK. All techniques suggest a crossover to a liquid-like state below about 40 K. The $^{7}$Li nuclear magnetic resonance (NMR) shift data suggest a non-zero $T$-independent spin susceptibility at low $T$. In zero field, $C_m/T$ shows $T^{-1}$ divergence which is consistent with vacancy-induced effects on low-energy excitations of the pristine Kitaev spin liquid. With field, power-law variations in the $^{7}$Li NMR spin-lattice relaxation rate 1/T$_{1}$ and magnetic heat capacity $C_{m}$ show quantitatively new scaling behaviors. A two-step entropy release in heat capacity is also observed putatively from $Z_{2}$ flux (low-$T$ step) and itinerant Majorana fermions (high-$T$ step). Based on these findings, we propose that (H,Li)$_{6}$Ru$_{2}$O$_{6}$ realizes a Kitaev spin liquid with no evidence of inherent magnetic ordering in zero field unlike $\alpha$-RuCl$_{3}$ where approximately $8$ Tesla field is required to suppress magnetic order., Comment: 5 pages, 6 figures of main text + 12 pages, 25 figures of supplementary information
- Published
- 2024
35. On the Coverage Required for Diploid Genome Assembly
- Author
-
Mahajan, Daanish, Jain, Chirag, and Kashyap, Navin
- Subjects
Computer Science - Information Theory ,Quantitative Biology - Genomics - Abstract
We investigate the information-theoretic conditions to achieve the complete reconstruction of a diploid genome. We also analyze the standard greedy and de-Bruijn graph-based algorithms and compare the coverage depth and read length requirements with the information-theoretic lower bound. Our results show that the gap between the two is considerable because both algorithms require the double repeats in the genome to be bridged., Comment: Accepted at ISIT'24
- Published
- 2024
36. Unbiasing Fermionic Auxiliary-Field Quantum Monte Carlo with Matrix Product State Trial Wavefunctions
- Author
-
Jiang, Tong, O'Gorman, Bryan, Mahajan, Ankit, and Lee, Joonho
- Subjects
Physics - Chemical Physics ,Quantum Physics - Abstract
In this work, we report, for the first time, an implementation of fermionic auxiliary-field quantum Monte Carlo (AFQMC) using matrix product state (MPS) trial wavefunctions, dubbed MPS-AFQMC. Calculating overlaps between an MPS trial and arbitrary Slater determinants up to a multiplicative error, a crucial subroutine in MPS-AFQMC, is proven to be #P-hard. Nonetheless, we tested several promising heuristics in successfully improving fermionic phaseless AFQMC energies. We also proposed a way to evaluate local energy and force bias evaluations free of matrix-product operators. This allows for larger basis set calculations without significant overhead. We showcase the utility of our approach on one- and two-dimensional hydrogen lattices, even when the MPS trial itself struggles to obtain high accuracy. Our work offers a new set of tools that can solve currently challenging electronic structure problems with future improvements., Comment: 24 pages, 20 figures
- Published
- 2024
37. Predicting Lung Disease Severity via Image-Based AQI Analysis using Deep Learning Techniques
- Author
-
Mahajan, Anvita, Mate, Sayali, Kulkarni, Chinmayee, and Sawant, Suraj
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Machine Learning - Abstract
Air pollution is a significant health concern worldwide, contributing to various respiratory diseases. Advances in air quality mapping, driven by the emergence of smart cities and the proliferation of Internet-of-Things sensor devices, have led to an increase in available data, fueling momentum in air pollution forecasting. The objective of this study is to devise an integrated approach for predicting air quality using image data and subsequently assessing lung disease severity based on Air Quality Index (AQI).The aim is to implement an integrated approach by refining existing techniques to improve accuracy in predicting AQI and lung disease severity. The study aims to forecast additional atmospheric pollutants like AQI, PM10, O3, CO, SO2, NO2 in addition to PM2.5 levels. Additionally, the study aims to compare the proposed approach with existing methods to show its effectiveness. The approach used in this paper uses VGG16 model for feature extraction in images and neural network for predicting AQI.In predicting lung disease severity, Support Vector Classifier (SVC) and K-Nearest Neighbors (KNN) algorithms are utilized. The neural network model for predicting AQI achieved training accuracy of 88.54 % and testing accuracy of 87.44%,which was measured using loss function, while the KNN model used for predicting lung disease severity achieved training accuracy of 98.4% and testing accuracy of 97.5% In conclusion, the integrated approach presented in this study forecasts air quality and evaluates lung disease severity, achieving high testing accuracies of 87.44% for AQI and 97.5% for lung disease severity using neural network, KNN, and SVC models. The future scope involves implementing transfer learning and advanced deep learning modules to enhance prediction capabilities. While the current study focuses on India, the objective is to expand its scope to encompass global coverage., Comment: 11 pages
- Published
- 2024
38. Scalable Ab Initio Electronic Structure Methods with Near Chemical Accuracy for Main Group Chemistry
- Author
-
Wei, Yujing, Debnath, Sibali, Weber, John L., Mahajan, Ankit, Reichman, David R., and Friesner, Richard A.
- Subjects
Physics - Chemical Physics - Abstract
This study evaluates the precision of widely recognized quantum chemical methodologies, CCSD(T), DLPNO-CCSD(T) and localized ph-AFQMC, for determining the thermochemistry of main group elements. DLPNO-CCSD(T) and localized ph-AFQMC, which offer greater scalability compared to canonical CCSD(T), have emerged over the last decade as pivotal in producing precise benchmark chemical data. Our investigation includes closed-shell, neutral molecules, focusing on their heat of formation and atomization energy sourced from four specific small molecule datasets. Firstly, we selected molecules from the G2 and G3 datasets, noted for their reliable experimental heat of formation data. Additionally, we incorporate molecules from the W4-11 and W4-17 sets, which provide high-level theoretical reference values for atomization energy at 0 K. Our findings reveal that both DLPNO-CCSD(T) and ph-AFQMC methods are capable of achieving a root-mean-square deviation (RMSD) of less than 1 kcal/mol across the combined dataset, aligning with the threshold for chemical accuracy. Moreover, we make efforts to confine the maximum deviations within 2 kcal/mol, a degree of precision that significantly broadens the applicability of these methods in fields such as biology and materials science.
- Published
- 2024
- Full Text
- View/download PDF
39. From stars to diverse mantles, melts, crusts and atmospheres of rocky exoplanets
- Author
-
Guimond, Claire Marie, Wang, Haiyang, Seidler, Fabian, Sossi, Paolo, Mahajan, Aprajit, and Shorttle, Oliver
- Subjects
Astrophysics - Earth and Planetary Astrophysics ,Astrophysics - Solar and Stellar Astrophysics ,Physics - Geophysics - Abstract
This review is focused on describing the logic by which we make predictions of exoplanetary compositions and mineralogies, and how these processes could lead to compositional diversity among rocky exoplanets. We use these predictions to determine the sensitivity of present-day and future observations to detecting compositional differences between rocky exoplanets and the four terrestrial planets. First, we review data on stellar abundances and infer how changes in composition may manifest themselves in the expected bulk compositions of rocky exoplanets (section 2). Converting this information in mass-radius relationships requires calculation of the stable mineral assemblages at a given temperature-pressure-composition (T-P-X), an exercise we describe in section 3. Should the planet be hot enough to engender partial melting of the mantle, then these liquids are likely to rise to the surface and erupt to form planetary crusts; the possible compositional and mineralogical variability of which we examine in section 4. Finally, the expected spectroscopic responses of such crusts are examined in section 5., Comment: Chapter 8 accepted for publication in the Reviews in Mineralogy and Geochemistry (RiMG) Volume 90 on "Exoplanets: Compositions, Mineralogy, and Evolution" edited by Natalie Hinkel, Keith Putirka, and Siyi Xu; 40 pages, 11 figures, and 3 equations
- Published
- 2024
40. Workload-Aware Hardware Accelerator Mining for Distributed Deep Learning Training
- Author
-
Adnan, Muhammad, Phanishayee, Amar, Kulkarni, Janardhan, Nair, Prashant J., and Mahajan, Divya
- Subjects
Computer Science - Hardware Architecture ,Computer Science - Distributed, Parallel, and Cluster Computing - Abstract
In this paper, we present a novel technique to search for hardware architectures of accelerators optimized for end-to-end training of deep neural networks (DNNs). Our approach addresses both single-device and distributed pipeline and tensor model parallel scenarios, latter being addressed for the first time. The search optimized accelerators for training relevant metrics such as throughput/TDP under a fixed area and power constraints. However, with the proliferation of specialized architectures and complex distributed training mechanisms, the design space exploration of hardware accelerators is very large. Prior work in this space has tried to tackle this by reducing the search space to either a single accelerator execution that too only for inference, or tuning the architecture for specific layers (e.g., convolution). Instead, we take a unique heuristic-based critical path-based approach to determine the best use of available resources (power and area) either for a set of DNN workloads or each workload individually. First, we perform local search to determine the architecture for each pipeline and tensor model stage. Specifically, the system iteratively generates architectural configurations and tunes the design using a novel heuristic-based approach that prioritizes accelerator resources and scheduling to critical operators in a machine learning workload. Second, to address the complexities of distributed training, the local search selects multiple (k) designs per stage. A global search then identifies an accelerator from the top-k sets to optimize training throughput across the stages. We evaluate this work on 11 different DNN models. Compared to a recent inference-only work Spotlight, our method converges to a design in, on average, 31x less time and offers 12x higher throughput. Moreover, designs generated using our method achieve 12% throughput improvement over TPU architecture.
- Published
- 2024
41. Evaluating Interventional Reasoning Capabilities of Large Language Models
- Author
-
Kasetty, Tejas, Mahajan, Divyat, Dziugaite, Gintare Karolina, Drouin, Alexandre, and Sridhar, Dhanya
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Computation and Language ,Statistics - Methodology - Abstract
Numerous decision-making tasks require estimating causal effects under interventions on different parts of a system. As practitioners consider using large language models (LLMs) to automate decisions, studying their causal reasoning capabilities becomes crucial. A recent line of work evaluates LLMs ability to retrieve commonsense causal facts, but these evaluations do not sufficiently assess how LLMs reason about interventions. Motivated by the role that interventions play in causal inference, in this paper, we conduct empirical analyses to evaluate whether LLMs can accurately update their knowledge of a data-generating process in response to an intervention. We create benchmarks that span diverse causal graphs (e.g., confounding, mediation) and variable types, and enable a study of intervention-based reasoning. These benchmarks allow us to isolate the ability of LLMs to accurately predict changes resulting from their ability to memorize facts or find other shortcuts. Our analysis on four LLMs highlights that while GPT- 4 models show promising accuracy at predicting the intervention effects, they remain sensitive to distracting factors in the prompts., Comment: 17 pages
- Published
- 2024
42. Unlocking Adaptive User Experience with Generative AI
- Author
-
Huang, Yutan, Kanij, Tanjila, Madugalla, Anuradha, Mahajan, Shruti, Arora, Chetan, and Grundy, John
- Subjects
Computer Science - Human-Computer Interaction ,Computer Science - Software Engineering - Abstract
Developing user-centred applications that address diverse user needs requires rigorous user research. This is time, effort and cost-consuming. With the recent rise of generative AI techniques based on Large Language Models (LLMs), there is a possibility that these powerful tools can be used to develop adaptive interfaces. This paper presents a novel approach to develop user personas and adaptive interface candidates for a specific domain using ChatGPT. We develop user personas and adaptive interfaces using both ChatGPT and a traditional manual process and compare these outcomes. To obtain data for the personas we collected data from 37 survey participants and 4 interviews in collaboration with a not-for-profit organisation. The comparison of ChatGPT generated content and manual content indicates promising results that encourage using LLMs in the adaptive interfaces design process.
- Published
- 2024
43. Age-stratified Assessment of Brain Volumetric Segmentation on the Indian Population Using Quantitative Magnetic Resonance Imaging
- Author
-
Syed Nasser, Nisha, Venugopal, Vasantha K., Veenstra, Cynthia, Johansson, Peter, Rajan, Sriram, Mahajan, Kabir, Naik, Swati, Masand, Ravi, Yadav, Pratiksha, Khanduri, Sachin, Singhal, Suman, Bhargava, Rajat, Kabra, Utkarsh, Gupta, Sanjay, Saggar, Kavita, Varaprasad, Balaji, Aggrawal, Kushagra, Rao, Adinarayana, K.S., Manoj, Dakhole, Atul, Kelkar, Abhimanyu, Benjamin, Geena, Sodani, Varsha, Goyal, Pradeep, and Mahajan, Harsh
- Published
- 2024
- Full Text
- View/download PDF
44. Test and treat model for tuberculosis preventive treatment among household contacts of pulmonary tuberculosis patients in selected districts of maharashtra: A mixed-methods study on care cascade, timeliness, and early implementation challenges
- Author
-
Mahajan, Palak, Soundappan, Kathirvel, Singla, Neeta, Mehta, Kedar, Nuken, Amenla, Thekkur, Pruthu, Nair, Divya, Rattan, Sampan, Thakur, Chaturanand, Sachdeva, Kuldeep Singh, and Kalottee, Bharati
- Published
- 2024
45. Transition to adulthood for youth with neurodevelopment disorders- family/caregiver survey and needs assessment
- Author
-
Talmi, Sydney, Mahajan, Gayatri, and Wilson, Machelle
- Abstract
Transition to adulthood is a challenging process that is especially difficult for youth with neurodevelopmental disorders (NDD) and their caregivers. Due to transition age youth with NDD having complex medical and behavioral health needs, youth and their caregivershave to navigate several systems of care to support them. Despite these challenges, there has been limited research into the needs of this population during transition to adulthood from a quality improvement perspective.
- Published
- 2024
46. A Study on the Use of Simulation in Synthesizing Path-Following Control Policies for Autonomous Ground Robots
- Author
-
Zhang, Harry, Caldararu, Stefan, Young, Aaron, Ruiz, Alexis, Unjhawala, Huzaifa, Mahajan, Ishaan, Ashokkumar, Sriram, Batagoda, Nevindu, Zhou, Zhenhao, Bakke, Luning, and Negrut, Dan
- Subjects
Computer Science - Robotics - Abstract
We report results obtained and insights gained while answering the following question: how effective is it to use a simulator to establish path following control policies for an autonomous ground robot? While the quality of the simulator conditions the answer to this question, we found that for the simulation platform used herein, producing four control policies for path planning was straightforward once a digital twin of the controlled robot was available. The control policies established in simulation and subsequently demonstrated in the real world are PID control, MPC, and two neural network (NN) based controllers. Training the two NN controllers via imitation learning was accomplished expeditiously using seven simple maneuvers: follow three circles clockwise, follow the same circles counter-clockwise, and drive straight. A test randomization process that employs random micro-simulations is used to rank the ``goodness'' of the four control policies. The policy ranking noted in simulation correlates well with the ranking observed when the control policies were tested in the real world. The simulation platform used is publicly available and BSD3-released as open source; a public Docker image is available for reproducibility studies. It contains a dynamics engine, a sensor simulator, a ROS2 bridge, and a ROS2 autonomy stack the latter employed both in the simulator and the real world experiments., Comment: 8 pages, 7 figures
- Published
- 2024
47. Relational Network Verification
- Author
-
Xu, Xieyang, Yuan, Yifei, Kincaid, Zachary, Krishnamurthy, Arvind, Mahajan, Ratul, Walker, David, and Zhai, Ennan
- Subjects
Computer Science - Networking and Internet Architecture - Abstract
Relational network verification is a new approach to validating network changes. In contrast to traditional network verification, which analyzes specifications for a single network snapshot, relational network verification analyzes specifications concerning two network snapshots (e.g., pre- and post-change snapshots) and captures their similarities and differences. Relational change specifications are compact and precise because they specify the flows or paths that change between snapshots and then simply mandate that other behaviors of the network "stay the same", without enumerating them. To achieve similar guarantees, single-snapshot specifications need to enumerate all flow and path behaviors that are not expected to change, so we can check that nothing has accidentally changed. Thus, precise single-snapshot specifications are proportional to network size, which makes them impractical to generate for many real-world networks. To demonstrate the value of relational reasoning, we develop a high-level relational specification language and a tool called Rela to validate network changes. Rela first compiles input specifications and network snapshot representations to finite state transducers. It then checks compliance using decision procedures for automaton equivalence. Our experiments using data on complex changes to a global backbone (with over 10^3 routers) find that Rela specifications need fewer than 10 terms for 93% of them and it validates 80% of them within 20 minutes.
- Published
- 2024
48. Accelerating Recommender Model Training by Dynamically Skipping Stale Embeddings
- Author
-
Maboud, Yassaman Ebrahimzadeh, Adnan, Muhammad, Mahajan, Divya, and Nair, Prashant J.
- Subjects
Computer Science - Information Retrieval ,Computer Science - Machine Learning - Abstract
Training recommendation models pose significant challenges regarding resource utilization and performance. Prior research has proposed an approach that categorizes embeddings into popular and non-popular classes to reduce the training time for recommendation models. We observe that, even among the popular embeddings, certain embeddings undergo rapid training and exhibit minimal subsequent variation, resulting in saturation. Consequently, updates to these embeddings lack any contribution to model quality. This paper presents Slipstream, a software framework that identifies stale embeddings on the fly and skips their updates to enhance performance. This capability enables Slipstream to achieve substantial speedup, optimize CPU-GPU bandwidth usage, and eliminate unnecessary memory access. SlipStream showcases training time reductions of 2x, 2.4x, 1.2x, and 1.175x across real-world datasets and configurations, compared to Baseline XDL, Intel-optimized DRLM, FAE, and Hotline, respectively.
- Published
- 2024
49. Accelerating String-Key Learned Index Structures via Memoization-based Incremental Training
- Author
-
Kim, Minsu, Hwang, Jinwoo, Heo, Guseul, Cho, Seiyeon, Mahajan, Divya, and Park, Jongse
- Subjects
Computer Science - Machine Learning ,Computer Science - Hardware Architecture ,Computer Science - Databases - Abstract
Learned indexes use machine learning models to learn the mappings between keys and their corresponding positions in key-value indexes. These indexes use the mapping information as training data. Learned indexes require frequent retrainings of their models to incorporate the changes introduced by update queries. To efficiently retrain the models, existing learned index systems often harness a linear algebraic QR factorization technique that performs matrix decomposition. This factorization approach processes all key-position pairs during each retraining, resulting in compute operations that grow linearly with the total number of keys and their lengths. Consequently, the retrainings create a severe performance bottleneck, especially for variable-length string keys, while the retrainings are crucial for maintaining high prediction accuracy and in turn, ensuring low query service latency. To address this performance problem, we develop an algorithm-hardware co-designed string-key learned index system, dubbed SIA. In designing SIA, we leverage a unique algorithmic property of the matrix decomposition-based training method. Exploiting the property, we develop a memoization-based incremental training scheme, which only requires computation over updated keys, while decomposition results of non-updated keys from previous computations can be reused. We further enhance SIA to offload a portion of this training process to an FPGA accelerator to not only relieve CPU resources for serving index queries (i.e., inference), but also accelerate the training itself. Our evaluation shows that compared to ALEX, LIPP, and SIndex, a state-of-the-art learned index systems, SIA-accelerated learned indexes offer 2.6x and 3.4x higher throughput on the two real-world benchmark suites, YCSB and Twitter cache trace, respectively., Comment: Accepted at VLDB '24; 12 pages + 2 pages (ref), 18 figures, 2 tables
- Published
- 2024
50. Quantifying the Sim2real Gap for GPS and IMU Sensors
- Author
-
Mahajan, Ishaan, Unjhawala, Huzaifa, Zhang, Harry, Zhou, Zhenhao, Young, Aaron, Ruiz, Alexis, Caldararu, Stefan, Batagoda, Nevindu, Ashokkumar, Sriram, and Negrut, Dan
- Subjects
Computer Science - Robotics - Abstract
Simulation can and should play a critical role in the development and testing of algorithms for autonomous agents. What might reduce its impact is the ``sim2real'' gap -- the algorithm response differs between operation in simulated versus real-world environments. This paper introduces an approach to evaluate this gap, focusing on the accuracy of sensor simulation -- specifically IMU and GPS -- in velocity estimation tasks for autonomous agents. Using a scaled autonomous vehicle, we conduct 40 real-world experiments across diverse environments then replicate the experiments in simulation with five distinct sensor noise models. We note that direct comparison of raw simulation and real sensor data fails to quantify the sim2real gap for robotics applications. We demonstrate that by using a state of the art state-estimation package as a ``judge'', and by evaluating the performance of this state-estimator in both real and simulated scenarios, we can isolate the sim2real discrepancies stemming from sensor simulations alone. The dataset generated is open-source and publicly available for unfettered use.
- Published
- 2024
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.