1,436,496 results on '"Tan AT"'
Search Results
2. All-optical autoencoder machine learning framework using diffractive processors
- Author
-
Feng, Peijie, Tan, Yong, Chong, Mingzhe, Li, Lintao, Zhang, Zongkun, Liu, Fubei, Tan, Yunhua, and Wen, Yongzheng
- Subjects
Physics - Applied Physics ,Physics - Optics - Abstract
Diffractive deep neural network (D2NN), known for its high speed, low power consumption, and strong parallelism, has been widely applied across various fields, including pattern recognition, image processing, and image transmission. However, existing network architectures primarily focus on data representation within the original domain, with limited exploration of the latent space, thereby restricting the information mining capabilities and multifunctional integration of D2NNs. Here, we propose an all-optical autoencoder (OAE) framework that can encode the input wavefield into a prior shape distribution in the latent space and decode the encoded pattern back to the original wavefield. By leveraging the non-reciprocal property of D2NN, the OAE models function as encoders in one direction of wave propagation and as decoders in the opposite direction. We further apply the models to three key areas: image denoising, noise-resistant reconfigurable image classification, and image generation. Proof-of-concept experiments have been conducted to validate numerical simulations. Our OAE framework fully exploits the potential of latent space representations, enabling a single set of diffractive processors to simultaneously achieve image reconstruction, representation, and generation. It can be viewed as both a counterpart and an extension of the electronic autoencoder model. This work not only offers fresh insights into the design of optical generative models but also paves the way for developing and applying multifunctional, highly integrated, and general optical intelligent systems., Comment: 21 pages, 7 figure
- Published
- 2024
3. Online 4D Ultrasound-Guided Robotic Tracking Enables 3D Ultrasound Localisation Microscopy with Large Tissue Displacements
- Author
-
Yan, Jipeng, Kawara, Shusei, Tan, Qingyuan, Zhu, Jingwen, Wang, Bingxue, Toulemonde, Matthieu, Liu, Honghai, Tan, Ying, and Tang, Meng-Xing
- Subjects
Electrical Engineering and Systems Science - Image and Video Processing ,Electrical Engineering and Systems Science - Systems and Control - Abstract
Super-Resolution Ultrasound (SRUS) imaging through localising and tracking microbubbles, also known as Ultrasound Localisation Microscopy (ULM), has demonstrated significant potential for reconstructing microvasculature and flows with sub-diffraction resolution in clinical diagnostics. However, imaging organs with large tissue movements, such as those caused by respiration, presents substantial challenges. Existing methods often require breath holding to maintain accumulation accuracy, which limits data acquisition time and ULM image saturation. To improve image quality in the presence of large tissue movements, this study introduces an approach integrating high-frame-rate ultrasound with online precise robotic probe control. Tested on a microvasculature phantom with translation motions up to 20 mm, twice the aperture size of the matrix array used, our method achieved real-time tracking of the moving phantom and imaging volume rate at 85 Hz, keeping majority of the target volume in the imaging field of view. ULM images of the moving cross channels in the phantom were successfully reconstructed in post-processing, demonstrating the feasibility of super-resolution imaging under large tissue motions. This represents a significant step towards ULM imaging of organs with large motion.
- Published
- 2024
4. Ordinal Learning: Longitudinal Attention Alignment Model for Predicting Time to Future Breast Cancer Events from Mammograms
- Author
-
Wang, Xin, Tan, Tao, Gao, Yuan, Marcus, Eric, Han, Luyi, Portaluri, Antonio, Zhang, Tianyu, Lu, Chunyao, Liang, Xinglong, Beets-Tan, Regina, Teuwen, Jonas, and Mann, Ritse
- Subjects
Electrical Engineering and Systems Science - Image and Video Processing ,Computer Science - Computer Vision and Pattern Recognition - Abstract
Precision breast cancer (BC) risk assessment is crucial for developing individualized screening and prevention. Despite the promising potential of recent mammogram (MG) based deep learning models in predicting BC risk, they mostly overlook the 'time-to-future-event' ordering among patients and exhibit limited explorations into how they track history changes in breast tissue, thereby limiting their clinical application. In this work, we propose a novel method, named OA-BreaCR, to precisely model the ordinal relationship of the time to and between BC events while incorporating longitudinal breast tissue changes in a more explainable manner. We validate our method on public EMBED and inhouse datasets, comparing with existing BC risk prediction and time prediction methods. Our ordinal learning method OA-BreaCR outperforms existing methods in both BC risk and time-to-future-event prediction tasks. Additionally, ordinal heatmap visualizations show the model's attention over time. Our findings underscore the importance of interpretable and precise risk assessment for enhancing BC screening and prevention efforts. The code will be accessible to the public.
- Published
- 2024
5. Comparative Study of Data-driven Area Inertia Estimation Approaches on WECC Power Systems
- Author
-
Tan, Bendong, Peng, Jiangkai, Gao, Ningchao, Zhao, Junbo, and Tan, Jin
- Subjects
Electrical Engineering and Systems Science - Systems and Control - Abstract
With the increasing integration of inverter-based resources into the power grid, there has been a notable reduction in system inertia, potentially compromising frequency stability. To assess the suitability of existing area inertia estimation techniques for real-world power systems, this paper presents a rigorous comparative analysis of system identification, measurement reconstruction, and electromechanical oscillation-based area inertia estimation methodologies, specifically applied to the large-scale and multi-area WECC 240-bus power system. Comprehensive results show that the system identification-based approach exhibits superior robustness and accuracy relative to its counterparts.
- Published
- 2024
6. Molecular Precision Engineering for Efficient Binary Organic Photovoltaics through Energy Level and Fibrillar Structure Modulation
- Author
-
Zeng, Rui, Xu, Shengjie, Deng, Jiawei, Tan, Senke, Zhou, Guanqing, Zhang, Ming, Zhu, Lei, Han, Fei, Xue, Xiaonan, Zhang, Anyang, Tan, Hongtao, Zhang, Lingjie, Zhu, Chenhui, Wang, Cheng, Wu, Xuefei, Fink, Zachary, Russell, Thomas P, Zhang, Yongming, and Liu, Feng
- Subjects
Engineering ,Macromolecular and Materials Chemistry ,Materials Engineering ,Chemical Sciences ,Affordable and Clean Energy ,Interdisciplinary Engineering ,Macromolecular and materials chemistry ,Materials engineering - Abstract
Adjusting the energy levels and fibrillar morphology is paramount to enhancing the power conversion efficiency (PCE) of organic solar cells (OSCs). In the present study, an increase in the open-circuit voltage (VOC) is facilitated through the elongation of the alkyl chain within AQx (namely AQx-8), aiming to decrease the free volume ratio (FVR). This reduction in FVR attenuates electron-phonon coupling, thereby augmenting emission efficiency and diminishing the non-radiative energy loss (ΔEnr). To further refine the energy levels and morphological characteristics, the external undecyl chain of AQx-8 is substituted with a shorter carbon chain and cyclohexane noted for its considerable steric hindrance (AQx-H). This alteration significantly mitigates intermolecular aggregation, expands the bandgap, and elevates the lowest unoccupied molecular orbital (LUMO) energy level, culminating in an elevated VOC of 0.923 V in devices based on AQx-H. Morphological analysis reveals that blends based on AQx-H exhibit an enhanced multi-length-scale fibrillar structure, which is conducive to exciton dissociation and charge transport, thereby contributing to a high fill factor (FF) nearing 80%. Consequently, this study reports one of the highest binary PCEs documented, standing at 19.5% (with certification at 19.0%).
- Published
- 2024
7. De novo variants in the RNU4-2 snRNA cause a frequent neurodevelopmental syndrome.
- Author
-
Chen, Yuyang, Dawes, Ruebena, Kim, Hyung, Ljungdahl, Alicia, Stenton, Sarah, Walker, Susan, Lord, Jenny, Lemire, Gabrielle, Martin-Geary, Alexandra, Ganesh, Vijay, Ma, Jialan, Ellingford, Jamie, Delage, Erwan, DSouza, Elston, Dong, Shan, Adams, David, Allan, Kirsten, Bakshi, Madhura, Baldwin, Erin, Berger, Seth, Bernstein, Jonathan, Bhatnagar, Ishita, Blair, Ed, Brown, Natasha, Burrage, Lindsay, Chapman, Kimberly, Coman, David, Compton, Alison, Cunningham, Chloe, DSouza, Precilla, Danecek, Petr, Délot, Emmanuèle, Dias, Kerith-Rae, Elias, Ellen, Elmslie, Frances, Evans, Care-Anne, Ewans, Lisa, Ezell, Kimberly, Fraser, Jamie, Gallacher, Lyndon, Genetti, Casie, Goriely, Anne, Grant, Christina, Haack, Tobias, Higgs, Jenny, Hinch, Anjali, Hurles, Matthew, Kuechler, Alma, Lachlan, Katherine, Lalani, Seema, Lecoquierre, François, Leitão, Elsa, Fevre, Anna, Leventer, Richard, Liebelt, Jan, Lindsay, Sarah, Lockhart, Paul, Ma, Alan, Macnamara, Ellen, Mansour, Sahar, Maurer, Taylor, Mendez, Hector, Metcalfe, Kay, Montgomery, Stephen, Moosajee, Mariya, Nassogne, Marie-Cécile, Neumann, Serena, ODonoghue, Michael, OLeary, Melanie, Palmer, Elizabeth, Pattani, Nikhil, Phillips, John, Pitsava, Georgia, Pysar, Ryan, Rehm, Heidi, Reuter, Chloe, Revencu, Nicole, Riess, Angelika, Rius, Rocio, Rodan, Lance, Roscioli, Tony, Rosenfeld, Jill, Sachdev, Rani, Shaw-Smith, Charles, Simons, Cas, Sisodiya, Sanjay, Snell, Penny, St Clair, Laura, Stark, Zornitza, Stewart, Helen, Tan, Tiong, Tan, Natalie, Temple, Suzanna, Thorburn, David, Tifft, Cynthia, Uebergang, Eloise, VanNoy, Grace, Vasudevan, Pradeep, Vilain, Eric, and Viskochil, David
- Subjects
Humans ,RNA ,Small Nuclear ,Neurodevelopmental Disorders ,Female ,Male ,Brain ,Heterozygote ,Alleles ,Syndrome ,Spliceosomes ,Animals - Abstract
Around 60% of individuals with neurodevelopmental disorders (NDD) remain undiagnosed after comprehensive genetic testing, primarily of protein-coding genes1. Large genome-sequenced cohorts are improving our ability to discover new diagnoses in the non-coding genome. Here we identify the non-coding RNA RNU4-2 as a syndromic NDD gene. RNU4-2 encodes the U4 small nuclear RNA (snRNA), which is a critical component of the U4/U6.U5 tri-snRNP complex of the major spliceosome2. We identify an 18 base pair region of RNU4-2 mapping to two structural elements in the U4/U6 snRNA duplex (the T-loop and stem III) that is severely depleted of variation in the general population, but in which we identify heterozygous variants in 115 individuals with NDD. Most individuals (77.4%) have the same highly recurrent single base insertion (n.64_65insT). In 54 individuals in whom it could be determined, the de novo variants were all on the maternal allele. We demonstrate that RNU4-2 is highly expressed in the developing human brain, in contrast to RNU4-1 and other U4 homologues. Using RNA sequencing, we show how 5 splice-site use is systematically disrupted in individuals with RNU4-2 variants, consistent with the known role of this region during spliceosome activation. Finally, we estimate that variants in this 18 base pair region explain 0.4% of individuals with NDD. This work underscores the importance of non-coding genes in rare disorders and will provide a diagnosis to thousands of individuals with NDD worldwide.
- Published
- 2024
8. cymyc -- Calabi-Yau Metrics, Yukawas, and Curvature
- Author
-
Berglund, Per, Butbaia, Giorgi, Hübsch, Tristan, Jejjala, Vishnu, Mishra, Challenger, Peña, Damián Mayorga, and Tan, Justin
- Subjects
High Energy Physics - Theory ,Computer Science - Machine Learning ,High Energy Physics - Phenomenology - Abstract
We introduce \texttt{cymyc}, a high-performance Python library for numerical investigation of the geometry of a large class of string compactification manifolds and their associated moduli spaces. We develop a well-defined geometric ansatz to numerically model tensor fields of arbitrary degree on a large class of Calabi-Yau manifolds. \texttt{cymyc} includes a machine learning component which incorporates this ansatz to model tensor fields of interest on these spaces by finding an approximate solution to the system of partial differential equations they should satisfy., Comment: 35 pages, 12 figures
- Published
- 2024
9. IPPON: Common Sense Guided Informative Path Planning for Object Goal Navigation
- Author
-
Qu, Kaixian, Tan, Jie, Zhang, Tingnan, Xia, Fei, Cadena, Cesar, and Hutter, Marco
- Subjects
Computer Science - Robotics ,Computer Science - Artificial Intelligence ,Computer Science - Computation and Language - Abstract
Navigating efficiently to an object in an unexplored environment is a critical skill for general-purpose intelligent robots. Recent approaches to this object goal navigation problem have embraced a modular strategy, integrating classical exploration algorithms-notably frontier exploration-with a learned semantic mapping/exploration module. This paper introduces a novel informative path planning and 3D object probability mapping approach. The mapping module computes the probability of the object of interest through semantic segmentation and a Bayes filter. Additionally, it stores probabilities for common objects, which semantically guides the exploration based on common sense priors from a large language model. The planner terminates when the current viewpoint captures enough voxels identified with high confidence as the object of interest. Although our planner follows a zero-shot approach, it achieves state-of-the-art performance as measured by the Success weighted by Path Length (SPL) and Soft SPL in the Habitat ObjectNav Challenge 2023, outperforming other works by more than 20%. Furthermore, we validate its effectiveness on real robots. Project webpage: https://ippon-paper.github.io/
- Published
- 2024
10. Analyzing Neural Network Robustness Using Graph Curvature
- Author
-
Tan, Shuhang, Sia, Jayson, Bogdan, Paul, and Ivanov, Radoslav
- Subjects
Computer Science - Machine Learning - Abstract
This paper presents a new look at the neural network (NN) robustness problem, from the point of view of graph theory analysis, specifically graph curvature. Graph curvature (e.g., Ricci curvature) has been used to analyze system dynamics and identify bottlenecks in many domains, including road traffic analysis and internet routing. We define the notion of neural Ricci curvature and use it to identify bottleneck NN edges that are heavily used to ``transport data" to the NN outputs. We provide an evaluation on MNIST that illustrates that such edges indeed occur more frequently for inputs where NNs are less robust. These results will serve as the basis for an alternative method of robust training, by minimizing the number of bottleneck edges.
- Published
- 2024
11. Enhancing Zero-Shot Vision Models by Label-Free Prompt Distribution Learning and Bias Correcting
- Author
-
Zhu, Xingyu, Zhu, Beier, Tan, Yi, Wang, Shuo, Hao, Yanbin, and Zhang, Hanwang
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Vision-language models, such as CLIP, have shown impressive generalization capacities when using appropriate text descriptions. While optimizing prompts on downstream labeled data has proven effective in improving performance, these methods entail labor costs for annotations and are limited by their quality. Additionally, since CLIP is pre-trained on highly imbalanced Web-scale data, it suffers from inherent label bias that leads to suboptimal performance. To tackle the above challenges, we propose a label-Free prompt distribution learning and bias correction framework, dubbed as **Frolic**, which boosts zero-shot performance without the need for labeled data. Specifically, our Frolic learns distributions over prompt prototypes to capture diverse visual representations and adaptively fuses these with the original CLIP through confidence matching. This fused model is further enhanced by correcting label bias via a label-free logit adjustment. Notably, our method is not only training-free but also circumvents the necessity for hyper-parameter tuning. Extensive experimental results across 16 datasets demonstrate the efficacy of our approach, particularly outperforming the state-of-the-art by an average of $2.6\%$ on 10 datasets with CLIP ViT-B/16 and achieving an average margin of $1.5\%$ on ImageNet and its five distribution shifts with CLIP ViT-B/16. Codes are available in https://github.com/zhuhsingyuu/Frolic., Comment: NeurIPS 2024 Spotlight
- Published
- 2024
12. ODDN: Addressing Unpaired Data Challenges in Open-World Deepfake Detection on Online Social Networks
- Author
-
Tao, Renshuai, Le, Manyi, Tan, Chuangchuang, Liu, Huan, Qin, Haotong, and Zhao, Yao
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Despite significant advances in deepfake detection, handling varying image quality, especially due to different compressions on online social networks (OSNs), remains challenging. Current methods succeed by leveraging correlations between paired images, whether raw or compressed. However, in open-world scenarios, paired data is scarce, with compressed images readily available but corresponding raw versions difficult to obtain. This imbalance, where unpaired data vastly outnumbers paired data, often leads to reduced detection performance, as existing methods struggle without corresponding raw images. To overcome this issue, we propose a novel approach named the open-world deepfake detection network (ODDN), which comprises two core modules: open-world data aggregation (ODA) and compression-discard gradient correction (CGC). ODA effectively aggregates correlations between compressed and raw samples through both fine-grained and coarse-grained analyses for paired and unpaired data, respectively. CGC incorporates a compression-discard gradient correction to further enhance performance across diverse compression methods in OSN. This technique optimizes the training gradient to ensure the model remains insensitive to compression variations. Extensive experiments conducted on 17 popular deepfake datasets demonstrate the superiority of the ODDN over SOTA baselines., Comment: 9 pages, 4 figures
- Published
- 2024
13. Topological 8d $\mathcal{N}=1$ Gauge Theory: Novel Floer Homologies, and $A_\infty$-categories of Six, Five, and Four-Manifolds
- Author
-
Er, Arif and Tan, Meng-Chwan
- Subjects
High Energy Physics - Theory ,Mathematics - Algebraic Geometry ,Mathematics - Differential Geometry ,Mathematics - Geometric Topology ,Mathematics - Symplectic Geometry - Abstract
This work is a continuation of the program initiated in [arXiv:2311.18302]. We show how one can define novel gauge-theoretic (holomorphic) Floer homologies of seven, six, and five-manifolds, from the physics of a topologically-twisted 8d $\mathcal{N}=1$ gauge theory on a Spin$(7)$-manifold via its supersymmetric quantum mechanics interpretation. They are associated with $G_2$ instanton, Donaldson-Thomas, and Haydys-Witten configurations on the seven, six, and five-manifolds, respectively. We also show how one can define hyperk\"ahler Floer homologies specified by hypercontact three-manifolds, and symplectic Floer homologies of instanton moduli spaces. In turn, this will allow us to derive Atiyah-Floer type dualities between the various gauge-theoretic Floer homologies and symplectic intersection Floer homologies of instanton moduli spaces. Via a 2d gauged Landau-Ginzburg model interpretation of the 8d theory, one can derive novel Fukaya-Seidel type $A_\infty$-categories that categorify Donaldson-Thomas, Haydys-Witten, and Vafa-Witten configurations on six, five, and four-manifolds, respectively, where an Atiyah-Floer type correspondence for the Donaldson-Thomas case can be established. Last but not least, topological invariance of the theory suggests a relation amongst these Floer homologies and Fukaya-Seidel type $A_\infty$-categories for certain Spin$(7)$-manifolds. Our work therefore furnishes purely physical proofs and generalizations of the conjectures by Donaldson-Thomas [2], Donaldson-Segal [3], Cherkis [4], Hohloch-Noetzel-Salamon [5], Salamon [6], Haydys [7], and Bousseau [8], and more., Comment: 82 pp
- Published
- 2024
14. PRACT: Optimizing Principled Reasoning and Acting of LLM Agent
- Author
-
Liu, Zhiwei, Yao, Weiran, Zhang, Jianguo, Murthy, Rithesh, Yang, Liangwei, Liu, Zuxin, Lan, Tian, Zhu, Ming, Tan, Juntao, Kokane, Shirley, Hoang, Thai, Niebles, Juan Carlos, Heinecke, Shelby, Wang, Huan, Savarese, Silvio, and Xiong, Caiming
- Subjects
Computer Science - Artificial Intelligence - Abstract
We introduce the Principled Reasoning and Acting (PRAct) framework, a novel method for learning and enforcing action principles from trajectory data. Central to our approach is the use of text gradients from a reflection and optimization engine to derive these action principles. To adapt action principles to specific task requirements, we propose a new optimization framework, Reflective Principle Optimization (RPO). After execution, RPO employs a reflector to critique current action principles and an optimizer to update them accordingly. We develop the RPO framework under two scenarios: Reward-RPO, which uses environmental rewards for reflection, and Self-RPO, which conducts self-reflection without external rewards. Additionally, two RPO methods, RPO-Traj and RPO-Batch, is introduced to adapt to different settings. Experimental results across four environments demonstrate that the PRAct agent, leveraging the RPO framework, effectively learns and applies action principles to enhance performance., Comment: Accepted to SIG CoNLL 2024
- Published
- 2024
15. Uncertainty-Error correlations in Evidential Deep Learning models for biomedical segmentation
- Author
-
Tan, Hai Siong, Wang, Kuancheng, and Mcbeth, Rafe
- Subjects
Electrical Engineering and Systems Science - Image and Video Processing ,Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Machine Learning ,Physics - Medical Physics - Abstract
In this work, we examine the effectiveness of an uncertainty quantification framework known as Evidential Deep Learning applied in the context of biomedical image segmentation. This class of models involves assigning Dirichlet distributions as priors for segmentation labels, and enables a few distinct definitions of model uncertainties. Using the cardiac and prostate MRI images available in the Medical Segmentation Decathlon for validation, we found that Evidential Deep Learning models with U-Net backbones generally yielded superior correlations between prediction errors and uncertainties relative to the conventional baseline equipped with Shannon entropy measure, Monte-Carlo Dropout and Deep Ensemble methods. We also examined these models' effectiveness in active learning, finding that relative to the standard Shannon entropy-based sampling, they yielded higher point-biserial uncertainty-error correlations while attaining similar performances in Dice-Sorensen coefficients. These superior features of EDL models render them well-suited for segmentation tasks that warrant a critical sensitivity in detecting large model errors., Comment: 15 pages
- Published
- 2024
16. Local regularity and finite-time singularity for a class of generalized SQG patches on the half-plane
- Author
-
Miao, Qianyun, Tan, Changhui, Xue, Liutang, and Xue, Zhilong
- Subjects
Mathematics - Analysis of PDEs - Abstract
In this paper, we investigate a class of inviscid generalized surface quasi-geostrophic (SQG) equations on the half-plane with a rigid boundary. Compared to the Biot-Savart law in the vorticity form of the 2D Euler equation, the velocity formula here includes an additional Fourier multiplier operator $m(\Lambda)$. When $m(\Lambda) = \Lambda^\alpha$, where $\Lambda = (-\Delta)^{1/2}$ and $\alpha\in (0,2)$, the equation reduces to the well-known $\alpha$-SQG equation. Finite-time singularity formation for patch solutions to the $\alpha$-SQG equation was famously discovered by Kiselev, Ryzhik, Yao, and Zlato\v{s} [Ann. Math., 184 (2016), pp. 909-948]. We establish finite-time singularity formation for patch solutions to the generalized SQG equations under the Osgood condition \[\int_2^\infty \frac{1}{r (\log r) m(r)} dr < \infty\] along with some additional mild conditions. Notably, our result fills the gap between the globally well-posed 2D Euler equation ($\alpha = 0$) and the $\alpha$-SQG equation ($\alpha > 0$). Furthermore, in line with Elgindi's global regularity results for 2D Loglog-Euler type equations [Arch. Rat. Mech. Anal., 211 (2014), pp. 965-990], our findings suggest that the Osgood condition serves as a sharp threshold that distinguishes global regularity and finite-time singularity in these models. In addition, we generalize the local regularity and finite-time singularity results for patch solutions to the $\alpha$-SQG equation, as established by Gancedo and Patel [Ann. PDE, 7 (2021), no. 1, Art. no. 4], extending them to cases where $m(r)$ behaves like $r^\alpha$ near infinity but does not have an explicit formulation., Comment: 62 pages, 4 figures
- Published
- 2024
17. Paths and Intersections: Characterization of Quasi-metrics in Directed Okamura-Seymour Instances
- Author
-
Chen, Yu and Tan, Zihan
- Subjects
Computer Science - Data Structures and Algorithms ,Computer Science - Discrete Mathematics ,Mathematics - Combinatorics - Abstract
We study the following distance realization problem. Given a quasi-metric $D$ on a set $T$ of terminals, does there exist a directed Okamura-Seymour graph that realizes $D$ as the (directed) shortest-path distance metric on $T$? We show that, if we are further given the circular ordering of terminals lying on the boundary, then Monge property is a sufficient and necessary condition. This generalizes previous results for undirected Okamura-Seymour instances. With the circular ordering, we give a greedy algorithm for constructing a directed Okamura-Seymour instance that realizes the input quasi-metric. The algorithm takes the dual perspective concerning flows and routings, and is based on a new way of analyzing graph structures, by viewing graphs as \emph{paths and their intersections}. We believe this new understanding is of independent interest and will prove useful in other problems in graph theory and graph algorithms. We also design an efficient algorithm for finding such a circular ordering that makes $D$ satisfy Monge property, if one exists. Combined with our result above, this gives an efficient algorithm for the distance realization problem.
- Published
- 2024
18. Imaging magnetic switching in orthogonally twisted stacks of a van der Waals antiferromagnet
- Author
-
Healey, Alexander J, Tan, Cheng, Gross, Boris, Scholten, Sam C, Xing, Kaijian, Chica, Daniel G, Johnson, Brett C, Poggio, Martino, Ziebel, Michael E, Roy, Xavier, Tetienne, Jean-Philippe, and Broadway, David A
- Subjects
Condensed Matter - Mesoscale and Nanoscale Physics ,Physics - Applied Physics - Abstract
Stacking van der Waals magnets holds promise for creating new hybrid materials with properties that do not exist in bulk materials. Here we investigate orthogonally twisted stacks of the van der Waals antiferromagnet CrSBr, aiming to exploit an extreme misalignment of magnetic anisotropy across the twisted interface.Using nitrogen-vacancy centre microscopy, we construct vector maps of the magnetisation, and track their evolution under an external field, in a range of twisted compensated and uncompensated configurations differing by the number of layers. We show that twisted stacking consistently modifies the local magnetic switching behaviour of constituent flakes, and that these modifications are spatially non-uniform. In the case of compensated component flakes (even number of layers), we demonstrate that the combination of dipolar coupling and stacking-induced strain can reduce the switching field by over an order of magnitude. Conversely, in uncompensated component flakes (odd number of layers), we observe indications of a non-zero interlayer exchange interaction between twisted flakes during magnetization reversal, which can persistently modify magnetic order. This work highlights the importance of spatial imaging in investigating stacking-induced magnetic effects, particularly in the case of twistronics where spatial variation is expected and can be conflated with structural imperfections., Comment: 6 main text pages, 4 figures plus 8 supplementary information pages, 9 figures
- Published
- 2024
19. Convex elements and Steinberg's cross-sections
- Author
-
Nie, Sian, Tan, Panjun, and Yu, Qingchao
- Subjects
Mathematics - Representation Theory ,Mathematics - Algebraic Geometry ,Mathematics - Group Theory ,20G99, 20F55 - Abstract
In this paper, we study convex elements in a (twisted) Weyl group introduced by Ivanov and the first named author. We show that each conjugacy class of the twisted Weyl group contains a convex element, and moreover, the Steinberg cross-sections exist for all convex elements. This result strictly enlarges the cases of Steinberg cross-sections from a new perspective, and will play an essential role in the study of higher Deligne-Lusztig representations., Comment: 11 pages
- Published
- 2024
20. WAFFLE: Multi-Modal Model for Automated Front-End Development
- Author
-
Liang, Shanchao, Jiang, Nan, Qian, Shangshu, and Tan, Lin
- Subjects
Computer Science - Software Engineering ,Computer Science - Computation and Language ,Computer Science - Computer Vision and Pattern Recognition - Abstract
Web development involves turning UI designs into functional webpages, which can be difficult for both beginners and experienced developers due to the complexity of HTML's hierarchical structures and styles. While Large Language Models (LLMs) have shown promise in generating source code, two major challenges persist in UI-to-HTML code generation: (1) effectively representing HTML's hierarchical structure for LLMs, and (2) bridging the gap between the visual nature of UI designs and the text-based format of HTML code. To tackle these challenges, we introduce Waffle, a new fine-tuning strategy that uses a structure-aware attention mechanism to improve LLMs' understanding of HTML's structure and a contrastive fine-tuning approach to align LLMs' understanding of UI images and HTML code. Models fine-tuned with Waffle show up to 9.00 pp (percentage point) higher HTML match, 0.0982 higher CW-SSIM, 32.99 higher CLIP, and 27.12 pp higher LLEM on our new benchmark WebSight-Test and an existing benchmark Design2Code, outperforming current fine-tuning methods.
- Published
- 2024
21. Trajectory Optimization for Spatial Microstructure Control in Electron Beam Metal Additive Manufacturing
- Author
-
Khrenov, Mikhail, Tan, Moon, Fitzwater, Lauren, Hobdari, Michelle, and Narra, Sneha Prabha
- Subjects
Electrical Engineering and Systems Science - Systems and Control ,Mathematics - Optimization and Control - Abstract
Metal additive manufacturing (AM) opens the possibility for spatial control of as-fabricated microstructure and properties. However, since the solid state diffusional transformations that drive microstructure outcomes are governed by nonlinear ODEs in terms of temperature, which is itself governed by PDEs over the entire part domain, solving for the system inputs needed to achieve desired microstructure distributions has proven difficult. In this work, we present a trajectory optimization approach for spatial control of microstructure in metal AM, which we demonstrate by controlling the hardness of a low-alloy steel in electron beam powder bed fusion (EB-PBF). To this end, we present models for thermal and microstructural dynamics. Next, we use experimental data to identify the parameters of the microstructure transformation dynamics. We then pose spatial microstructure control as a finite-horizon optimal control problem. The optimal power field trajectory is computed using an augmented Lagrangian differential dynamic programming (AL-DDP) method with GPU acceleration. The resulting time-varying power fields are then realized on an EB-PBF machine through an approximation scheme. Measurements of the resultant hardness shows that the optimized power field trajectory is able to closely produce the desired hardness distribution., Comment: 6 pages, 6 figures
- Published
- 2024
22. Measurements of $\psi{(2S)}$ and $\chi_{c1}(3872)$ production within fully reconstructed jets
- Author
-
LHCb collaboration, Aaij, R., Abdelmotteleb, A. S. W., Beteta, C. Abellan, Abudinén, F., Ackernley, T., Adefisoye, A. A., Adeva, B., Adinolfi, M., Adlarson, P., Agapopoulou, C., Aidala, C. A., Ajaltouni, Z., Akar, S., Akiba, K., Albicocco, P., Albrecht, J., Alessio, F., Alexander, M., Aliouche, Z., Cartelle, P. Alvarez, Amalric, R., Amato, S., Amey, J. L., Amhis, Y., An, L., Anderlini, L., Andersson, M., Andreianov, A., Andreola, P., Andreotti, M., Andreou, D., Anelli, A., Ao, D., Archilli, F., Argenton, M., Cuendis, S. Arguedas, Artamonov, A., Artuso, M., Aslanides, E., Da Silva, R. Ataíde, Atzeni, M., Audurier, B., Bacher, D., Perea, I. Bachiller, Bachmann, S., Bachmayer, M., Back, J. J., Rodriguez, P. Baladron, Balagura, V., Baldini, W., Balzani, L., Bao, H., Leite, J. Baptista de Souza, Pretel, C. Barbero, Barbetti, M., Barbosa, I. R., Barlow, R. J., Barnyakov, M., Barsuk, S., Barter, W., Bartolini, M., Bartz, J., Basels, J. M., Bashir, S., Bassi, G., Batsukh, B., Battista, P. B., Bay, A., Beck, A., Becker, M., Bedeschi, F., Bediaga, I. B., Behling, N. A., Belin, S., Bellee, V., Belous, K., Belov, I., Belyaev, I., Benane, G., Bencivenni, G., Ben-Haim, E., Berezhnoy, A., Bernet, R., Andres, S. Bernet, Bertolin, A., Betancourt, C., Betti, F., Bex, J., Bezshyiko, Ia., Bhom, J., Bieker, M. S., Biesuz, N. V., Billoir, P., Biolchini, A., Birch, M., Bishop, F. C. R., Bitadze, A., Bizzeti, A., Blake, T., Blanc, F., Blank, J. E., Blusk, S., Bocharnikov, V., Boelhauve, J. A., Garcia, O. Boente, Boettcher, T., Bohare, A., Boldyrev, A., Bolognani, C. S., Bolzonella, R., Bondar, N., Bordelius, A., Borgato, F., Borghi, S., Borsato, M., Borsuk, J. T., Bouchiba, S. A., Bovill, M., Bowcock, T. J. V., Boyer, A., Bozzi, C., Rodriguez, A. Brea, Breer, N., Brodzicka, J., Gonzalo, A. Brossa, Brown, J., Brundu, D., Buchanan, E., Buonaura, A., Buonincontri, L., Burke, A. T., Burr, C., Butter, J. S., Buytaert, J., Byczynski, W., Cadeddu, S., Cai, H., Caillet, A. C., Calabrese, R., Ramirez, S. Calderon, Calefice, L., Cali, S., Calvi, M., Gomez, M. Calvo, Magalhaes, P. Camargo, Bouzas, J. I. Cambon, Campana, P., Perez, D. H. Campora, Quezada, A. F. Campoverde, Capelli, S., Capriotti, L., Caravaca-Mora, R., Carbone, A., Salgado, L. Carcedo, Cardinale, R., Cardini, A., Carniti, P., Carus, L., Vidal, A. Casais, Caspary, R., Casse, G., Godinez, J. Castro, Cattaneo, M., Cavallero, G., Cavallini, V., Celani, S., Cervenkov, D., Cesare, S., Chadwick, A. J., Chahrour, I., Charles, M., Charpentier, Ph., Chatzianagnostou, E., Chefdeville, M., Chen, C., Chen, S., Chen, Z., Chernov, A., Chernyshenko, S., Chiotopoulos, X., Chobanova, V., Cholak, S., Chrzaszcz, M., Chubykin, A., Chulikov, V., Ciambrone, P., Vidal, X. Cid, Ciezarek, G., Cifra, P., Clarke, P. E. L., Clemencic, M., Cliff, H. V., Closier, J., Toapaxi, C. Cocha, Coco, V., Cogan, J., Cogneras, E., Cojocariu, L., Collins, P., Colombo, T., Colonna, M. C., Comerma-Montells, A., Congedo, L., Contu, A., Cooke, N., Corredoira, I., Correia, A., Corti, G., Meldrum, J. J. Cottee, Couturier, B., Craik, D. C., Torres, M. Cruz, Rivera, E. Curras, Currie, R., Da Silva, C. L., Dadabaev, S., Dai, L., Dai, X., Dall'Occo, E., Dalseno, J., D'Ambrosio, C., Daniel, J., Danilina, A., d'Argent, P., Davidson, A., Davies, J. E., Davis, A., Francisco, O. De Aguiar, De Angelis, C., De Benedetti, F., de Boer, J., De Bruyn, K., De Capua, S., De Cian, M., Da Graca, U. De Freitas Carneiro, De Lucia, E., De Miranda, J. M., De Paula, L., De Serio, M., De Simone, P., De Vellis, F., de Vries, J. A., Debernardis, F., Decamp, D., Dedu, V., Dekkers, S., Del Buono, L., Delaney, B., Dembinski, H. -P., Deng, J., Denysenko, V., Deschamps, O., Dettori, F., Dey, B., Di Nezza, P., Diachkov, I., Didenko, S., Ding, S., Dittmann, L., Dobishuk, V., Docheva, A. D., Dong, C., Donohoe, A. M., Dordei, F., Reis, A. C. dos, Dowling, A. D., Duan, W., Duda, P., Dudek, M. W., Dufour, L., Duk, V., Durante, P., Duras, M. M., Durham, J. M., Durmus, O. D., Dziurda, A., Dzyuba, A., Easo, S., Eckstein, E., Egede, U., Egorychev, A., Egorychev, V., Eisenhardt, S., Ejopu, E., Eklund, L., Elashri, M., Ellbracht, J., Ely, S., Ene, A., Epple, E., Eschle, J., Esen, S., Evans, T., Fabiano, F., Falcao, L. N., Fan, Y., Fang, B., Fantini, L., Faria, M., Farmer, K., Fazzini, D., Felkowski, L., Feng, M., Feo, M., Casani, A. Fernandez, Gomez, M. Fernandez, Fernez, A. D., Ferrari, F., Rodrigues, F. Ferreira, Ferrillo, M., Ferro-Luzzi, M., Filippov, S., Fini, R. A., Fiorini, M., Fischer, K. L., Fitzgerald, D. S., Fitzpatrick, C., Fleuret, F., Fontana, M., Foreman, L. F., Forty, R., Foulds-Holt, D., Lima, V. Franco, Sevilla, M. Franco, Frank, M., Franzoso, E., Frau, G., Frei, C., Friday, D. A., Fu, J., Führing, Q., Fujii, Y., Fulghesu, T., Gabriel, E., Galati, G., Galati, M. D., Torreira, A. Gallas, Galli, D., Gambetta, S., Gandelman, M., Gandini, P., Ganie, B., Gao, H., Gao, R., Gao, T. Q., Gao, Y., Garau, M., Martin, L. M. Garcia, Moreno, P. Garcia, Pardiñas, J. García, Garg, K. G., Garrido, L., Gaspar, C., Geertsema, R. E., Gerken, L. L., Gersabeck, E., Gersabeck, M., Gershon, T., Ghizzo, S. G., Ghorbanimoghaddam, Z., Giambastiani, L., Giasemis, F. I., Gibson, V., Giemza, H. K., Gilman, A. L., Giovannetti, M., Gioventù, A., Girardey, L., Gironell, P. Gironella, Giugliano, C., Giza, M. A., Gkougkousis, E. L., Glaser, F. C., Gligorov, V. V., Göbel, C., Golobardes, E., Golubkov, D., Golutvin, A., Gomes, A., Fernandez, S. Gomez, Abrantes, F. Goncalves, Goncerz, M., Gong, G., Gooding, J. A., Gorelov, I. V., Gotti, C., Grabowski, J. P., Cardoso, L. A. Granado, Graugés, E., Graverini, E., Grazette, L., Graziani, G., Grecu, A. T., Greeven, L. M., Grieser, N. A., Grillo, L., Gromov, S., Gu, C., Guarise, M., Guerry, L., Guittiere, M., Guliaeva, V., Günther, P. A., Guseinov, A. -K., Gushchin, E., Guz, Y., Gys, T., Habermann, K., Hadavizadeh, T., Hadjivasiliou, C., Haefeli, G., Haen, C., Haimberger, J., Hajheidari, M., Hallett, G., Halvorsen, M. M., Hamilton, P. M., Hammerich, J., Han, Q., Han, X., Hansmann-Menzemer, S., Hao, L., Harnew, N., Hartmann, M., Hashmi, S., He, J., Hemmer, F., Henderson, C., Henderson, R. D. L., Hennequin, A. M., Hennessy, K., Henry, L., Herd, J., Gascon, P. Herrero, Heuel, J., Hicheur, A., Mendizabal, G. Hijano, Hill, D., Hollitt, S. E., Horswill, J., Hou, R., Hou, Y., Howarth, N., Hu, J., Hu, W., Hu, X., Huang, W., Hulsbergen, W., Hunter, R. J., Hushchyn, M., Hutchcroft, D., Ilin, D., Ilten, P., Inglessi, A., Iniukhin, A., Ishteev, A., Ivshin, K., Jacobsson, R., Jage, H., Elles, S. J. Jaimes, Jakobsen, S., Jans, E., Jashal, B. K., Jawahery, A., Jevtic, V., Jiang, E., Jiang, X., Jiang, Y., Jiang, Y. J., John, M., Rajan, A. John Rubesh, Johnson, D., Jones, C. R., Jones, T. P., Joshi, S., Jost, B., Castella, J. Juan, Jurik, N., Juszczak, I., Kaminaris, D., Kandybei, S., Kane, M., Kang, Y., Kar, C., Karacson, M., Karpenkov, D., Kauniskangas, A., Kautz, J. W., Kazanecki, M. K., Keizer, F., Kenzie, M., Ketel, T., Khanji, B., Kharisova, A., Kholodenko, S., Khreich, G., Kirn, T., Kirsebom, V. S., Kitouni, O., Klaver, S., Kleijne, N., Klimaszewski, K., Kmiec, M. R., Koliiev, S., Kolk, L., Konoplyannikov, A., Kopciewicz, P., Koppenburg, P., Korolev, M., Kostiuk, I., Kot, O., Kotriakhova, S., Kozachuk, A., Kravchenko, P., Kravchuk, L., Kreps, M., Krokovny, P., Krupa, W., Krzemien, W., Kshyvanskyi, O., Kubat, J., Kubis, S., Kucharczyk, M., Kudryavtsev, V., Kulikova, E., Kupsc, A., Kutsenko, B. K., Lacarrere, D., Gonzalez, P. Laguarta, Lai, A., Lampis, A., Lancierini, D., Gomez, C. Landesa, Lane, J. J., Lane, R., Lanfranchi, G., Langenbruch, C., Langer, J., Lantwin, O., Latham, T., Lazzari, F., Lazzeroni, C., Gac, R. Le, Lee, H., Lefèvre, R., Leflat, A., Legotin, S., Lehuraux, M., Cid, E. Lemos, Leroy, O., Lesiak, T., Lesser, E. D., Leverington, B., Li, A., Li, C., Li, H., Li, K., Li, L., Li, P., Li, P. -R., Li, Q., Li, S., Li, T., Li, Y., Lian, Z., Liang, X., Libralon, S., Lin, C., Lin, T., Lindner, R., Lisovskyi, V., Litvinov, R., Liu, F. L., Liu, G., Liu, K., Liu, S., Liu, W., Liu, Y., Liu, Y. L., Salvia, A. Lobo, Loi, A., Castro, J. Lomba, Long, T., Lopes, J. H., Huertas, A. Lopez, Soliño, S. López, Lu, Q., Lucarelli, C., Lucchesi, D., Martinez, M. Lucio, Lukashenko, V., Luo, Y., Lupato, A., Luppi, E., Lynch, K., Lyu, X. -R., Ma, G. M., Ma, R., Maccolini, S., Machefert, F., Maciuc, F., Mack, B., Mackay, I., Mackey, L. M., Mohan, L. R. Madhan, Madurai, M. J., Maevskiy, A., Magdalinski, D., Maisuzenko, D., Majewski, M. W., Malczewski, J. J., Malde, S., Malentacca, L., Malinin, A., Maltsev, T., Manca, G., Mancinelli, G., Mancuso, C., Escalero, R. Manera, Manuzzi, D., Marangotto, D., Marchand, J. F., Marchevski, R., Marconi, U., Mariani, E., Mariani, S., Benito, C. Marin, Marks, J., Marshall, A. M., Martel, L., Martelli, G., Martellotti, G., Martinazzoli, L., Martinelli, M., Santos, D. Martinez, Vidal, F. Martinez, Massafferri, A., Matev, R., Mathad, A., Matiunin, V., Matteuzzi, C., Mattioli, K. R., Mauri, A., Maurice, E., Mauricio, J., Mayencourt, P., de Cos, J. Mazorra, Mazurek, M., McCann, M., Mcconnell, L., McGrath, T. H., McHugh, N. T., McNab, A., McNulty, R., Meadows, B., Meier, G., Melnychuk, D., Meng, F. M., Merk, M., Merli, A., Garcia, L. Meyer, Miao, D., Miao, H., Mikhasenko, M., Milanes, D. A., Minotti, A., Minucci, E., Miralles, T., Mitreska, B., Mitzel, D. S., Modak, A., Mohammed, R. A., Moise, R. D., Mokhnenko, S., Cardenas, E. F. Molina, Mombächer, T., Monk, M., Monteil, S., Gomez, A. Morcillo, Morello, G., Morello, M. J., Morgenthaler, M. P., Morris, A. B., Morris, A. G., Mountain, R., Mu, H., Mu, Z. M., Muhammad, E., Muheim, F., Mulder, M., Müller, K., Muñoz-Rojas, F., Murta, R., Naik, P., Nakada, T., Nandakumar, R., Nanut, T., Nasteva, I., Needham, M., Neri, N., Neubert, S., Neufeld, N., Neustroev, P., Nicolini, J., Nicotra, D., Niel, E. M., Nikitin, N., Nogarolli, P., Nogga, P., Nolte, N. S., Normand, C., Fernandez, J. Novoa, Nowak, G., Nunez, C., Nur, H. N., Oblakowska-Mucha, A., Obraztsov, V., Oeser, T., Okamura, S., Okhotnikov, A., Okhrimenko, O., Oldeman, R., Oliva, F., Olocco, M., Onderwater, C. J. G., O'Neil, R. H., Osthues, D., Goicochea, J. M. Otalora, Owen, P., Oyanguren, A., Ozcelik, O., Paciolla, F., Padee, A., Padeken, K. O., Pagare, B., Pais, P. R., Pajero, T., Palano, A., Palutan, M., Panshin, G., Paolucci, L., Papanestis, A., Pappagallo, M., Pappalardo, L. L., Pappenheimer, C., Parkes, C., Passalacqua, B., Passaleva, G., Passaro, D., Pastore, A., Patel, M., Patoc, J., Patrignani, C., Paul, A., Pawley, C. J., Pellegrino, A., Peng, J., Altarelli, M. Pepe, Perazzini, S., Pereima, D., Da Costa, H. Pereira, Castro, A. Pereiro, Perret, P., Perro, A., Petridis, K., Petrolini, A., Pfaller, J. P., Pham, H., Pica, L., Piccini, M., Pietrzyk, B., Pietrzyk, G., Pinci, D., Pisani, F., Pizzichemi, M., Placinta, V., Casasus, M. Plo, Poeschl, T., Polci, F., Lener, M. Poli, Poluektov, A., Polukhina, N., Polyakov, I., Polycarpo, E., Ponce, S., Popov, D., Poslavskii, S., Prasanth, K., Prouve, C., Provenzano, D., Pugatch, V., Punzi, G., Qasim, S., Qian, Q. Q., Qian, W., Qin, N., Qu, S., Quagliani, R., Trejo, R. I. Rabadan, Rademacker, J. H., Rama, M., García, M. Ramírez, De Oliveira, V. Ramos, Pernas, M. Ramos, Rangel, M. S., Ratnikov, F., Raven, G., De Miguel, M. Rebollo, Redi, F., Reich, J., Reiss, F., Ren, Z., Resmi, P. K., Ribatti, R., Ricart, G. R., Riccardi, D., Ricciardi, S., Richardson, K., Richardson-Slipper, M., Rinnert, K., Robbe, P., Robertson, G., Rodrigues, E., Fernandez, E. Rodriguez, Lopez, J. A. Rodriguez, Rodriguez, E. Rodriguez, Roensch, J., Rogachev, A., Rogovskiy, A., Rolf, D. L., Roloff, P., Romanovskiy, V., Lamas, M. Romero, Vidal, A. Romero, Romolini, G., Ronchetti, F., Rong, T., Rotondo, M., Roy, S. R., Rudolph, M. S., Diaz, M. Ruiz, Fernandez, R. A. Ruiz, Vidal, J. Ruiz, Ryzhikov, A., Ryzka, J., Saavedra-Arias, J. J., Silva, J. J. Saborido, Sadek, R., Sagidova, N., Sahoo, D., Sahoo, N., Saitta, B., Salomoni, M., Gras, C. Sanchez, Sanderswood, I., Santacesaria, R., Rios, C. Santamarina, Santimaria, M., Santoro, L., Santovetti, E., Saputi, A., Saranin, D., Sarnatskiy, A., Sarpis, G., Sarpis, M., Satriano, C., Satta, A., Saur, M., Savrina, D., Sazak, H., Sborzacchi, F., Smead, L. G. Scantlebury, Scarabotto, A., Schael, S., Scherl, S., Schiller, M., Schindler, H., Schmelling, M., Schmidt, B., Schmitt, S., Schmitz, H., Schneider, O., Schopper, A., Schulte, N., Schulte, S., Schune, M. H., Schwemmer, R., Schwering, G., Sciascia, B., Sciuccati, A., Sellam, S., Semennikov, A., Senger, T., Soares, M. Senghi, Sergi, A., Serra, N., Sestini, L., Seuthe, A., Shang, Y., Shangase, D. M., Shapkin, M., Sharma, R. S., Shchemerov, I., Shchutska, L., Shears, T., Shekhtman, L., Shen, Z., Sheng, S., Shevchenko, V., Shi, B., Shi, Q., Shimizu, Y., Shmanin, E., Shorkin, R., Shupperd, J. D., Coutinho, R. Silva, Simi, G., Simone, S., Skidmore, N., Skwarnicki, T., Slater, M. W., Smallwood, J. C., Smith, E., Smith, K., Smith, M., Snoch, A., Lavra, L. Soares, Sokoloff, M. D., Soler, F. J. P., Solomin, A., Solovev, A., Solovyev, I., Song, R., Song, Y., Song, Y. S., De Almeida, F. L. Souza, De Paula, B. Souza, Norella, E. Spadaro, Spedicato, E., Speer, J. G., Spiridenkov, E., Spradlin, P., Sriskaran, V., Stagni, F., Stahl, M., Stahl, S., Stanislaus, S., Stein, E. N., Steinkamp, O., Stenyakin, O., Stevens, H., Strekalina, D., Su, Y., Suljik, F., Sun, J., Sun, L., Sun, Y., Sundfeld, D., Sutcliffe, W., Swallow, P. N., Swystun, F., Szabelski, A., Szumlak, T., Tan, Y., Tat, M. D., Terentev, A., Terzuoli, F., Teubert, F., Thomas, E., Thompson, D. J. D., Tilquin, H., Tisserand, V., T'Jampens, S., Tobin, M., Tomassetti, L., Tonani, G., Tong, X., Machado, D. Torres, Toscano, L., Tou, D. Y., Trippl, C., Tuci, G., Tuning, N., Uecker, L. H., Ukleja, A., Unverzagt, D. J., Ursov, E., Usachov, A., Ustyuzhanin, A., Uwer, U., Vagnoni, V., Cadenas, V. Valcarce, Valenti, G., Canudas, N. Valls, Van Hecke, H., van Herwijnen, E., Van Hulse, C. B., Van Laak, R., van Veghel, M., Vasquez, G., Gomez, R. Vazquez, Regueiro, P. Vazquez, Sierra, C. Vázquez, Vecchi, S., Velthuis, J. J., Veltri, M., Venkateswaran, A., Vesterinen, M., Benet, D. Vico, Villalba, P. Vidrier, Diaz, M. Vieites, Vilasis-Cardona, X., Figueras, E. Vilella, Villa, A., Vincent, P., Volle, F. C., Bruch, D. vom, Voropaev, N., Vos, K., Vouters, G., Vrahas, C., Wagner, J., Walsh, J., Walton, E. J., Wan, G., Wang, C., Wang, G., Wang, J., Wang, M., Wang, N. W., Wang, R., Wang, X., Wang, X. W., Wang, Y., Wang, Z., Ward, J. A., Waterlaat, M., Watson, N. K., Websdale, D., Wei, Y., Wendel, J., Westhenry, B. D. C., White, C., Whitehead, M., Whiter, E., Wiederhold, A. R., Wiedner, D., Wilkinson, G., Wilkinson, M. K., Williams, M., Williams, M. R. J., Williams, R., Williams, Z., Wilson, F. F., Wislicki, W., Witek, M., Witola, L., Wormser, G., Wotton, S. A., Wu, H., Wu, J., Wu, Y., Wu, Z., Wyllie, K., Xian, S., Xiang, Z., Xie, Y., Xu, A., Xu, J., Xu, L., Xu, M., Xu, Z., Yang, D., Yang, K., Yang, S., Yang, X., Yang, Y., Yang, Z., Yeroshenko, V., Yeung, H., Yin, H., Yin, X., Yu, C. Y., Yu, J., Yuan, X., Yuan, Y, Zaffaroni, E., Zavertyaev, M., Zdybal, M., Zenesini, F., Zeng, C., Zeng, M., Zhang, C., Zhang, D., Zhang, J., Zhang, L., Zhang, S., Zhang, Y., Zhang, Y. Z., Zhao, Y., Zharkova, A., Zhelezov, A., Zheng, S. Z., Zheng, X. Z., Zheng, Y., Zhou, T., Zhou, X., Zhou, Y., Zhovkovska, V., Zhu, L. Z., Zhu, X., Zhukov, V., Zhuo, J., Zou, Q., Zuliani, D., and Zunica, G.
- Subjects
High Energy Physics - Experiment - Abstract
This paper presents the first measurement of $\psi{(2S)}$ and $\chi_{c1}(3872)$ meson production within fully reconstructed jets. Each quarkonium state (tag) is reconstructed via its decay to the $J/\psi$($\rightarrow\mu^+\mu^-$)$\pi^+\pi^-$ final state in the forward region using proton-proton collision data collected by the LHCb experiment at the center-of-mass-energy of $13 \text{TeV}$ in 2016, corresponding to an integrated luminosity of $1.64 \text{fb}^{-1}$. The fragmentation function, presented as the ratio of the quarkonium-tag transverse momentum to the full jet transverse momentum ($p_{\mathrm{T}}(\text{tag})/p_{\mathrm{T}}(\text{jet})$), is measured differentially in $p_{\mathrm{T}}(\text{jet})$ and $p_{\mathrm{T}}(\text{tag})$ bins. The distributions are separated into promptly produced quarkonia from proton-proton collisions and quarkonia produced from displaced $b$-hadron decays. While the displaced quarkonia fragmentation functions are in general well described by parton-shower predictions, the prompt quarkonium distributions differ significantly from fixed-order non-relativistic QCD (NRQCD) predictions followed by a QCD parton shower., Comment: All figures and tables, along with machine-readable versions and any supplementary material and additional information, are available at https://lbfence.cern.ch/alcm/public/analysis/full-details/1618/ (LHCb public pages)
- Published
- 2024
23. ProFL: Performative Robust Optimal Federated Learning
- Author
-
Zheng, Xue, Xie, Tian, Tan, Xuwei, Yener, Aylin, Zhang, Xueru, Payani, Ali, and Lee, Myungjin
- Subjects
Computer Science - Machine Learning ,Computer Science - Information Theory - Abstract
Performative prediction (PP) is a framework that captures distribution shifts that occur during the training of machine learning models due to their deployment. As the trained model is used, its generated data could cause the model to evolve, leading to deviations from the original data distribution. The impact of such model-induced distribution shifts in the federated learning (FL) setup remains unexplored despite being increasingly likely to transpire in real-life use cases. Although Jin et al. (2024) recently extended PP to FL in a straightforward manner, the resulting model only converges to a performative stable point, which may be far from optimal. The methods in Izzo et al. (2021); Miller et al. (2021) can find a performative optimal point in centralized settings, but they require the performative risk to be convex and the training data to be noiseless, assumptions often violated in realistic FL systems. This paper overcomes all of these shortcomings and proposes Performative robust optimal Federated Learning (ProFL), an algorithm that finds performative optimal points in FL from noisy and contaminated data. We present the convergence analysis under the Polyak-Lojasiewicz condition, which applies to non-convex objectives. Extensive experiments on multiple datasets validate our proposed algorithms' efficiency., Comment: 27 pages with Appendix, 18 figures. The paper has been submitted and is currently under review
- Published
- 2024
24. Disordered charge density waves in the kagome metal FeGe
- Author
-
Tan, Hengxin and Yan, Binghai
- Subjects
Condensed Matter - Materials Science ,Condensed Matter - Disordered Systems and Neural Networks ,Condensed Matter - Strongly Correlated Electrons - Abstract
The discovery of a charge density wave (CDW) in the antiferromagnetic kagome metal FeGe has prompted interest in the interplay between kagome physics, CDW, and magnetism. However, a crucial aspect for understanding these emergent phenomena-the precise CDW structure-remains ambiguous. Recent studies have assumed uniformly distributed Ge dimers, but this assumption is problematic. The predicted band structure based on this model exhibits an abrupt disappearance of a Ge-$p$ band in the Fermi surface, contradicting experimental observations from angle-resolved photoemission spectroscopy (ARPES). In this study, we propose that a CDW phase with disordered Ge dimers can reconcile theoretical predictions with ARPES results. This model reproduces the observed CDW gaps while preserving the Ge-$p$ band. Depending on experimental conditions, Ge dimers can be randomly distributed or exhibit phase separation from pristine regions. Our findings reveal the crucial role of Ge dimer disorder in the FeGe CDW and suggest potential implications of this disorder for other properties, such as magnetism and transport, in this system., Comment: Supplementary included
- Published
- 2024
25. LongRAG: A Dual-Perspective Retrieval-Augmented Generation Paradigm for Long-Context Question Answering
- Author
-
Zhao, Qingfei, Wang, Ruobing, Cen, Yukuo, Zha, Daren, Tan, Shicheng, Dong, Yuxiao, and Tang, Jie
- Subjects
Computer Science - Computation and Language - Abstract
Long-Context Question Answering (LCQA), a challenging task, aims to reason over long-context documents to yield accurate answers to questions. Existing long-context Large Language Models (LLMs) for LCQA often struggle with the "lost in the middle" issue. Retrieval-Augmented Generation (RAG) mitigates this issue by providing external factual evidence. However, its chunking strategy disrupts the global long-context information, and its low-quality retrieval in long contexts hinders LLMs from identifying effective factual details due to substantial noise. To this end, we propose LongRAG, a general, dual-perspective, and robust LLM-based RAG system paradigm for LCQA to enhance RAG's understanding of complex long-context knowledge (i.e., global information and factual details). We design LongRAG as a plug-and-play paradigm, facilitating adaptation to various domains and LLMs. Extensive experiments on three multi-hop datasets demonstrate that LongRAG significantly outperforms long-context LLMs (up by 6.94%), advanced RAG (up by 6.16%), and Vanilla RAG (up by 17.25%). Furthermore, we conduct quantitative ablation studies and multi-dimensional analyses, highlighting the effectiveness of the system's components and fine-tuning strategies. Data and code are available at https://github.com/QingFei1/LongRAG., Comment: EMNLP 2024 Main
- Published
- 2024
26. Evaluating the performance of machine-learning-based phase pickers when applied to ocean bottom seismic data: Blanco oceanic transform fault as a case study
- Author
-
Liu, Min and Tan, Yen Joe
- Subjects
Physics - Geophysics - Abstract
Machine-learning-based phase pickers have been successfully leveraged to build high-resolution earthquake catalogs using seismic data on land. However, their performance when applied to ocean bottom seismic (OBS) data remains to be evaluated. In this study, we first adopt three machine-learning-based phase pickers - EQTransformer, Pickblue, and OBSTansformer - to build three earthquake catalogs for the 350-km-long Blanco oceanic transform fault (BTF) based on a year-long OBS deployment. We then systematically compare these catalogs with an existing catalog which utilized a traditional workflow. Results indicate that the Pickblue-based catalog documents more events and/or provides better-constrained locations than the other catalogs. The different performances of the three phase pickers suggest that detailed assessment of catalogs built using automatic workflows is necessary to prevent misinterpretations, especially when applied to regions without training samples. The Pickblue-based catalog reveals seismicity gaps in three extensional segments of BTF which likely represent aseismic slip zones affected by seawater infiltration. Furthermore, most earthquakes are shallower than the 600-degree isotherm predicted by a half-space conductive cooling model, except for the Blanco Ridge segment which has hosted 80% of the Mw > 6.0 earthquakes along BTF since 1976. These Blanco Ridge deep earthquake clusters can be explained by hydrothermal cooling or the serpentinization of mantle peridotite due to seawater infiltration along conduits created by the deeper ruptures of large earthquakes. Our analyses also demonstrate the importance of careful examination of automatically produced earthquake catalogs since mislocated events can lead to very different interpretations of fault slip modes from seismicity distribution., Comment: 38 pages and 16 figures
- Published
- 2024
27. Stick-breaking Attention
- Author
-
Tan, Shawn, Shen, Yikang, Yang, Songlin, Courville, Aaron, and Panda, Rameswar
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Computation and Language - Abstract
The self-attention mechanism traditionally relies on the softmax operator, necessitating positional embeddings like RoPE, or position biases to account for token order. But current methods using still face length generalisation challenges. We propose an alternative attention mechanism based on the stick-breaking process: For each token before the current, we determine a break point $\beta_{i,j}$, which represents the proportion of the remaining stick to allocate to the current token. We repeat the process until the stick is fully allocated, resulting in a sequence of attention weights. This process naturally incorporates recency bias, which has linguistic motivations for grammar parsing (Shen et. al., 2017). We study the implications of replacing the conventional softmax-based attention mechanism with stick-breaking attention. We then discuss implementation of numerically stable stick-breaking attention and adapt Flash Attention to accommodate this mechanism. When used as a drop-in replacement for current softmax+RoPE attention systems, we find that stick-breaking attention performs competitively with current methods on length generalisation and downstream tasks. Stick-breaking also performs well at length generalisation, allowing a model trained with $2^{11}$ context window to perform well at $2^{14}$ with perplexity improvements.
- Published
- 2024
28. OmniFlatten: An End-to-end GPT Model for Seamless Voice Conversation
- Author
-
Zhang, Qinglin, Cheng, Luyao, Deng, Chong, Chen, Qian, Wang, Wen, Zheng, Siqi, Liu, Jiaqing, Yu, Hai, and Tan, Chaohong
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Sound ,Electrical Engineering and Systems Science - Audio and Speech Processing - Abstract
Full-duplex spoken dialogue systems significantly advance over traditional turn-based dialogue systems, as they allow simultaneous bidirectional communication, closely mirroring human-human interactions. However, achieving low latency and natural interactions in full-duplex dialogue systems remains a significant challenge, especially considering human conversation dynamics such as interruptions, backchannels, and overlapping speech. In this paper, we introduce a novel End-to-End GPT-based model OmniFlatten for full-duplex conversation, capable of effectively modeling the complex behaviors inherent to natural conversations with low latency. To achieve full-duplex communication capabilities, we propose a multi-stage post-training scheme that progressively adapts a text-based large language model (LLM) backbone into a speech-text dialogue LLM, capable of generating text and speech in real time, without modifying the architecture of the backbone LLM. The training process comprises three stages: modality alignment, half-duplex dialogue learning, and full-duplex dialogue learning. Throughout all training stages, we standardize the data using a flattening operation, which allows us to unify the training methods and the model architecture across different modalities and tasks. Our approach offers a straightforward modeling technique and a promising research direction for developing efficient and natural end-to-end full-duplex spoken dialogue systems. Audio samples of dialogues generated by OmniFlatten can be found at this web site (https://omniflatten.github.io/)., Comment: Work in progress
- Published
- 2024
29. Deterministic formation of carbon-functionalized quantum emitters in hexagonal boron nitride
- Author
-
Luo, Manlin, Ge, Junyu, Huang, Pengru, Yu, Yi, Seo, In Cheol, Lu, Kunze, Sun, Hao, Tan, Jian Kwang, Kim, Sejeong, Gao, Weibo, Li, Hong, and Nam, Donguk
- Subjects
Physics - Applied Physics ,Quantum Physics - Abstract
Forming single-photon emitters (SPEs) in insulating hexagonal boron nitride (hBN) has sparked wide interests in the quantum photonics. Despite significant progress, it remains challenging to deterministically create SPEs at precise locations with a specific type of element for creating defects. In this study, we present a straightforward approach to generate site-deterministic carbon-functionalized quantum emitters in hBN by harnessing ultrasonic nanoindentation. The obtained SPEs are high-quality and can be scaled up to large arrays in a single fabrication step. Comprehensive experimental analyses reveal that the insertion of carbon atoms into the hBN lattice is the source of the robust quantum emission. Complementary theoretical studies suggest possible candidates for the structural origin of the defects based on our experimental results. This rapid and scalable nanoindentation method provides a new way to create SPE arrays with specific types of atoms, enabling the comprehensive investigation of the origins and mechanics of SPE formations in two-dimensional (2D) materials and beyond.
- Published
- 2024
30. Neutrinoless Double Beta Decay Sensitivity of the XLZD Rare Event Observatory
- Author
-
XLZD Collaboration, Aalbers, J., Abe, K., Adrover, M., Maouloud, S. Ahmed, Akerib, D. S., Musalhi, A. K. Al, Alder, F., Althueser, L., Amaral, D. W. P., Amarasinghe, C. S., Ames, A., Andrieu, B., Angelides, N., Angelino, E., Antunovic, B., Aprile, E., Araújo, H. M., Armstrong, J. E., Arthurs, M., Babicz, M., Bajpai, D., Baker, A., Balzer, M., Bang, J., Barberio, E., Bargemann, J. W., Barillier, E., Basharina-Freshville, A., Baudis, L., Bauer, D., Bazyk, M., Beattie, K., Beaupere, N., Bell, N. F., Bellagamba, L., Benson, T., Bhatti, A., Biesiadzinski, T. P., Biondi, R., Biondi, Y., Birch, H. J., Bishop, E., Bismark, A., Boehm, C., Boese, K., Bolotnikov, A., Brás, P., Braun, R., Breskin, A., Brew, C. A. J., Brommer, S., Brown, A., Bruni, G., Budnik, R., Burdin, S., Cai, C., Capelli, C., Carini, G., Carmona-Benitez, M. C., Carter, M., Chauvin, A., Chawla, A., Chen, H., Cherwinka, J. J., Chin, Y. T., Chott, N. I., Chavez, A. P. Cimental, Clark, K., Colijn, A. P., Colling, D. J., Conrad, J., Converse, M. V., Coronel, R., Costanzo, D., Cottle, A., Cox, G., Cuenca-García, J. J., Curran, D., Cussans, D., D'Andrea, V., Garcia, L. C. Daniel, Darlington, I., Dave, S., David, A., Davies, G. J., Decowski, M. P., Deisting, A., Delgaudio, J., Dey, S., Di Donato, C., Di Felice, L., Di Gangi, P., Diglio, S., Ding, C., Dobson, J. E. Y., Doerenkamp, M., Drexlin, G., Druszkiewicz, E., Dunbar, C. L., Eitel, K., Elykov, A., Engel, R., Eriksen, S. R., Fayer, S., Fearon, N. M., Ferella, A. D., Ferrari, C., Fieldhouse, N., Fischer, H., Flaecher, H., Flehmke, T., Flierman, M., Fraser, E. D., Fruth, T. M. A., Fujikawa, K., Fulgione, W., Fuselli, C., Gaemers, P., Gaior, R., Gaitskell, R. J., Gallice, N., Galloway, M., Gao, F., Garroum, N., Geffre, A., Genovesi, J., Ghag, C., Ghosh, S., Giacomobono, R., Gibbons, R., Girard, F., Glade-Beucke, R., Glück, F., Gokhale, S., Grandi, L., Green, J., Grigat, J., van der Grinten, M. G. D., Größle, R., Guan, H., Guida, M., Gyorgy, P., Haiston, J. J., Hall, C. R., Hall, T., Hammann, R., Hannen, V., Hansmann-Menzemer, S., Hargittai, N., Hartigan-O'Connor, E., Haselschwardt, S. J., Hernandez, M., Hertel, S. A., Higuera, A., Hils, C., Hiraoka, K., Hoetzsch, L., Hoferichter, M., Homenides, G. J., Hood, N. F., Horn, M., Huang, D. Q., Hughes, S., Hunt, D., Iacovacci, M., Itow, Y., Jacquet, E., Jakob, J., James, R. S., Joerg, F., Jones, S., Kaboth, A. C., Kahlert, F., Kamaha, A. C., Kaminaga, Y., Kara, M., Kavrigin, P., Kazama, S., Keller, M., Kemp-Russell, P., Khaitan, D., Kharbanda, P., Kilminster, B., Kim, J., Kirk, R., Kleifges, M., Klute, M., Kobayashi, M., Kodroff, D., Koke, D., Kopec, A., Korolkova, E. V., Kraus, H., Kravitz, S., Kreczko, L., von Krosigk, B., Kudryavtsev, V. A., Kuger, F., Kurita, N., Landsman, H., Lang, R. F., Lawes, C., Lee, J., Lehnert, B., Leonard, D. S., Lesko, K. T., Levinson, L., Li, A., Li, I., Li, S., Liang, S., Liang, Z., Lin, J., Lin, Y. -T., Lindemann, S., Linden, S., Lindner, M., Lindote, A., Lippincott, W. H., Liu, K., Loizeau, J., Lombardi, F., Lopes, J. A. M., Lopes, M. I., Lorenzon, W., Loutit, M., Lu, C., Lucchetti, G. M., Luce, T., Luitz, S., Ma, Y., Macolino, C., Mahlstedt, J., Maier, B., Majewski, P. A., Manalaysay, A., Mancuso, A., Manenti, L., Mannino, R. L., Marignetti, F., Marley, T., Undagoitia, T. Marrodán, Martens, K., Masbou, J., Masson, E., Mastroianni, S., Maupin, C., McCabe, C., McCarthy, M. E., McKinsey, D. N., McLaughlin, J. B., Melchiorre, A., Menéndez, J., Messina, M., Miller, E. H., Milosovic, B., Milutinovic, S., Miuchi, K., Miyata, R., Mizrachi, E., Molinario, A., Monteiro, C. M. B., Monzani, M. E., Morå, K., Moriyama, S., Morrison, E., Morteau, E., Mosbacher, Y., Mount, B. J., Müller, J., Murdy, M., Murphy, A. St. J., Murra, M., Naylor, A., Nelson, H. N., Neves, F., Newstead, J. L., Nguyen, A., Ni, K., O'Hare, C., Oberlack, U., Obradovic, M., Olcina, I., Oliver-Mallory, K. C., Gann, G. D. Orebi, Orpwood, J., Ostrowskiy, I., Ouahada, S., Oyulmaz, K., Paetsch, B., Palladino, K. J., Palmer, J., Pan, Y., Pandurovic, M., Pannifer, N. J., Paramesvaran, S., Patton, S. J., Pellegrini, Q., Penning, B., Pereira, G., Peres, R., Perry, E., Pershing, T., Piastra, F., Pienaar, J., Piepke, A., Pierre, M., Plante, G., Pollmann, T. R., Principe, L., Qi, J., Qiao, K., Qie, Y., Qin, J., Radeka, S., Radeka, V., Rajado, M., García, D. Ramírez, Ravindran, A., Razeto, A., Reichenbacher, J., Rhyne, C. A., Richards, A., Rischbieter, G. R. C., Riyat, H. S., Rosero, R., Roy, A., Rushton, T., Rynders, D., Saakyan, R., Sanchez, L., Sanchez-Lucas, P., Santone, D., Santos, J. M. F. dos, Sartorelli, G., Sazzad, A. B. M. R., Scaffidi, A., Schnee, R. W., Schreiner, J., Schulte, P., Schulze, H., Eißing, Schumann, M., Schwenck, A., Schwenk, A., Lavina, L. Scotto, Selvi, M., Semeria, F., Shagin, P., Sharma, S., Shaw, S., Shen, W., Sherman, L., Shi, S., Shi, S. Y., Shimada, T., Shutt, T., Silk, J. J., Silva, C., Simgen, H., Sinev, G., Singh, R., Siniscalco, J., Solmaz, M., Solovov, V. N., Song, Z., Sorensen, P., Soria, J., Stanley, O., Steidl, M., Stenhouse, T., Stevens, A., Stifter, K., Sumner, T. J., Takeda, A., Tan, P. -L., Taylor, D. J., Taylor, W. C., Thers, D., Thümmler, T., Tiedt, D. R., Tönnies, F., Tong, Z., Toschi, F., Tovey, D. R., Tranter, J., Trask, M., Trinchero, G., Tripathi, M., Tronstad, D. R., Trotta, R., Tunnell, C. D., Urquijo, P., Usón, A., Utoyama, M., Vaitkus, A. C., Valentino, O., Valerius, K., Vecchi, S., Velan, V., Vetter, S., de Viveiros, L., Volta, G., Vorkapic, D., Wang, A., Wang, J. J., Wang, W., Wang, Y., Waters, D., Weerman, K. M., Weinheimer, C., Weiss, M., Wenz, D., Whitis, T. J., Wild, K., Williams, M., Wilson, M., Wilson, S. T., Wittweg, C., Wolf, J., Wolfs, F. L. H., Woodford, S., Woodward, D., Worcester, M., Wright, C. J., Wu, V. H. S., üstling, S. W, Wurm, M., Xia, Q., Xing, Y., Xu, D., Xu, J., Xu, Y., Xu, Z., Yamashita, M., Yang, L., Ye, J., Yeh, M., Yu, B., Zavattini, G., Zha, W., Zhong, M., and Zuber, K.
- Subjects
Physics - Instrumentation and Detectors ,High Energy Physics - Experiment ,Nuclear Experiment - Abstract
The XLZD collaboration is developing a two-phase xenon time projection chamber with an active mass of 60 to 80 t capable of probing the remaining WIMP-nucleon interaction parameter space down to the so-called neutrino fog. In this work we show that, based on the performance of currently operating detectors using the same technology and a realistic reduction of radioactivity in detector materials, such an experiment will also be able to competitively search for neutrinoless double beta decay in $^{136}$Xe using a natural-abundance xenon target. XLZD can reach a 3$\sigma$ discovery potential half-life of 5.7$\times$10$^{27}$ yr (and a 90% CL exclusion of 1.3$\times$10$^{28}$ yr) with 10 years of data taking, corresponding to a Majorana mass range of 7.3-31.3 meV (4.8-20.5 meV). XLZD will thus exclude the inverted neutrino mass ordering parameter space and will start to probe the normal ordering region for most of the nuclear matrix elements commonly considered by the community., Comment: 29 pages, 7 figures
- Published
- 2024
31. Thermodynamics of high order correction for Schwarzschild-AdS black hole in non-commutative geometry
- Author
-
Tan, Baoyu
- Subjects
General Relativity and Quantum Cosmology ,High Energy Physics - Theory - Abstract
Under the premise that quantum gravity becomes non-negligible, higher-order corrections of non-commutative geometry dominate. In this paper, we studied the thermodynamics of high-order corrections for Schwarzschild-AdS black hole with Lorentz distribution in the framework of non-commutative geometry. Our results indicate that when high-order corrections dominate, the thermodynamic behavior of Schwarzschild-AdS black hole in non-commutative geometry will gradually approach that of ordinary Schwarzschild-AdS black hole. In addition, we also studied the Joule-Thomson effect of Schwarzschild-AdS black hole under high-order corrections.
- Published
- 2024
32. Diffusion Priors for Variational Likelihood Estimation and Image Denoising
- Author
-
Cheng, Jun and Tan, Shan
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Real-world noise removal is crucial in low-level computer vision. Due to the remarkable generation capabilities of diffusion models, recent attention has shifted towards leveraging diffusion priors for image restoration tasks. However, existing diffusion priors-based methods either consider simple noise types or rely on approximate posterior estimation, limiting their effectiveness in addressing structured and signal-dependent noise commonly found in real-world images. In this paper, we build upon diffusion priors and propose adaptive likelihood estimation and MAP inference during the reverse diffusion process to tackle real-world noise. We introduce an independent, non-identically distributed likelihood combined with the noise precision (inverse variance) prior and dynamically infer the precision posterior using variational Bayes during the generation process. Meanwhile, we rectify the estimated noise variance through local Gaussian convolution. The final denoised image is obtained by propagating intermediate MAP solutions that balance the updated likelihood and diffusion prior. Additionally, we explore the local diffusion prior inherent in low-resolution diffusion models, enabling direct handling of high-resolution noisy images. Extensive experiments and analyses on diverse real-world datasets demonstrate the effectiveness of our method. Code is available at https://github.com/HUST-Tan/DiffusionVI., Comment: Accepted by NeurIPS2024 as Spotlight
- Published
- 2024
33. Literature Meets Data: A Synergistic Approach to Hypothesis Generation
- Author
-
Liu, Haokun, Zhou, Yangqiaoyu, Li, Mingxuan, Yuan, Chenfei, and Tan, Chenhao
- Subjects
Computer Science - Artificial Intelligence ,Computer Science - Computation and Language ,Computer Science - Computers and Society ,Computer Science - Machine Learning - Abstract
AI holds promise for transforming scientific processes, including hypothesis generation. Prior work on hypothesis generation can be broadly categorized into theory-driven and data-driven approaches. While both have proven effective in generating novel and plausible hypotheses, it remains an open question whether they can complement each other. To address this, we develop the first method that combines literature-based insights with data to perform LLM-powered hypothesis generation. We apply our method on five different datasets and demonstrate that integrating literature and data outperforms other baselines (8.97\% over few-shot, 15.75\% over literature-based alone, and 3.37\% over data-driven alone). Additionally, we conduct the first human evaluation to assess the utility of LLM-generated hypotheses in assisting human decision-making on two challenging tasks: deception detection and AI generated content detection. Our results show that human accuracy improves significantly by 7.44\% and 14.19\% on these tasks, respectively. These findings suggest that integrating literature-based and data-driven approaches provides a comprehensive and nuanced framework for hypothesis generation and could open new avenues for scientific inquiry., Comment: 30 pages, 7 figures, code link: https://github.com/ChicagoHAI/hypothesis-generation
- Published
- 2024
34. Altogether: Image Captioning via Re-aligning Alt-text
- Author
-
Xu, Hu, Huang, Po-Yao, Tan, Xiaoqing Ellen, Yeh, Ching-Feng, Kahn, Jacob, Jou, Christine, Ghosh, Gargi, Levy, Omer, Zettlemoyer, Luke, Yih, Wen-tau, Li, Shang-Wen, Xie, Saining, and Feichtenhofer, Christoph
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Computation and Language - Abstract
This paper focuses on creating synthetic data to improve the quality of image captions. Existing works typically have two shortcomings. First, they caption images from scratch, ignoring existing alt-text metadata, and second, lack transparency if the captioners' training data (e.g. GPT) is unknown. In this paper, we study a principled approach Altogether based on the key idea to edit and re-align existing alt-texts associated with the images. To generate training data, we perform human annotation where annotators start with the existing alt-text and re-align it to the image content in multiple rounds, consequently constructing captions with rich visual concepts. This differs from prior work that carries out human annotation as a one-time description task solely based on images and annotator knowledge. We train a captioner on this data that generalizes the process of re-aligning alt-texts at scale. Our results show our Altogether approach leads to richer image captions that also improve text-to-image generation and zero-shot image classification tasks., Comment: accepted by EMNLP 2024; MetaCLIPv2
- Published
- 2024
35. LVSM: A Large View Synthesis Model with Minimal 3D Inductive Bias
- Author
-
Jin, Haian, Jiang, Hanwen, Tan, Hao, Zhang, Kai, Bi, Sai, Zhang, Tianyuan, Luan, Fujun, Snavely, Noah, and Xu, Zexiang
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Graphics ,Computer Science - Machine Learning - Abstract
We propose the Large View Synthesis Model (LVSM), a novel transformer-based approach for scalable and generalizable novel view synthesis from sparse-view inputs. We introduce two architectures: (1) an encoder-decoder LVSM, which encodes input image tokens into a fixed number of 1D latent tokens, functioning as a fully learned scene representation, and decodes novel-view images from them; and (2) a decoder-only LVSM, which directly maps input images to novel-view outputs, completely eliminating intermediate scene representations. Both models bypass the 3D inductive biases used in previous methods -- from 3D representations (e.g., NeRF, 3DGS) to network designs (e.g., epipolar projections, plane sweeps) -- addressing novel view synthesis with a fully data-driven approach. While the encoder-decoder model offers faster inference due to its independent latent representation, the decoder-only LVSM achieves superior quality, scalability, and zero-shot generalization, outperforming previous state-of-the-art methods by 1.5 to 3.5 dB PSNR. Comprehensive evaluations across multiple datasets demonstrate that both LVSM variants achieve state-of-the-art novel view synthesis quality. Notably, our models surpass all previous methods even with reduced computational resources (1-2 GPUs). Please see our website for more details: https://haian-jin.github.io/projects/LVSM/ ., Comment: project page: https://haian-jin.github.io/projects/LVSM/
- Published
- 2024
36. Few-shot In-Context Preference Learning Using Large Language Models
- Author
-
Yu, Chao, Lu, Hong, Gao, Jiaxuan, Tan, Qixin, Yang, Xinting, Wang, Yu, Wu, Yi, and Vinitsky, Eugene
- Subjects
Computer Science - Artificial Intelligence ,Computer Science - Machine Learning - Abstract
Designing reward functions is a core component of reinforcement learning but can be challenging for truly complex behavior. Reinforcement Learning from Human Feedback (RLHF) has been used to alleviate this challenge by replacing a hand-coded reward function with a reward function learned from preferences. However, it can be exceedingly inefficient to learn these rewards as they are often learned tabula rasa. We investigate whether Large Language Models (LLMs) can reduce this query inefficiency by converting an iterative series of human preferences into code representing the rewards. We propose In-Context Preference Learning (ICPL), a method that uses the grounding of an LLM to accelerate learning reward functions from preferences. ICPL takes the environment context and task description, synthesizes a set of reward functions, and then repeatedly updates the reward functions using human rankings of videos of the resultant policies. Using synthetic preferences, we demonstrate that ICPL is orders of magnitude more efficient than RLHF and is even competitive with methods that use ground-truth reward functions instead of preferences. Finally, we perform a series of human preference-learning trials and observe that ICPL extends beyond synthetic settings and can work effectively with humans-in-the-loop. Additional information and videos are provided at https://sites.google.com/view/few-shot-icpl/home.
- Published
- 2024
37. VoiceBench: Benchmarking LLM-Based Voice Assistants
- Author
-
Chen, Yiming, Yue, Xianghu, Zhang, Chen, Gao, Xiaoxue, Tan, Robby T., and Li, Haizhou
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Sound ,Electrical Engineering and Systems Science - Audio and Speech Processing - Abstract
Building on the success of large language models (LLMs), recent advancements such as GPT-4o have enabled real-time speech interactions through LLM-based voice assistants, offering a significantly improved user experience compared to traditional text-based interactions. However, the absence of benchmarks designed to evaluate these speech interaction capabilities has hindered progress of LLM-based voice assistants development. Current evaluations focus primarily on automatic speech recognition (ASR) or general knowledge evaluation with clean speeches, neglecting the more intricate, real-world scenarios that involve diverse speaker characteristics, environmental and content factors. To address this, we introduce VoiceBench, the first benchmark designed to provide a multi-faceted evaluation of LLM-based voice assistants. VoiceBench also includes both real and synthetic spoken instructions that incorporate the above three key real-world variations. Extensive experiments reveal the limitations of current LLM-based voice assistant models and offer valuable insights for future research and development in this field., Comment: Work in progress. Data is available at https://github.com/MatthewCYM/VoiceBench
- Published
- 2024
38. Temporal and Spectral Analysis of the Unique and Second Brightest Gamma-Ray Burst GRB 230307A: Insights from GECAM and Fermi/GBM Observations
- Author
-
Moradi, R., Wang, C. W., Zhang, B., Wang, Y., Xiong, S. -L., Yi, S. -X., Tan, W. -J., Karlica, M., and Zhang, S. -N.
- Subjects
Astrophysics - High Energy Astrophysical Phenomena - Abstract
In this study, we present the pulse profile of the unique and the second brightest gamma-ray burst GRB 230307A, and analyze its temporal behavior using a joint GECAM--Fermi/GBM time-resolved spectral analysis. The utilization of GECAM data is advantageous as it successfully captured significant data during the pile-up period of the Fermi/GBM. We investigate the evolution of its flux, photon fluence, photon flux, peak energy, and the corresponding hardness-intensity and hardness-flux correlations. The findings within the first 27 seconds exhibit consistent patterns reported previously, providing valuable insights for comparing observations with predictions from the synchrotron radiation model invoking an expanding shell. Beyond the initial 27 seconds, we observe a notable transition in the emitted radiation, attributed to high latitude emission (HLE), influenced by the geometric properties of the shells and the relativistic Doppler effects. By modeling the data within the framework of the large-radius internal shock model, we discuss the required parameters as well as the limitations of the model. We conclude that a more complicated synchrotron emission model is needed to fully describe the observational data of GRB 230307A., Comment: Accepted for publication in The Astrophysical Journal (ApJ)
- Published
- 2024
39. The XLZD Design Book: Towards the Next-Generation Liquid Xenon Observatory for Dark Matter and Neutrino Physics
- Author
-
XLZD Collaboration, Aalbers, J., Abe, K., Adrover, M., Maouloud, S. Ahmed, Akerib, D. S., Musalhi, A. K. Al, Alder, F., Althueser, L., Amaral, D. W. P., Amarasinghe, C. S., Ames, A., Andrieu, B., Angelides, N., Angelino, E., Antunovic, B., Aprile, E., Araújo, H. M., Armstrong, J. E., Arthurs, M., Babicz, M., Bajpai, D., Baker, A., Balzer, M., Bang, J., Barberio, E., Bargemann, J. W., Barillier, E., Basharina-Freshville, A., Baudis, L., Bauer, D., Bazyk, M., Beattie, K., Beaupere, N., Bell, N. F., Bellagamba, L., Benson, T., Bhatti, A., Biesiadzinski, T. P., Biondi, R., Biondi, Y., Birch, H. J., Bishop, E., Bismark, A., Boehm, C., Boese, K., Bolotnikov, A., Brás, P., Braun, R., Breskin, A., Brew, C. A. J., Brommer, S., Brown, A., Bruni, G., Budnik, R., Burdin, S., Cai, C., Capelli, C., Carini, G., Carmona-Benitez, M. C., Carter, M., Chauvin, A., Chawla, A., Chen, H., Cherwinka, J. J., Chin, Y. T., Chott, N. I., Chavez, A. P. Cimental, Clark, K., Colijn, A. P., Colling, D. J., Conrad, J., Converse, M. V., Coronel, R., Costanzo, D., Cottle, A., Cox, G., Cuenca-García, J. J., Curran, D., Cussans, D., D'Andrea, V., Garcia, L. C. Daniel, Darlington, I., Dave, S., David, A., Davies, G. J., Decowski, M. P., Deisting, A., Delgaudio, J., Dey, S., Di Donato, C., Di Felice, L., Di Gangi, P., Diglio, S., Ding, C., Dobson, J. E. Y., Doerenkamp, M., Drexlin, G., Druszkiewicz, E., Dunbar, C. L., Eitel, K., Elykov, A., Engel, R., Eriksen, S. R., Fayer, S., Fearon, N. M., Ferella, A. D., Ferrari, C., Fieldhouse, N., Fischer, H., Flaecher, H., Flehmke, T., Flierman, M., Fraser, E. D., Fruth, T. M. A., Fujikawa, K., Fulgione, W., Fuselli, C., Gaemers, P., Gaior, R., Gaitskell, R. J., Gallice, N., Galloway, M., Gao, F., Garroum, N., Geffre, A., Genovesi, J., Ghag, C., Ghosh, S., Giacomobono, R., Gibbons, R., Girard, F., Glade-Beucke, R., Glück, F., Gokhale, S., Grandi, L., Green, J., Grigat, J., van der Grinten, M. G. D., Größle, R., Guan, H., Guida, M., Gyorgy, P., Haiston, J. J., Hall, C. R., Hall, T., Hammann, R., Hannen, V., Hansmann-Menzemer, S., Hargittai, N., Hartigan-O'Connor, E., Haselschwardt, S. J., Hernandez, M., Hertel, S. A., Higuera, A., Hils, C., Hiraoka, K., Hoetzsch, L., Hoferichter, M., Homenides, G. J., Hood, N. F., Horn, M., Huang, D. Q., Hughes, S., Hunt, D., Iacovacci, M., Itow, Y., Jacquet, E., Jakob, J., James, R. S., Joerg, F., Jones, S., Kaboth, A. C., Kahlert, F., Kamaha, A. C., Kaminaga, Y., Kara, M., Kavrigin, P., Kazama, S., Keller, M., Kemp-Russell, P., Khaitan, D., Kharbanda, P., Kilminster, B., Kim, J., Kirk, R., Kleifges, M., Klute, M., Kobayashi, M., Kodroff, D., Koke, D., Kopec, A., Korolkova, E. V., Kraus, H., Kravitz, S., Kreczko, L., von Krosigk, B., Kudryavtsev, V. A., Kuger, F., Kurita, N., Landsman, H., Lang, R. F., Lawes, C., Lee, J., Lehnert, B., Leonard, D. S., Lesko, K. T., Levinson, L., Li, A., Li, I., Li, S., Liang, S., Liang, Z., Lin, J., Lin, Y. -T., Lindemann, S., Linden, S., Lindner, M., Lindote, A., Lippincott, W. H., Liu, K., Loizeau, J., Lombardi, F., Lopes, J. A. M., Lopes, M. I., Lorenzon, W., Loutit, M., Lu, C., Lucchetti, G. M., Luce, T., Luitz, S., Ma, Y., Macolino, C., Mahlstedt, J., Maier, B., Majewski, P. A., Manalaysay, A., Mancuso, A., Manenti, L., Mannino, R. L., Marignetti, F., Marley, T., Undagoitia, T. Marrodán, Martens, K., Masbou, J., Masson, E., Mastroianni, S., Maupin, C., McCabe, C., McCarthy, M. E., McKinsey, D. N., McLaughlin, J. B., Melchiorre, A., Menéndez, J., Messina, M., Miller, E. H., Milosovic, B., Milutinovic, S., Miuchi, K., Miyata, R., Mizrachi, E., Molinario, A., Monteiro, C. M. B., Monzani, M. E., Morå, K., Moriyama, S., Morrison, E., Morteau, E., Mosbacher, Y., Mount, B. J., Müller, J., Murdy, M., Murphy, A. St. J., Murra, M., Naylor, A., Nelson, H. N., Neves, F., Newstead, J. L., Nguyen, A., Ni, K., O'Hare, C., Oberlack, U., Obradovic, M., Olcina, I., Oliver-Mallory, K. C., Gann, G. D. Orebi, Orpwood, J., Ostrowskiy, I., Ouahada, S., Oyulmaz, K., Paetsch, B., Palladino, K. J., Palmer, J., Pan, Y., Pandurovic, M., Pannifer, N. J., Paramesvaran, S., Patton, S. J., Pellegrini, Q., Penning, B., Pereira, G., Peres, R., Perry, E., Pershing, T., Piastra, F., Pienaar, J., Piepke, A., Pierre, M., Plante, G., Pollmann, T. R., Principe, L., Qi, J., Qiao, K., Qie, Y., Qin, J., Radeka, S., Radeka, V., Rajado, M., García, D. Ramírez, Ravindran, A., Razeto, A., Reichenbacher, J., Rhyne, C. A., Richards, A., Rischbieter, G. R. C., Riyat, H. S., Rosero, R., Roy, A., Rushton, T., Rynders, D., Saakyan, R., Sanchez, L., Sanchez-Lucas, P., Santone, D., Santos, J. M. F. dos, Sartorelli, G., Sazzad, A. B. M. R., Scaffidi, A., Schnee, R. W., Schreiner, J., Schulte, P., Schulze, H., Eißing, Schumann, M., Schwenck, A., Schwenk, A., Lavina, L. Scotto, Selvi, M., Semeria, F., Shagin, P., Sharma, S., Shaw, S., Shen, W., Sherman, L., Shi, S., Shi, S. Y., Shimada, T., Shutt, T., Silk, J. J., Silva, C., Simgen, H., Sinev, G., Singh, R., Siniscalco, J., Solmaz, M., Solovov, V. N., Song, Z., Sorensen, P., Soria, J., Stanley, O., Steidl, M., Stenhouse, T., Stevens, A., Stifter, K., Sumner, T. J., Takeda, A., Tan, P. -L., Taylor, D. J., Taylor, W. C., Thers, D., Thümmler, T., Tiedt, D. R., Tönnies, F., Tong, Z., Toschi, F., Tovey, D. R., Tranter, J., Trask, M., Trinchero, G., Tripathi, M., Tronstad, D. R., Trotta, R., Tunnell, C. D., Urquijo, P., Usón, A., Utoyama, M., Vaitkus, A. C., Valentino, O., Valerius, K., Vecchi, S., Velan, V., Vetter, S., de Viveiros, L., Volta, G., Vorkapic, D., Wang, A., Wang, J. J., Wang, W., Wang, Y., Waters, D., Weerman, K. M., Weinheimer, C., Weiss, M., Wenz, D., Whitis, T. J., Wild, K., Williams, M., Wilson, M., Wilson, S. T., Wittweg, C., Wolf, J., Wolfs, F. L. H., Woodford, S., Woodward, D., Worcester, M., Wright, C. J., Wu, V. H. S., üstling, S. W, Wurm, M., Xia, Q., Xing, Y., Xu, D., Xu, J., Xu, Y., Xu, Z., Yamashita, M., Yang, L., Ye, J., Yeh, M., Yu, B., Zavattini, G., Zha, W., Zhong, M., and Zuber, K.
- Subjects
High Energy Physics - Experiment ,High Energy Physics - Phenomenology ,Physics - Instrumentation and Detectors - Abstract
This report describes the experimental strategy and technologies for a next-generation xenon observatory sensitive to dark matter and neutrino physics. The detector will have an active liquid xenon target mass of 60-80 tonnes and is proposed by the XENON-LUX-ZEPLIN-DARWIN (XLZD) collaboration. The design is based on the mature liquid xenon time projection chamber technology of the current-generation experiments, LZ and XENONnT. A baseline design and opportunities for further optimization of the individual detector components are discussed. The experiment envisaged here has the capability to explore parameter space for Weakly Interacting Massive Particle (WIMP) dark matter down to the neutrino fog, with a 3$\sigma$ evidence potential for the spin-independent WIMP-nucleon cross sections as low as $3\times10^{-49}\rm cm^2$ (at 40 GeV/c$^2$ WIMP mass). The observatory is also projected to have a 3$\sigma$ observation potential of neutrinoless double-beta decay of $^{136}$Xe at a half-life of up to $5.7\times 10^{27}$ years. Additionally, it is sensitive to astrophysical neutrinos from the atmosphere, sun, and galactic supernovae., Comment: 32 pages, 14 figures
- Published
- 2024
40. Order Matters: Exploring Order Sensitivity in Multimodal Large Language Models
- Author
-
Tan, Zhijie, Chu, Xu, Li, Weiping, and Mo, Tong
- Subjects
Computer Science - Artificial Intelligence - Abstract
Multimodal Large Language Models (MLLMs) utilize multimodal contexts consisting of text, images, or videos to solve various multimodal tasks. However, we find that changing the order of multimodal input can cause the model's performance to fluctuate between advanced performance and random guessing. This phenomenon exists in both single-modality (text-only or image-only) and mixed-modality (image-text-pair) contexts. Furthermore, we demonstrate that popular MLLMs pay special attention to certain multimodal context positions, particularly the beginning and end. Leveraging this special attention, we place key video frames and important image/text content in special positions within the context and submit them to the MLLM for inference. This method results in average performance gains of 14.7% for video-caption matching and 17.8% for visual question answering tasks. Additionally, we propose a new metric, Position-Invariant Accuracy (PIA), to address order bias in MLLM evaluation. Our research findings contribute to a better understanding of Multi-Modal In-Context Learning (MMICL) and provide practical strategies for enhancing MLLM performance without increasing computational costs.
- Published
- 2024
41. AttentionPainter: An Efficient and Adaptive Stroke Predictor for Scene Painting
- Author
-
Tang, Yizhe, Wang, Yue, Hu, Teng, Yi, Ran, Tan, Xin, Ma, Lizhuang, Lai, Yu-Kun, and Rosin, Paul L.
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Stroke-based Rendering (SBR) aims to decompose an input image into a sequence of parameterized strokes, which can be rendered into a painting that resembles the input image. Recently, Neural Painting methods that utilize deep learning and reinforcement learning models to predict the stroke sequences have been developed, but suffer from longer inference time or unstable training. To address these issues, we propose AttentionPainter, an efficient and adaptive model for single-step neural painting. First, we propose a novel scalable stroke predictor, which predicts a large number of stroke parameters within a single forward process, instead of the iterative prediction of previous Reinforcement Learning or auto-regressive methods, which makes AttentionPainter faster than previous neural painting methods. To further increase the training efficiency, we propose a Fast Stroke Stacking algorithm, which brings 13 times acceleration for training. Moreover, we propose Stroke-density Loss, which encourages the model to use small strokes for detailed information, to help improve the reconstruction quality. Finally, we propose a new stroke diffusion model for both conditional and unconditional stroke-based generation, which denoises in the stroke parameter space and facilitates stroke-based inpainting and editing applications helpful for human artists design. Extensive experiments show that AttentionPainter outperforms the state-of-the-art neural painting methods.
- Published
- 2024
42. Modulations of Gravitational Waves due to Non-static Gravitational Lenses
- Author
-
Yang, Xing-Yu, Chen, Tan, and Cai, Rong-Gen
- Subjects
Astrophysics - Cosmology and Nongalactic Astrophysics ,General Relativity and Quantum Cosmology - Abstract
Gravitational waves (GWs) offer a new observational window into the universe, providing insights into compact objects and cosmic structures. Gravitational lensing, commonly studied in electromagnetic waves, also affects GWs, introducing magnification, time delays, and multiple images. While existing studies focus on static lenses, many astrophysical lenses are dynamic, with time-varying mass distributions such as moving stars or orbiting binaries. We develop a general theoretical framework to describe non-static lenses and demonstrate how they modulate GW signals, inducing unique time-varying amplitude modulations and spectral broadening. By examining uniformly moving and orbiting binary lenses, we show that these modulations provide new observational signatures, enhancing our understanding of lensing objects and the dynamics of the universe. Our findings have important implications for GW astronomy, offering novel ways to probe lens dynamics and improve the interpretations of GW signals., Comment: 20 pages, 8 figures
- Published
- 2024
43. Formalising CXL Cache Coherence
- Author
-
Tan, Chengsong, Donaldson, Alastair F., and Wickerson, John
- Subjects
Computer Science - Hardware Architecture ,Computer Science - Programming Languages - Abstract
We report our experience formally modelling and verifying CXL.cache, the inter-device cache coherence protocol of the Compute Express Link standard. We have used the Isabelle proof assistant to create a formal model for CXL.cache based on the prose English specification. This led to us identifying and proposing fixes to several problems we identified as unclear, ambiguous or inaccurate, some of which could lead to incoherence if left unfixed. Nearly all our issues and proposed fixes have been confirmed and tentatively accepted by the CXL consortium for adoption, save for one which is still under discussion. To validate the faithfulness of our model we performed scenario verification of essential restrictions such as "Snoop-pushes-GO", and produced a fully mechanised proof of a coherence property of the model. The considerable size of this proof, comprising tens of thousands of lemmas, prompted us to develop new proof automation tools, which we have made available for other Isabelle users working with similarly cumbersome proofs., Comment: 12 pages
- Published
- 2024
44. Acoustic Model Optimization over Multiple Data Sources: Merging and Valuation
- Author
-
Wei, Victor Junqiu, Wang, Weicheng, Jiang, Di, Tan, Conghui, and Lian, Rongzhong
- Subjects
Computer Science - Sound ,Computer Science - Computation and Language ,Electrical Engineering and Systems Science - Audio and Speech Processing - Abstract
Due to the rising awareness of privacy protection and the voluminous scale of speech data, it is becoming infeasible for Automatic Speech Recognition (ASR) system developers to train the acoustic model with complete data as before. For example, the data may be owned by different curators, and it is not allowed to share with others. In this paper, we propose a novel paradigm to solve salient problems plaguing the ASR field. In the first stage, multiple acoustic models are trained based upon different subsets of the complete speech data, while in the second phase, two novel algorithms are utilized to generate a high-quality acoustic model based upon those trained on data subsets. We first propose the Genetic Merge Algorithm (GMA), which is a highly specialized algorithm for optimizing acoustic models but suffers from low efficiency. We further propose the SGD-Based Optimizational Merge Algorithm (SOMA), which effectively alleviates the efficiency bottleneck of GMA and maintains superior model accuracy. Extensive experiments on public data show that the proposed methods can significantly outperform the state-of-the-art. Furthermore, we introduce Shapley Value to estimate the contribution score of the trained models, which is useful for evaluating the effectiveness of the data and providing fair incentives to their curators.
- Published
- 2024
45. CPE-Pro: A Structure-Sensitive Deep Learning Method for Protein Representation and Origin Evaluation
- Author
-
Gou, Wenrui, Ge, Wenhui, Tan, Yang, Li, Mingchen, Fan, Guisheng, and Yu, Huiqun
- Subjects
Quantitative Biology - Biomolecules ,Computer Science - Computation and Language ,Computer Science - Machine Learning ,Quantitative Biology - Quantitative Methods - Abstract
Protein structures are important for understanding their functions and interactions. Currently, many protein structure prediction methods are enriching the structure database. Discriminating the origin of structures is crucial for distinguishing between experimentally resolved and computationally predicted structures, evaluating the reliability of prediction methods, and guiding downstream biological studies. Building on works in structure prediction, We developed a structure-sensitive supervised deep learning model, Crystal vs Predicted Evaluator for Protein Structure (CPE-Pro), to represent and discriminate the origin of protein structures. CPE-Pro learns the structural information of proteins and captures inter-structural differences to achieve accurate traceability on four data classes, and is expected to be extended to more. Simultaneously, we utilized Foldseek to encode protein structures into "structure-sequences" and trained a protein Structural Sequence Language Model, SSLM. Preliminary experiments demonstrated that, compared to large-scale protein language models pre-trained on vast amounts of amino acid sequences, the "structure-sequence" enables the language model to learn more informative protein features, enhancing and optimizing structural representations. We have provided the code, model weights, and all related materials on https://github.com/GouWenrui/CPE-Pro-main.git.
- Published
- 2024
46. Gradient Rewiring for Editable Graph Neural Network Training
- Author
-
Jiang, Zhimeng, Liu, Zirui, Han, Xiaotian, Feng, Qizhang, Jin, Hongye, Tan, Qiaoyu, Zhou, Kaixiong, Zou, Na, and Hu, Xia
- Subjects
Computer Science - Machine Learning - Abstract
Deep neural networks are ubiquitously adopted in many applications, such as computer vision, natural language processing, and graph analytics. However, well-trained neural networks can make prediction errors after deployment as the world changes. \textit{Model editing} involves updating the base model to correct prediction errors with less accessible training data and computational resources. Despite recent advances in model editors in computer vision and natural language processing, editable training in graph neural networks (GNNs) is rarely explored. The challenge with editable GNN training lies in the inherent information aggregation across neighbors, which can lead model editors to affect the predictions of other nodes unintentionally. In this paper, we first observe the gradient of cross-entropy loss for the target node and training nodes with significant inconsistency, which indicates that directly fine-tuning the base model using the loss on the target node deteriorates the performance on training nodes. Motivated by the gradient inconsistency observation, we propose a simple yet effective \underline{G}radient \underline{R}ewiring method for \underline{E}ditable graph neural network training, named \textbf{GRE}. Specifically, we first store the anchor gradient of the loss on training nodes to preserve the locality. Subsequently, we rewire the gradient of the loss on the target node to preserve performance on the training node using anchor gradient. Experiments demonstrate the effectiveness of GRE on various model architectures and graph datasets in terms of multiple editing situations. The source code is available at \url{https://github.com/zhimengj0326/Gradient_rewiring_editing}, Comment: NeurIPS 2024
- Published
- 2024
47. Bayesian Concept Bottleneck Models with LLM Priors
- Author
-
Feng, Jean, Kothari, Avni, Zier, Luke, Singh, Chandan, and Tan, Yan Shuo
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Statistics - Machine Learning - Abstract
Concept Bottleneck Models (CBMs) have been proposed as a compromise between white-box and black-box models, aiming to achieve interpretability without sacrificing accuracy. The standard training procedure for CBMs is to predefine a candidate set of human-interpretable concepts, extract their values from the training data, and identify a sparse subset as inputs to a transparent prediction model. However, such approaches are often hampered by the tradeoff between enumerating a sufficiently large set of concepts to include those that are truly relevant versus controlling the cost of obtaining concept extractions. This work investigates a novel approach that sidesteps these challenges: BC-LLM iteratively searches over a potentially infinite set of concepts within a Bayesian framework, in which Large Language Models (LLMs) serve as both a concept extraction mechanism and prior. BC-LLM is broadly applicable and multi-modal. Despite imperfections in LLMs, we prove that BC-LLM can provide rigorous statistical inference and uncertainty quantification. In experiments, it outperforms comparator methods including black-box models, converges more rapidly towards relevant concepts and away from spuriously correlated ones, and is more robust to out-of-distribution samples.
- Published
- 2024
48. How Aligned are Generative Models to Humans in High-Stakes Decision-Making?
- Author
-
Tan, Sarah, Mallari, Keri, Adebayo, Julius, Gordo, Albert, Wells, Martin T., and Inkpen, Kori
- Subjects
Computer Science - Artificial Intelligence ,Computer Science - Machine Learning - Abstract
Large generative models (LMs) are increasingly being considered for high-stakes decision-making. This work considers how such models compare to humans and predictive AI models on a specific case of recidivism prediction. We combine three datasets -- COMPAS predictive AI risk scores, human recidivism judgements, and photos -- into a dataset on which we study the properties of several state-of-the-art, multimodal LMs. Beyond accuracy and bias, we focus on studying human-LM alignment on the task of recidivism prediction. We investigate if these models can be steered towards human decisions, the impact of adding photos, and whether anti-discimination prompting is effective. We find that LMs can be steered to outperform humans and COMPAS using in context-learning. We find anti-discrimination prompting to have unintended effects, causing some models to inhibit themselves and significantly reduce their number of positive predictions.
- Published
- 2024
49. Contextual Augmented Multi-Model Programming (CAMP): A Hybrid Local-Cloud Copilot Framework
- Author
-
Wang, Yuchen, Guo, Shangxin, and Tan, Chee Wei
- Subjects
Computer Science - Artificial Intelligence - Abstract
The advancements in cloud-based Large Languages Models (LLMs) have revolutionized AI-assisted programming. However, their integration into certain local development environments like ones within the Apple software ecosystem (e.g., iOS apps, macOS) remains challenging due to computational demands and sandboxed constraints. This paper presents CAMP, a multi-model AI-assisted programming framework that consists of a local model that employs Retrieval-Augmented Generation (RAG) to retrieve contextual information from the codebase to facilitate context-aware prompt construction thus optimizing the performance of the cloud model, empowering LLMs' capabilities in local Integrated Development Environments (IDEs). The methodology is actualized in Copilot for Xcode, an AI-assisted programming tool crafted for Xcode that employs the RAG module to address software constraints and enables diverse generative programming tasks, including automatic code completion, documentation, error detection, and intelligent user-agent interaction. The results from objective experiments on generated code quality and subjective experiments on user adoption collectively demonstrate the pilot success of the proposed system and mark its significant contributions to the realm of AI-assisted programming., Comment: 12 pages, 3 figures, 4 tables
- Published
- 2024
50. Stool Recognition for Colorectal Cancer Detection through Deep Learning
- Author
-
Tan, Glenda Hui En, Karin, Goh Xin Ru, and Bingquan, Shen
- Subjects
Electrical Engineering and Systems Science - Image and Video Processing ,Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Machine Learning - Abstract
Colorectal cancer is the most common cancer in Singapore and the third most common cancer worldwide. Blood in a person's stool is a symptom of this disease, and it is usually detected by the faecal occult blood test (FOBT). However, the FOBT presents several limitations - the collection process for the stool samples is tedious and unpleasant, the waiting period for results is about 2 weeks and costs are involved. In this research, we propose a simple-to-use, fast and cost-free alternative - a stool recognition neural network that determines if there is blood in one's stool (which indicates a possible risk of colorectal cancer) from an image of it. As this is a new classification task, there was limited data available, hindering classifier performance. Hence, various Generative Adversarial Networks (GANs) (DiffAugment StyleGAN2, DCGAN, Conditional GAN) were trained to generate images of high fidelity to supplement the dataset. Subsequently, images generated by the GAN with the most realistic images (DiffAugment StyleGAN2) were concatenated to the classifier's training batch on-the-fly, improving accuracy to 94%. This model was then deployed to a mobile app - Poolice, where users can take a photo of their stool and obtain instantaneous results if there is blood in their stool, prompting those who do to seek medical advice. As "early detection saves lives", we hope our app built on our stool recognition neural network can help people detect colorectal cancer earlier, so they can seek treatment and have higher chances of survival., Comment: 21 pages, 28 figures
- Published
- 2024
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.