1,219,241 results on '"CHAN, A"'
Search Results
2. The Impact of a Transnational Background on Family Migration Considerations Amid Political Uncertainty: Second-Generation Returnees in Hong Kong
- Author
-
Ngan, Lucille Lok-Sun, Chan, Anita Kit-Wa, Chan, Rami Hin-Yeung, and Siu, Queenie Kwan-Yee
- Published
- 2023
3. The Virtues of the Mass : A Taxonomy of a Late Middle English Genre of Liturgical Significance
- Author
-
Chan, Antje Elisa
- Published
- 2023
4. New Zealand, the Australian Commonwealth and ‘Plain Nonsense’
- Author
-
Chan, Adrian
- Published
- 2022
5. Exploring Writing Anxiety during Writing Process: An Analysis of Perceptions in Chinese English as a Foreign Language (EFL) Learners
- Author
-
Jing Sun, Saeid Motevalli, and Nee Nee Chan
- Abstract
Writing anxiety has been identified as a significant obstacle for English as a Foreign Language (EFL) learners in China, with previous studies indicating that it can negatively affect writing performance. Despite this, most research on writing anxiety in the Chinese EFL context has focused on the relationship between writing anxiety and writing performance, with limited attention paid to exploring writing anxiety during the writing process and its sources in depth. This study applied a qualitative method to explore Chinese EFL learners' writing anxiety in the writing process. Thematic analysis was used for analyzing data collected through semistructured interviews with 18 Chinese EFL learners. The results revealed seven primary themes, including lack of knowledge about the writing topic, inexperience with the genre, challenges with brainstorming or coming up with ideas, trouble with structuring or arranging information, difficulty with integrating sources, linguistic difficulty, and negative or no feedback from instructors. The findings indicated that writing anxiety accompanies throughout the writing process. EFL learners need strategies to alleviate writing anxiety, including clear instructions on how to approach writing tasks, provision of appropriate resources, individualized feedback, and a supportive learning environment.
- Published
- 2024
6. Chaucer’s Prayers: Writing Christian and Pagan Devotion by Megan E. Murton (review)
- Author
-
Chan, Antje Elisa
- Published
- 2022
7. Search for gravitational waves emitted from SN 2023ixf
- Author
-
The LIGO Scientific Collaboration, the Virgo Collaboration, the KAGRA Collaboration, Abac, A. G., Abbott, R., Abouelfettouh, I., Acernese, F., Ackley, K., Adhicary, S., Adhikari, N., Adhikari, R. X., Adkins, V. K., Agarwal, D., Agathos, M., Abchouyeh, M. Aghaei, Aguiar, O. D., Aguilar, I., Aiello, L., Ain, A., Akutsu, T., Albanesi, S., Alfaidi, R. A., Al-Jodah, A., Alléné, C., Allocca, A., Al-Shammari, S., Altin, P. A., Alvarez-Lopez, S., Amato, A., Amez-Droz, L., Amorosi, A., Amra, C., Ananyeva, A., Anderson, S. B., Anderson, W. G., Andia, M., Ando, M., Andrade, T., Andres, N., Andrés-Carcasona, M., Andrić, T., Anglin, J., Ansoldi, S., Antelis, J. M., Antier, S., Aoumi, M., Appavuravther, E. Z., Appert, S., Apple, S. K., Arai, K., Araya, A., Araya, M. C., Areeda, J. S., Argianas, L., Aritomi, N., Armato, F., Arnaud, N., Arogeti, M., Aronson, S. M., Ashton, G., Aso, Y., Assiduo, M., Melo, S. Assis de Souza, Aston, S. M., Astone, P., Attadio, F., Aubin, F., AultONeal, K., Avallone, G., Babak, S., Badaracco, F., Badger, C., Bae, S., Bagnasco, S., Bagui, E., Baier, J. G., Baiotti, L., Bajpai, R., Baka, T., Ball, M., Ballardin, G., Ballmer, S. W., Banagiri, S., Banerjee, B., Bankar, D., Baral, P., Barayoga, J. C., Barish, B. C., Barker, D., Barneo, P., Barone, F., Barr, B., Barsotti, L., Barsuglia, M., Barta, D., Bartoletti, A. M., Barton, M. A., Bartos, I., Basak, S., Basalaev, A., Bassiri, R., Basti, A., Bates, D. E., Bawaj, M., Baxi, P., Bayley, J. C., Baylor, A. C., Baynard II, P. A., Bazzan, M., Bedakihale, V. M., Beirnaert, F., Bejger, M., Belardinelli, D., Bell, A. S., Benedetto, V., Benoit, W., Bentley, J. D., Yaala, M. Ben, Bera, S., Berbel, M., Bergamin, F., Berger, B. K., Bernuzzi, S., Beroiz, M., Bersanetti, D., Bertolini, A., Betzwieser, J., Beveridge, D., Bevins, N., Bhandare, R., Bhardwaj, U., Bhatt, R., Bhattacharjee, D., Bhaumik, S., Bhowmick, S., Bianchi, A., Bilenko, I. A., Billingsley, G., Binetti, A., Bini, S., Birnholtz, O., Biscoveanu, S., Bisht, A., Bitossi, M., Bizouard, M. -A., Blackburn, J. K., Blagg, L. A., Blair, C. D., Blair, D. G., Bobba, F., Bode, N., Boileau, G., Boldrini, M., Bolingbroke, G. N., Bolliand, A., Bonavena, L. D., Bondarescu, R., Bondu, F., Bonilla, E., Bonilla, M. S., Bonino, A., Bonnand, R., Booker, P., Borchers, A., Boschi, V., Bose, S., Bossilkov, V., Boudart, V., Boudon, A., Bozzi, A., Bradaschia, C., Brady, P. R., Braglia, M., Branch, A., Branchesi, M., Brandt, J., Braun, I., Breschi, M., Briant, T., Brillet, A., Brinkmann, M., Brockill, P., Brockmueller, E., Brooks, A. F., Brown, B. C., Brown, D. D., Brozzetti, M. L., Brunett, S., Bruno, G., Bruntz, R., Bryant, J., Bucci, F., Buchanan, J., Bulashenko, O., Bulik, T., Bulten, H. J., Buonanno, A., Burtnyk, K., Buscicchio, R., Buskulic, D., Buy, C., Byer, R. L., Davies, G. S. Cabourn, Cabras, G., Cabrita, R., Cáceres-Barbosa, V., Cadonati, L., Cagnoli, G., Cahillane, C., Bustillo, J. Calderón, Callister, T. A., Calloni, E., Camp, J. B., Canepa, M., Santoro, G. Caneva, Cannon, K. C., Cao, H., Capistran, L. A., Capocasa, E., Capote, E., Carapella, G., Carbognani, F., Carlassara, M., Carlin, J. B., Carpinelli, M., Carrillo, G., Carter, J. J., Carullo, G., Diaz, J. Casanueva, Casentini, C., Castro-Lucas, S. Y., Caudill, S., Cavaglià, M., Cavalieri, R., Cella, G., Cerdá-Durán, P., Cesarini, E., Chaibi, W., Chakraborty, P., Subrahmanya, S. Chalathadka, Chan, J. C. L., Chan, M., Chandra, K., Chang, R. -J., Chao, S., Charlton, E. L., Charlton, P., Chassande-Mottin, E., Chatterjee, C., Chatterjee, Debarati, Chatterjee, Deep, Chaturvedi, M., Chaty, S., Chen, A., Chen, A. H. -Y., Chen, D., Chen, H., Chen, H. Y., Chen, J., Chen, K. H., Chen, Y., Chen, Yanbei, Chen, Yitian, Cheng, H. P., Chessa, P., Cheung, H. T., Cheung, S. Y., Chiadini, F., Chiarini, G., Chierici, R., Chincarini, A., Chiofalo, M. L., Chiummo, A., Chou, C., Choudhary, S., Christensen, N., Chua, S. S. Y., Chugh, P., Ciani, G., Ciecielag, P., Cieślar, M., Cifaldi, M., Ciolfi, R., Clara, F., Clark, J. A., Clarke, J., Clarke, T. A., Clearwater, P., Clesse, S., Coccia, E., Codazzo, E., Cohadon, P. -F., Colace, S., Colleoni, M., Collette, C. G., Collins, J., Colloms, S., Colombo, A., Colpi, M., Compton, C. M., Connolly, G., Conti, L., Corbitt, T. R., Cordero-Carrión, I., Corezzi, S., Cornish, N. J., Corsi, A., Cortese, S., Costa, C. A., Cottingham, R., Coughlin, M. W., Couineaux, A., Coulon, J. -P., Countryman, S. T., Coupechoux, J. -F., Couvares, P., Coward, D. M., Cowart, M. J., Coyne, R., Craig, K., Creed, R., Creighton, J. D. E., Creighton, T. D., Cremonese, P., Criswell, A. W., Crockett-Gray, J. C. G., Crook, S., Crouch, R., Csizmazia, J., Cudell, J. R., Cullen, T. J., Cumming, A., Cuoco, E., Cusinato, M., Dabadie, P., Canton, T. Dal, Dall'Osso, S., Pra, S. Dal, Dálya, G., D'Angelo, B., Danilishin, S., D'Antonio, S., Danzmann, K., Darroch, K. E., Dartez, L. P., Dasgupta, A., Datta, S., Dattilo, V., Daumas, A., Davari, N., Dave, I., Davenport, A., Davier, M., Davies, T. F., Davis, D., Davis, L., Davis, M. C., Davis, P. J., Dax, M., De Bolle, J., Deenadayalan, M., Degallaix, J., De Laurentis, M., Deléglise, S., De Lillo, F., Dell'Aquila, D., Del Pozzo, W., De Marco, F., De Matteis, F., D'Emilio, V., Demos, N., Dent, T., Depasse, A., DePergola, N., De Pietri, R., De Rosa, R., De Rossi, C., DeSalvo, R., De Simone, R., Dhani, A., Diab, R., Díaz, M. C., Di Cesare, M., Dideron, G., Didio, N. A., Dietrich, T., Di Fiore, L., Di Fronzo, C., Di Giovanni, M., Di Girolamo, T., Diksha, D., Di Michele, A., Ding, J., Di Pace, S., Di Palma, I., Di Renzo, F., Divyajyoti, Dmitriev, A., Doctor, Z., Dohmen, E., Doleva, P. P., Dominguez, D., D'Onofrio, L., Donovan, F., Dooley, K. L., Dooney, T., Doravari, S., Dorosh, O., Drago, M., Driggers, J. C., Ducoin, J. -G., Dunn, L., Dupletsa, U., D'Urso, D., Duval, H., Duverne, P. -A., Dwyer, S. E., Eassa, C., Ebersold, M., Eckhardt, T., Eddolls, G., Edelman, B., Edo, T. B., Edy, O., Effler, A., Eichholz, J., Einsle, H., Eisenmann, M., Eisenstein, R. A., Ejlli, A., Eleveld, R. M., Emma, M., Endo, K., Engl, A. J., Enloe, E., Errico, L., Essick, R. C., Estellés, H., Estevez, D., Etzel, T., Evans, M., Evstafyeva, T., Ewing, B. E., Ezquiaga, J. M., Fabrizi, F., Faedi, F., Fafone, V., Fairhurst, S., Farah, A. M., Farr, B., Farr, W. M., Favaro, G., Favata, M., Fays, M., Fazio, M., Feicht, J., Fejer, M. M., Felicetti, R., Fenyvesi, E., Ferguson, D. L., Ferraiuolo, S., Ferrante, I., Ferreira, T. A., Fidecaro, F., Figura, P., Fiori, A., Fiori, I., Fishbach, M., Fisher, R. P., Fittipaldi, R., Fiumara, V., Flaminio, R., Fleischer, S. M., Fleming, L. S., Floden, E., Foley, E. M., Fong, H., Font, J. A., Fornal, B., Forsyth, P. W. F., Franceschetti, K., Franchini, N., Frasca, S., Frasconi, F., Mascioli, A. Frattale, Frei, Z., Freise, A., Freitas, O., Frey, R., Frischhertz, W., Fritschel, P., Frolov, V. V., Fronzé, G. G., Fuentes-Garcia, M., Fujii, S., Fujimori, T., Fulda, P., Fyffe, M., Gadre, B., Gair, J. R., Galaudage, S., Galdi, V., Gallagher, H., Gallardo, S., Gallego, B., Gamba, R., Gamboa, A., Ganapathy, D., Ganguly, A., Garaventa, B., García-Bellido, J., Núñez, C. García, García-Quirós, C., Gardner, J. W., Gardner, K. A., Gargiulo, J., Garron, A., Garufi, F., Gasbarra, C., Gateley, B., Gayathri, V., Gemme, G., Gennai, A., Gennari, V., George, J., George, R., Gerberding, O., Gergely, L., Ghosh, Archisman, Ghosh, Sayantan, Ghosh, Shaon, Ghosh, Shrobana, Ghosh, Suprovo, Ghosh, Tathagata, Giacoppo, L., Giaime, J. A., Giardina, K. D., Gibson, D. R., Gibson, D. T., Gier, C., Giri, P., Gissi, F., Gkaitatzis, S., Glanzer, J., Glotin, F., Godfrey, J., Godwin, P., Goebbels, N. L., Goetz, E., Golomb, J., Lopez, S. Gomez, Goncharov, B., Gong, Y., González, G., Goodarzi, P., Goode, S., Goodwin-Jones, A. W., Gosselin, M., Göttel, A. S., Gouaty, R., Gould, D. W., Govorkova, K., Goyal, S., Grace, B., Grado, A., Graham, V., Granados, A. E., Granata, M., Granata, V., Gras, S., Grassia, P., Gray, A., Gray, C., Gray, R., Greco, G., Green, A. C., Green, S. M., Green, S. R., Gretarsson, A. M., Gretarsson, E. M., Griffith, D., Griffiths, W. L., Griggs, H. L., Grignani, G., Grimaldi, A., Grimaud, C., Grote, H., Guerra, D., Guetta, D., Guidi, G. M., Guimaraes, A. R., Gulati, H. K., Gulminelli, F., Gunny, A. M., Guo, H., Guo, W., Guo, Y., Gupta, Anchal, Gupta, Anuradha, Gupta, Ish, Gupta, N. C., Gupta, P., Gupta, S. K., Gupta, T., Gupte, N., Gurs, J., Gutierrez, N., Guzman, F., H, H. -Y., Haba, D., Haberland, M., Haino, S., Hall, E. D., Hamilton, E. Z., Hammond, G., Han, W. -B., Haney, M., Hanks, J., Hanna, C., Hannam, M. D., Hannuksela, O. A., Hanselman, A. G., Hansen, H., Hanson, J., Harada, R., Hardison, A. R., Haris, K., Harmark, T., Harms, J., Harry, G. M., Harry, I. W., Hart, J., Haskell, B., Haster, C. -J., Hathaway, J. S., Haughian, K., Hayakawa, H., Hayama, K., Hayes, R., Heffernan, A., Heidmann, A., Heintze, M. C., Heinze, J., Heinzel, J., Heitmann, H., Hellman, F., Hello, P., Helmling-Cornell, A. F., Hemming, G., Henderson-Sapir, O., Hendry, M., Heng, I. S., Hennes, E., Henshaw, C., Hertog, T., Heurs, M., Hewitt, A. L., Heyns, J., Higginbotham, S., Hild, S., Hill, S., Himemoto, Y., Hirata, N., Hirose, C., Hoang, S., Hochheim, S., Hofman, D., Holland, N. A., Holley-Bockelmann, K., Holmes, Z. J., Holz, D. E., Honet, L., Hong, C., Hornung, J., Hoshino, S., Hough, J., Hourihane, S., Howell, E. J., Hoy, C. G., Hrishikesh, C. A., Hsieh, H. -F., Hsiung, C., Hsu, H. C., Hsu, W. -F., Hu, P., Hu, Q., Huang, H. Y., Huang, Y. -J., Huddart, A. D., Hughey, B., Hui, D. C. Y., Hui, V., Husa, S., Huxford, R., Huynh-Dinh, T., Iampieri, L., Iandolo, G. A., Ianni, M., Iess, A., Imafuku, H., Inayoshi, K., Inoue, Y., Iorio, G., Iqbal, M. H., Irwin, J., Ishikawa, R., Isi, M., Ismail, M. A., Itoh, Y., Iwanaga, H., Iwaya, M., Iyer, B. R., JaberianHamedan, V., Jacquet, C., Jacquet, P. -E., Jadhav, S. J., Jadhav, S. P., Jain, T., James, A. L., James, P. A., Jamshidi, R., Janquart, J., Janssens, K., Janthalur, N. N., Jaraba, S., Jaranowski, P., Jaume, R., Javed, W., Jennings, A., Jia, W., Jiang, J., Kubisz, J., Johanson, C., Johns, G. R., Johnson, N. A., Johnston, M. C., Johnston, R., Johny, N., Jones, D. H., Jones, D. I., Jones, R., Jose, S., Joshi, P., Ju, L., Jung, K., Junker, J., Juste, V., Kajita, T., Kaku, I., Kalaghatgi, C., Kalogera, V., Kamiizumi, M., Kanda, N., Kandhasamy, S., Kang, G., Kanner, J. B., Kapadia, S. J., Kapasi, D. P., Karat, S., Karathanasis, C., Kashyap, R., Kasprzack, M., Kastaun, W., Kato, T., Katsavounidis, E., Katzman, W., Kaushik, R., Kawabe, K., Kawamoto, R., Kazemi, A., Keitel, D., Kelley-Derzon, J., Kennington, J., Kesharwani, R., Key, J. S., Khadela, R., Khadka, S., Khalili, F. Y., Khan, F., Khan, I., Khanam, T., Khursheed, M., Khusid, N. M., Kiendrebeogo, W., Kijbunchoo, N., Kim, C., Kim, J. C., Kim, K., Kim, M. H., Kim, S., Kim, Y. -M., Kimball, C., Kinley-Hanlon, M., Kinnear, M., Kissel, J. S., Klimenko, S., Knee, A. M., Knust, N., Kobayashi, K., Obergaulinger, M., Koch, P., Koehlenbeck, S. M., Koekoek, G., Kohri, K., Kokeyama, K., Koley, S., Kolitsidou, P., Kolstein, M., Komori, K., Kong, A. K. H., Kontos, A., Korobko, M., Kossak, R. V., Kou, X., Koushik, A., Kouvatsos, N., Kovalam, M., Kozak, D. B., Kranzhoff, S. L., Kringel, V., Krishnendu, N. V., Królak, A., Kruska, K., Kuehn, G., Kuijer, P., Kulkarni, S., Ramamohan, A. Kulur, Kumar, A., Kumar, Praveen, Kumar, Prayush, Kumar, Rahul, Kumar, Rakesh, Kume, J., Kuns, K., Kuntimaddi, N., Kuroyanagi, S., Kurth, N. J., Kuwahara, S., Kwak, K., Kwan, K., Kwok, J., Lacaille, G., Lagabbe, P., Laghi, D., Lai, S., Laity, A. H., Lakkis, M. H., Lalande, E., Lalleman, M., Lalremruati, P. C., Landry, M., Lane, B. B., Lang, R. N., Lange, J., Lantz, B., La Rana, A., La Rosa, I., Lartaux-Vollard, A., Lasky, P. D., Lawrence, J., Lawrence, M. N., Laxen, M., Lazzarini, A., Lazzaro, C., Leaci, P., Lecoeuche, Y. K., Lee, H. M., Lee, H. W., Lee, K., Lee, R. -K., Lee, R., Lee, S., Lee, Y., Legred, I. N., Lehmann, J., Lehner, L., Jean, M. Le, Lemaître, A., Lenti, M., Leonardi, M., Lequime, M., Leroy, N., Lesovsky, M., Letendre, N., Lethuillier, M., Levin, S. E., Levin, Y., Leyde, K., Li, A. K. Y., Li, K. L., Li, T. G. F., Li, X., Li, Z., Lihos, A., Lin, C-Y., Lin, C. -Y., Lin, E. T., Lin, F., Lin, H., Lin, L. C. -C., Lin, Y. -C., Linde, F., Linker, S. D., Littenberg, T. B., Liu, A., Liu, G. C., Liu, Jian, Villarreal, F. Llamas, Llobera-Querol, J., Lo, R. K. L., Locquet, J. -P., London, L. T., Longo, A., Lopez, D., Portilla, M. Lopez, Lorenzini, M., Lorenzo-Medina, A., Loriette, V., Lormand, M., Losurdo, G., Lott IV, T. P., Lough, J. D., Loughlin, H. A., Lousto, C. O., Lowry, M. J., Lu, N., Lück, H., Lumaca, D., Lundgren, A. P., Lussier, A. W., Ma, L. -T., Ma, S., Ma'arif, M., Macas, R., Macedo, A., MacInnis, M., Maciy, R. R., Macleod, D. M., MacMillan, I. A. O., Macquet, A., Macri, D., Maeda, K., Maenaut, S., Hernandez, I. Magaña, Magare, S. S., Magazzù, C., Magee, R. M., Maggio, E., Maggiore, R., Magnozzi, M., Mahesh, M., Mahesh, S., Maini, M., Majhi, S., Majorana, E., Makarem, C. N., Makelele, E., Malaquias-Reis, J. A., Mali, U., Maliakal, S., Malik, A., Man, N., Mandic, V., Mangano, V., Mannix, B., Mansell, G. L., Mansingh, G., Manske, M., Mantovani, M., Mapelli, M., Marchesoni, F., Pina, D. Marín, Marion, F., Márka, S., Márka, Z., Markosyan, A. S., Markowitz, A., Maros, E., Marsat, S., Martelli, F., Martin, I. W., Martin, R. M., Martinez, B. B., Martinez, M., Martinez, V., Martini, A., Martinovic, K., Martins, J. C., Martynov, D. V., Marx, E. J., Massaro, L., Masserot, A., Masso-Reid, M., Mastrodicasa, M., Mastrogiovanni, S., Matcovich, T., Matiushechkina, M., Matsuyama, M., Mavalvala, N., Maxwell, N., McCarrol, G., McCarthy, R., McClelland, D. E., McCormick, S., McCuller, L., McEachin, S., McElhenny, C., McGhee, G. I., McGinn, J., McGowan, K. B. M., McIver, J., McLeod, A., McRae, T., Meacher, D., Meijer, Q., Melatos, A., Mellaerts, S., Menendez-Vazquez, A., Menoni, C. S., Mera, F., Mercer, R. A., Mereni, L., Merfeld, K., Merilh, E. L., Mérou, J. R., Merritt, J. D., Merzougui, M., Messenger, C., Messick, C., Meyer-Conde, M., Meylahn, F., Mhaske, A., Miani, A., Miao, H., Michaloliakos, I., Michel, C., Michimura, Y., Middleton, H., Miller, A. L., Miller, S., Millhouse, M., Milotti, E., Milotti, V., Minenkov, Y., Mio, N., Mir, Ll. M., Mirasola, L., Miravet-Tenés, M., Miritescu, C. -A., Mishra, A. K., Mishra, A., Mishra, C., Mishra, T., Mitchell, A. L., Mitchell, J. G., Mitra, S., Mitrofanov, V. P., Mittleman, R., Miyakawa, O., Miyamoto, S., Miyoki, S., Mo, G., Mobilia, L., Mohapatra, S. R. P., Mohite, S. R., Molina-Ruiz, M., Mondal, C., Mondin, M., Montani, M., Moore, C. J., Moraru, D., More, A., More, S., Moreno, G., Morgan, C., Morisaki, S., Moriwaki, Y., Morras, G., Moscatello, A., Mourier, P., Mours, B., Mow-Lowry, C. M., Muciaccia, F., Mukherjee, Arunava, Mukherjee, D., Mukherjee, Samanwaya, Mukherjee, Soma, Mukherjee, Subroto, Mukherjee, Suvodip, Mukund, N., Mullavey, A., Munch, J., Mundi, J., Mungioli, C. L., Oberg, W. R. Munn, Murakami, Y., Murakoshi, M., Murray, P. G., Muusse, S., Nabari, D., Nadji, S. L., Nagar, A., Nagarajan, N., Nagler, K. N., Nakagaki, K., Nakamura, K., Nakano, H., Nakano, M., Nandi, D., Napolano, V., Narayan, P., Nardecchia, I., Narikawa, T., Narola, H., Naticchioni, L., Nayak, R. K., Neilson, J., Nelson, A., Nelson, T. J. N., Nery, M., Neunzert, A., Ng, S., Quynh, L. Nguyen, Nichols, S. A., Nielsen, A. B., Nieradka, G., Niko, A., Nishino, Y., Nishizawa, A., Nissanke, S., Nitoglia, E., Niu, W., Nocera, F., Norman, M., North, C., Novak, J., Siles, J. F. Nuño, Nuttall, L. K., Obayashi, K., Oberling, J., O'Dell, J., Oertel, M., Offermans, A., Oganesyan, G., Oh, J. J., Oh, K., O'Hanlon, T., Ohashi, M., Ohkawa, M., Ohme, F., Oliveira, A. S., Oliveri, R., O'Neal, B., Oohara, K., O'Reilly, B., Ormsby, N. D., Orselli, M., O'Shaughnessy, R., O'Shea, S., Oshima, Y., Oshino, S., Ossokine, S., Osthelder, C., Ota, I., Ottaway, D. J., Ouzriat, A., Overmier, H., Owen, B. J., Pace, A. E., Pagano, R., Page, M. A., Pai, A., Pal, A., Pal, S., Palaia, M. A., Pálfi, M., Palma, P. P., Palomba, C., Palud, P., Pan, H., Pan, J., Pan, K. C., Panai, R., Panda, P. K., Pandey, S., Panebianco, L., Pang, P. T. H., Pannarale, F., Pannone, K. A., Pant, B. C., Panther, F. H., Paoletti, F., Paolone, A., Papalexakis, E. E., Papalini, L., Papigkiotis, G., Paquis, A., Parisi, A., Park, B. -J., Park, J., Parker, W., Pascale, G., Pascucci, D., Pasqualetti, A., Passaquieti, R., Passenger, L., Passuello, D., Patane, O., Pathak, D., Pathak, M., Patra, A., Patricelli, B., Patron, A. S., Paul, K., Paul, S., Payne, E., Pearce, T., Pedraza, M., Pegna, R., Pele, A., Arellano, F. E. Peña, Penn, S., Penuliar, M. D., Perego, A., Pereira, Z., Perez, J. J., Périgois, C., Perna, G., Perreca, A., Perret, J., Perriès, S., Perry, J. W., Pesios, D., Petracca, S., Petrillo, C., Pfeiffer, H. P., Pham, H., Pham, K. A., Phukon, K. S., Phurailatpam, H., Piarulli, M., Piccari, L., Piccinni, O. J., Pichot, M., Piendibene, M., Piergiovanni, F., Pierini, L., Pierra, G., Pierro, V., Pietrzak, M., Pillas, M., Pilo, F., Pinard, L., Pinto, I. M., Pinto, M., Piotrzkowski, B. J., Pirello, M., Pitkin, M. D., Placidi, A., Placidi, E., Planas, M. L., Plastino, W., Poggiani, R., Polini, E., Pompili, L., Poon, J., Porcelli, E., Porter, E. K., Posnansky, C., Poulton, R., Powell, J., Pracchia, M., Pradhan, B. K., Pradier, T., Prajapati, A. K., Prasai, K., Prasanna, R., Prasia, P., Pratten, G., Principe, G., Principe, M., Prodi, G. A., Prokhorov, L., Prosposito, P., Puecher, A., Pullin, J., Punturo, M., Puppo, P., Pürrer, M., Qi, H., Qin, J., Quéméner, G., Quetschke, V., Quigley, C., Quinonez, P. J., Raab, F. J., Raabith, S. S., Raaijmakers, G., Raja, S., Rajan, C., Rajbhandari, B., Ramirez, K. E., Vidal, F. A. Ramis, Ramos-Buades, A., Rana, D., Ranjan, S., Ransom, K., Rapagnani, P., Ratto, B., Rawat, S., Ray, A., Raymond, V., Razzano, M., Read, J., Payo, M. Recaman, Regimbau, T., Rei, L., Reid, S., Reitze, D. H., Relton, P., Renzini, A. I., Rettegno, P., Revenu, B., Reyes, R., Rezaei, A. S., Ricci, F., Ricci, M., Ricciardone, A., Richardson, J. W., Richardson, M., Rijal, A., Riles, K., Riley, H. K., Rinaldi, S., Rittmeyer, J., Robertson, C., Robinet, F., Robinson, M., Rocchi, A., Rolland, L., Rollins, J. G., Romano, A. E., Romano, R., Romero, A., Romero-Shaw, I. M., Romie, J. H., Ronchini, S., Roocke, T. J., Rosa, L., Rosauer, T. J., Rose, C. A., Rosińska, D., Ross, M. P., Rossello, M., Rowan, S., Roy, S. K., Roy, S., Rozza, D., Ruggi, P., Ruhama, N., Morales, E. Ruiz, Ruiz-Rocha, K., Sachdev, S., Sadecki, T., Sadiq, J., Saffarieh, P., Sah, M. R., Saha, S. S., Saha, S., Sainrat, T., Menon, S. Sajith, Sakai, K., Sakellariadou, M., Sakon, S., Salafia, O. S., Salces-Carcoba, F., Salconi, L., Saleem, M., Salemi, F., Sallé, M., Salvador, S., Sanchez, A., Sanchez, E. J., Sanchez, J. H., Sanchez, L. E., Sanchis-Gual, N., Sanders, J. R., Sänger, E. M., Santoliquido, F., Saravanan, T. R., Sarin, N., Sasaoka, S., Sasli, A., Sassi, P., Sassolas, B., Satari, H., Sato, R., Sato, Y., Sauter, O., Savage, R. L., Sawada, T., Sawant, H. L., Sayah, S., Scacco, V., Schaetzl, D., Scheel, M., Schiebelbein, A., Schiworski, M. G., Schmidt, P., Schmidt, S., Schnabel, R., Schneewind, M., Schofield, R. M. S., Schouteden, K., Schulte, B. W., Schutz, B. F., Schwartz, E., Scialpi, M., Scott, J., Scott, S. M., Seetharamu, T. C., Seglar-Arroyo, M., Sekiguchi, Y., Sellers, D., Sengupta, A. S., Sentenac, D., Seo, E. G., Seo, J. W., Sequino, V., Serra, M., Servignat, G., Sevrin, A., Shaffer, T., Shah, U. S., Shaikh, M. A., Shao, L., Sharma, A. K., Sharma, P., Sharma-Chaudhary, S., Shaw, M. R., Shawhan, P., Shcheblanov, N. S., Sheridan, E., Shikano, Y., Shikauchi, M., Shimode, K., Shinkai, H., Shiota, J., Shoemaker, D. H., Shoemaker, D. M., Short, R. W., ShyamSundar, S., Sider, A., Siegel, H., Sieniawska, M., Sigg, D., Silenzi, L., Simmonds, M., Singer, L. P., Singh, A., Singh, D., Singh, M. K., Singh, S., Singha, A., Sintes, A. M., Sipala, V., Skliris, V., Slagmolen, B. J. J., Slaven-Blair, T. J., Smetana, J., Smith, J. R., Smith, L., Smith, R. J. E., Smith, W. J., Soldateschi, J., Somiya, K., Song, I., Soni, K., Soni, S., Sordini, V., Sorrentino, F., Sorrentino, N., Sotani, H., Soulard, R., Southgate, A., Spagnuolo, V., Spencer, A. P., Spera, M., Spinicelli, P., Spoon, J. B., Sprague, C. A., Srivastava, A. K., Stachurski, F., Steer, D. A., Steinlechner, J., Steinlechner, S., Stergioulas, N., Stevens, P., StPierre, M., Stratta, G., Strong, M. D., Strunk, A., Sturani, R., Stuver, A. L., Suchenek, M., Sudhagar, S., Sueltmann, N., Suleiman, L., Sullivan, K. D., Sun, L., Sunil, S., Suresh, J., Sutton, P. J., Suzuki, T., Suzuki, Y., Swinkels, B. L., Syx, A., Szczepańczyk, M. J., Szewczyk, P., Tacca, M., Tagoshi, H., Tait, S. C., Takahashi, H., Takahashi, R., Takamori, A., Takase, T., Takatani, K., Takeda, H., Takeshita, K., Talbot, C., Tamaki, M., Tamanini, N., Tanabe, D., Tanaka, K., Tanaka, S. J., Tanaka, T., Tang, D., Tanioka, S., Tanner, D. B., Tao, L., Tapia, R. D., Martín, E. N. Tapia San, Tarafder, R., Taranto, C., Taruya, A., Tasson, J. D., Teloi, M., Tenorio, R., Themann, H., Theodoropoulos, A., Thirugnanasambandam, M. P., Thomas, L. M., Thomas, M., Thomas, P., Thompson, J. E., Thondapu, S. R., Thorne, K. A., Thrane, E., Tissino, J., Tiwari, A., Tiwari, P., Tiwari, S., Tiwari, V., Todd, M. R., Toivonen, A. M., Toland, K., Tolley, A. E., Tomaru, T., Tomita, K., Tomura, T., Tong-Yu, C., Toriyama, A., Toropov, N., Torres-Forné, A., Torrie, C. I., Toscani, M., Melo, I. Tosta e, Tournefier, E., Trapananti, A., Travasso, F., Traylor, G., Trevor, M., Tringali, M. C., Tripathee, A., Troian, G., Troiano, L., Trovato, A., Trozzo, L., Trudeau, R. J., Tsang, T. T. L., Tso, R., Tsuchida, S., Tsukada, L., Tsutsui, T., Turbang, K., Turconi, M., Turski, C., Ubach, H., Uchikata, N., Uchiyama, T., Udall, R. P., Uehara, T., Uematsu, M., Ueno, K., Ueno, S., Undheim, V., Ushiba, T., Vacatello, M., Vahlbruch, H., Vaidya, N., Vajente, G., Vajpeyi, A., Valdes, G., Valencia, J., Valentini, M., Vallejo-Peña, S. A., Vallero, S., Valsan, V., van Bakel, N., van Beuzekom, M., van Dael, M., Brand, J. F. J. van den, Broeck, C. Van Den, Vander-Hyde, D. C., van der Sluys, M., Van de Walle, A., van Dongen, J., Vandra, K., van Haevermaet, H., van Heijningen, J. V., Van Hove, P., VanKeuren, M., Vanosky, J., van Putten, M. H. P. M., van Ranst, Z., van Remortel, N., Vardaro, M., Vargas, A. F., Varghese, J. J., Varma, V., Vasúth, M., Vecchio, A., Vedovato, G., Veitch, J., Veitch, P. J., Venikoudis, S., Venneberg, J., Verdier, P., Verkindt, D., Verma, B., Verma, P., Verma, Y., Vermeulen, S. M., Vetrano, F., Veutro, A., Vibhute, A. M., Viceré, A., Vidyant, S., Viets, A. D., Vijaykumar, A., Vilkha, A., Villa-Ortega, V., Vincent, E. T., Vinet, J. -Y., Viret, S., Virtuoso, A., Vitale, S., Vives, A., Vocca, H., Voigt, D., von Reis, E. R. G., von Wrangel, J. S. A., Vyatchanin, S. P., Wade, L. E., Wade, M., Wagner, K. J., Wajid, A., Walker, M., Wallace, G. S., Wallace, L., Wang, H., Wang, J. Z., Wang, W. H., Wang, Z., Waratkar, G., Warner, J., Was, M., Washimi, T., Washington, N. Y., Watarai, D., Wayt, K. E., Weaver, B. R., Weaver, B., Weaving, C. R., Webster, S. A., Weinert, M., Weinstein, A. J., Weiss, R., Wellmann, F., Wen, L., Weßels, P., Wette, K., Whelan, J. T., Whiting, B. F., Whittle, C., Wildberger, J. B., Wilk, O. S., Wilken, D., Wilkin, A. T., Willadsen, D. J., Willetts, K., Williams, D., Williams, M. J., Williams, N. S., Willis, J. L., Willke, B., Wils, M., Winterflood, J., Wipf, C. C., Woan, G., Woehler, J., Wofford, J. K., Wolfe, N. E., Wong, H. T., Wong, H. W. Y., Wong, I. C. F., Wright, J. L., Wright, M., Wu, C., Wu, D. S., Wu, H., Wuchner, E., Wysocki, D. M., Xu, V. A., Xu, Y., Yadav, N., Yamamoto, H., Yamamoto, K., Yamamoto, T. S., Yamamoto, T., Yamamura, S., Yamazaki, R., Yan, S., Yan, T., Yang, F. W., Yang, F., Yang, K. Z., Yang, Y., Yarbrough, Z., Yasui, H., Yeh, S. -W., Yelikar, A. B., Yin, X., Yokoyama, J., Yokozawa, T., Yoo, J., Yu, H., Yuan, S., Yuzurihara, H., Zadrożny, A., Zanolin, M., Zeeshan, M., Zelenova, T., Zendri, J. -P., Zeoli, M., Zerrad, M., Zevin, M., Zhang, A. C., Zhang, L., Zhang, R., Zhang, T., Zhang, Y., Zhao, C., Zhao, Yue, Zhao, Yuhang, Zheng, Y., Zhong, H., Zhou, R., Zhu, X. -J., Zhu, Z. -H., Zimmerman, A. B., Zucker, M. E., and Zweizig, J.
- Subjects
Astrophysics - High Energy Astrophysical Phenomena - Abstract
We present the results of a search for gravitational-wave transients associated with core-collapse supernova SN 2023ixf, which was observed in the galaxy Messier 101 via optical emission on 2023 May 19th, during the LIGO-Virgo-KAGRA 15th Engineering Run. We define a five-day on-source window during which an accompanying gravitational-wave signal may have occurred. No gravitational waves have been identified in data when at least two gravitational-wave observatories were operating, which covered $\sim 14\%$ of this five-day window. We report the search detection efficiency for various possible gravitational-wave emission models. Considering the distance to M101 (6.7 Mpc), we derive constraints on the gravitational-wave emission mechanism of core-collapse supernovae across a broad frequency spectrum, ranging from 50 Hz to 2 kHz where we assume the GW emission occurred when coincident data are available in the on-source window. Considering an ellipsoid model for a rotating proto-neutron star, our search is sensitive to gravitational-wave energy $1 \times 10^{-5} M_{\odot} c^2$ and luminosity $4 \times 10^{-5} M_{\odot} c^2/\text{s}$ for a source emitting at 50 Hz. These constraints are around an order of magnitude more stringent than those obtained so far with gravitational-wave data. The constraint on the ellipticity of the proto-neutron star that is formed is as low as $1.04$, at frequencies above $1200$ Hz, surpassing results from SN 2019ejj., Comment: Main paper: 6 pages, 4 figures and 1 table. Total with appendices: 20 pages, 4 figures, and 1 table
- Published
- 2024
8. A search using GEO600 for gravitational waves coincident with fast radio bursts from SGR 1935+2154
- Author
-
The LIGO Scientific Collaboration, the Virgo Collaboration, the KAGRA Collaboration, Abac, A. G., Abbott, R., Abouelfettouh, I., Acernese, F., Ackley, K., Adhicary, S., Adhikari, N., Adhikari, R. X., Adkins, V. K., Agarwal, D., Agathos, M., Abchouyeh, M. Aghaei, Aguiar, O. D., Aguilar, I., Aiello, L., Ain, A., Ajith, P., Akutsu, T., Albanesi, S., Alfaidi, R. A., Al-Jodah, A., Alléné, C., Allocca, A., Al-Shammari, S., Altin, P. A., Alvarez-Lopez, S., Amato, A., Amez-Droz, L., Amorosi, A., Amra, C., Ananyeva, A., Anderson, S. B., Anderson, W. G., Andia, M., Ando, M., Andrade, T., Andres, N., Andrés-Carcasona, M., Andrić, T., Anglin, J., Ansoldi, S., Antelis, J. M., Antier, S., Aoumi, M., Appavuravther, E. Z., Appert, S., Apple, S. K., Arai, K., Araya, A., Araya, M. C., Areeda, J. S., Argianas, L., Aritomi, N., Armato, F., Arnaud, N., Arogeti, M., Aronson, S. M., Ashton, G., Aso, Y., Assiduo, M., Melo, S. Assis de Souza, Aston, S. M., Astone, P., Attadio, F., Aubin, F., AultONeal, K., Avallone, G., Azrad, D., Babak, S., Badaracco, F., Badger, C., Bae, S., Bagnasco, S., Bagui, E., Baier, J. G., Baiotti, L., Bajpai, R., Baka, T., Ball, M., Ballardin, G., Ballmer, S. W., Banagiri, S., Banerjee, B., Bankar, D., Baral, P., Barayoga, J. C., Barish, B. C., Barker, D., Barneo, P., Barone, F., Barr, B., Barsotti, L., Barsuglia, M., Barta, D., Bartoletti, A. M., Barton, M. A., Bartos, I., Basak, S., Basalaev, A., Bassiri, R., Basti, A., Bates, D. E., Bawaj, M., Baxi, P., Bayley, J. C., Baylor, A. C., Baynard II, P. A., Bazzan, M., Bedakihale, V. M., Beirnaert, F., Bejger, M., Belardinelli, D., Bell, A. S., Benedetto, V., Benoit, W., Bentley, J. D., Yaala, M. Ben, Bera, S., Berbel, M., Bergamin, F., Berger, B. K., Bernuzzi, S., Beroiz, M., Bersanetti, D., Bertolini, A., Betzwieser, J., Beveridge, D., Bevins, N., Bhandare, R., Bhardwaj, U., Bhatt, R., Bhattacharjee, D., Bhaumik, S., Bhowmick, S., Bianchi, A., Bilenko, I. A., Billingsley, G., Binetti, A., Bini, S., Birnholtz, O., Biscoveanu, S., Bisht, A., Bitossi, M., Bizouard, M. -A., Blackburn, J. K., Blagg, L. A., Blair, C. D., Blair, D. G., Bobba, F., Bode, N., Boileau, G., Boldrini, M., Bolingbroke, G. N., Bolliand, A., Bonavena, L. D., Bondarescu, R., Bondu, F., Bonilla, E., Bonilla, M. S., Bonino, A., Bonnand, R., Booker, P., Borchers, A., Boschi, V., Bose, S., Bossilkov, V., Boudart, V., Boudon, A., Bozzi, A., Bradaschia, C., Brady, P. R., Braglia, M., Branch, A., Branchesi, M., Brandt, J., Braun, I., Breschi, M., Briant, T., Brillet, A., Brinkmann, M., Brockill, P., Brockmueller, E., Brooks, A. F., Brown, B. C., Brown, D. D., Brozzetti, M. L., Brunett, S., Bruno, G., Bruntz, R., Bryant, J., Bucci, F., Buchanan, J., Bulashenko, O., Bulik, T., Bulten, H. J., Buonanno, A., Burtnyk, K., Buscicchio, R., Buskulic, D., Buy, C., Byer, R. L., Davies, G. S. Cabourn, Cabras, G., Cabrita, R., Cáceres-Barbosa, V., Cadonati, L., Cagnoli, G., Cahillane, C., Bustillo, J. Calderón, Callister, T. A., Calloni, E., Camp, J. B., Canepa, M., Santoro, G. Caneva, Cannon, K. C., Cao, H., Capistran, L. A., Capocasa, E., Capote, E., Carapella, G., Carbognani, F., Carlassara, M., Carlin, J. B., Carpinelli, M., Carrillo, G., Carter, J. J., Carullo, G., Diaz, J. Casanueva, Casentini, C., Castro-Lucas, S. Y., Caudill, S., Cavaglià, M., Cavalieri, R., Cella, G., Cerdá-Durán, P., Cesarini, E., Chaibi, W., Chakraborty, P., Subrahmanya, S. Chalathadka, Chan, J. C. L., Chan, M., Chandra, K., Chang, R. -J., Chao, S., Charlton, E. L., Charlton, P., Chassande-Mottin, E., Chatterjee, C., Chatterjee, Debarati, Chatterjee, Deep, Chaturvedi, M., Chaty, S., Chen, A., Chen, A. H. -Y., Chen, D., Chen, H., Chen, H. Y., Chen, J., Chen, K. H., Chen, Y., Chen, Yanbei, Chen, Yitian, Cheng, H. P., Chessa, P., Cheung, H. T., Cheung, S. Y., Chiadini, F., Chiarini, G., Chierici, R., Chincarini, A., Chiofalo, M. L., Chiummo, A., Chou, C., Choudhary, S., Christensen, N., Chua, S. S. Y., Chugh, P., Ciani, G., Ciecielag, P., Cieślar, M., Cifaldi, M., Ciolfi, R., Clara, F., Clark, J. A., Clarke, J., Clarke, T. A., Clearwater, P., Clesse, S., Coccia, E., Codazzo, E., Cohadon, P. -F., Colace, S., Colleoni, M., Collette, C. G., Collins, J., Colloms, S., Colombo, A., Colpi, M., Compton, C. M., Connolly, G., Conti, L., Corbitt, T. R., Cordero-Carrión, I., Corezzi, S., Cornish, N. J., Corsi, A., Cortese, S., Costa, C. A., Cottingham, R., Coughlin, M. W., Couineaux, A., Coulon, J. -P., Countryman, S. T., Coupechoux, J. -F., Couvares, P., Coward, D. M., Cowart, M. J., Coyne, R., Craig, K., Creed, R., Creighton, J. D. E., Creighton, T. D., Cremonese, P., Criswell, A. W., Crockett-Gray, J. C. G., Crook, S., Crouch, R., Csizmazia, J., Cudell, J. R., Cullen, T. J., Cumming, A., Cuoco, E., Cusinato, M., Dabadie, P., Canton, T. Dal, Dall'Osso, S., Pra, S. Dal, Dálya, G., D'Angelo, B., Danilishin, S., D'Antonio, S., Danzmann, K., Darroch, K. E., Dartez, L. P., Dasgupta, A., Datta, S., Dattilo, V., Daumas, A., Davari, N., Dave, I., Davenport, A., Davier, M., Davies, T. F., Davis, D., Davis, L., Davis, M. C., Davis, P. J., Dax, M., De Bolle, J., Deenadayalan, M., Degallaix, J., De Laurentis, M., Deléglise, S., De Lillo, F., Dell'Aquila, D., Del Pozzo, W., De Marco, F., De Matteis, F., D'Emilio, V., Demos, N., Dent, T., Depasse, A., DePergola, N., De Pietri, R., De Rosa, R., De Rossi, C., DeSalvo, R., De Simone, R., Dhani, A., Diab, R., Díaz, M. C., Di Cesare, M., Dideron, G., Didio, N. A., Dietrich, T., Di Fiore, L., Di Fronzo, C., Di Giovanni, M., Di Girolamo, T., Diksha, D., Di Michele, A., Ding, J., Di Pace, S., Di Palma, I., Di Renzo, F., Divyajyoti, Dmitriev, A., Doctor, Z., Dohmen, E., Doleva, P. P., Dominguez, D., D'Onofrio, L., Donovan, F., Dooley, K. L., Dooney, T., Doravari, S., Dorosh, O., Drago, M., Driggers, J. C., Ducoin, J. -G., Dunn, L., Dupletsa, U., D'Urso, D., Duval, H., Duverne, P. -A., Dwyer, S. E., Eassa, C., Ebersold, M., Eckhardt, T., Eddolls, G., Edelman, B., Edo, T. B., Edy, O., Effler, A., Eichholz, J., Einsle, H., Eisenmann, M., Eisenstein, R. A., Ejlli, A., Eleveld, R. M., Emma, M., Endo, K., Engl, A. J., Enloe, E., Errico, L., Essick, R. C., Estellés, H., Estevez, D., Etzel, T., Evans, M., Evstafyeva, T., Ewing, B. E., Ezquiaga, J. M., Fabrizi, F., Faedi, F., Fafone, V., Fairhurst, S., Farah, A. M., Farr, B., Farr, W. M., Favaro, G., Favata, M., Fays, M., Fazio, M., Feicht, J., Fejer, M. M., Felicetti, R. ., Fenyvesi, E., Ferguson, D. L., Ferraiuolo, S., Ferrante, I., Ferreira, T. A., Fidecaro, F., Figura, P., Fiori, A., Fiori, I., Fishbach, M., Fisher, R. P., Fittipaldi, R., Fiumara, V., Flaminio, R., Fleischer, S. M., Fleming, L. S., Floden, E., Foley, E. M., Fong, H., Font, J. A., Fornal, B., Forsyth, P. W. F., Franceschetti, K., Franchini, N., Frasca, S., Frasconi, F., Mascioli, A. Frattale, Frei, Z., Freise, A., Freitas, O., Frey, R., Frischhertz, W., Fritschel, P., Frolov, V. V., Fronzé, G. G., Fuentes-Garcia, M., Fujii, S., Fujimori, T., Fulda, P., Fyffe, M., Gadre, B., Gair, J. R., Galaudage, S., Galdi, V., Gallagher, H., Gallardo, S., Gallego, B., Gamba, R., Gamboa, A., Ganapathy, D., Ganguly, A., Garaventa, B., García-Bellido, J., Núñez, C. García, García-Quirós, C., Gardner, J. W., Gardner, K. A., Gargiulo, J., Garron, A., Garufi, F., Gasbarra, C., Gateley, B., Gayathri, V., Gemme, G., Gennai, A., Gennari, V., George, J., George, R., Gerberding, O., Gergely, L., Ghonge, S., Ghosh, Archisman, Ghosh, Sayantan, Ghosh, Shaon, Ghosh, Shrobana, Ghosh, Suprovo, Ghosh, Tathagata, Giacoppo, L., Giaime, J. A., Giardina, K. D., Gibson, D. R., Gibson, D. T., Gier, C., Giri, P., Gissi, F., Gkaitatzis, S., Glanzer, J., Glotin, F., Godfrey, J., Godwin, P., Goebbels, N. L., Goetz, E., Golomb, J., Lopez, S. Gomez, Goncharov, B., Gong, Y., González, G., Goodarzi, P., Goode, S., Goodwin-Jones, A. W., Gosselin, M., Göttel, A. S., Gouaty, R., Gould, D. W., Govorkova, K., Goyal, S., Grace, B., Grado, A., Graham, V., Granados, A. E., Granata, M., Granata, V., Gras, S., Grassia, P., Gray, A., Gray, C., Gray, R., Greco, G., Green, A. C., Green, S. M., Green, S. R., Gretarsson, A. M., Gretarsson, E. M., Griffith, D., Griffiths, W. L., Griggs, H. L., Grignani, G., Grimaldi, A., Grimaud, C., Grote, H., Guerra, D., Guetta, D., Guidi, G. M., Guimaraes, A. R., Gulati, H. K., Gulminelli, F., Gunny, A. M., Guo, H., Guo, W., Guo, Y., Gupta, Anchal, Gupta, Anuradha, Gupta, Ish, Gupta, N. C., Gupta, P., Gupta, S. K., Gupta, T., Gupte, N., Gurs, J., Gutierrez, N., Guzman, F., H, H. -Y., Haba, D., Haberland, M., Haino, S., Hall, E. D., Hamilton, E. Z., Hammond, G., Han, W. -B., Haney, M., Hanks, J., Hanna, C., Hannam, M. D., Hannuksela, O. A., Hanselman, A. G., Hansen, H., Hanson, J., Harada, R., Hardison, A. R., Haris, K., Harmark, T., Harms, J., Harry, G. M., Harry, I. W., Hart, J., Haskell, B., Haster, C. -J., Hathaway, J. S., Haughian, K., Hayakawa, H., Hayama, K., Hayes, R., Heffernan, A., Heidmann, A., Heintze, M. C., Heinze, J., Heinzel, J., Heitmann, H., Hellman, F., Hello, P., Helmling-Cornell, A. F., Hemming, G., Henderson-Sapir, O., Hendry, M., Heng, I. S., Hennes, E., Henshaw, C., Hertog, T., Heurs, M., Hewitt, A. L., Heyns, J., Higginbotham, S., Hild, S., Hill, S., Himemoto, Y., Hirata, N., Hirose, C., Ho, W. C. G., Hoang, S., Hochheim, S., Hofman, D., Holland, N. A., Holley-Bockelmann, K., Holmes, Z. J., Holz, D. E., Honet, L., Hong, C., Hornung, J., Hoshino, S., Hough, J., Hourihane, S., Howell, E. J., Hoy, C. G., Hrishikesh, C. A., Hsieh, H. -F., Hsiung, C., Hsu, H. C., Hsu, W. -F., Hu, P., Hu, Q., Huang, H. Y., Huang, Y. -J., Huddart, A. D., Hughey, B., Hui, D. C. Y., Hui, V., Husa, S., Huxford, R., Huynh-Dinh, T., Iampieri, L., Iandolo, G. A., Ianni, M., Iess, A., Imafuku, H., Inayoshi, K., Inoue, Y., Iorio, G., Iqbal, M. H., Irwin, J., Ishikawa, R., Isi, M., Ismail, M. A., Itoh, Y., Iwanaga, H., Iwaya, M., Iyer, B. R., JaberianHamedan, V., Jacquet, C., Jacquet, P. -E., Jadhav, S. J., Jadhav, S. P., Jain, T., James, A. L., James, P. A., Jamshidi, R., Janquart, J., Janssens, K., Janthalur, N. N., Jaraba, S., Jaranowski, P., Jaume, R., Javed, W., Jennings, A., Jia, W., Jiang, J., Kubisz, J., Johanson, C., Johns, G. R., Johnson, N. A., Johnston, M. C., Johnston, R., Johny, N., Jones, D. H., Jones, D. I., Jones, R., Jose, S., Joshi, P., Ju, L., Jung, K., Junker, J., Juste, V., Kajita, T., Kaku, I., Kalaghatgi, C., Kalogera, V., Kamiizumi, M., Kanda, N., Kandhasamy, S., Kang, G., Kanner, J. B., Kapadia, S. J., Kapasi, D. P., Karat, S., Karathanasis, C., Kashyap, R., Kasprzack, M., Kastaun, W., Kato, T., Katsavounidis, E., Katzman, W., Kaushik, R., Kawabe, K., Kawamoto, R., Kazemi, A., Keitel, D., Kelley-Derzon, J., Kennington, J., Kesharwani, R., Key, J. S., Khadela, R., Khadka, S., Khalili, F. Y., Khan, F., Khan, I., Khanam, T., Khursheed, M., Khusid, N. M., Kiendrebeogo, W., Kijbunchoo, N., Kim, C., Kim, J. C., Kim, K., Kim, M. H., Kim, S., Kim, Y. -M., Kimball, C., Kinley-Hanlon, M., Kinnear, M., Kissel, J. S., Klimenko, S., Knee, A. M., Knust, N., Kobayashi, K., Koch, P., Koehlenbeck, S. M., Koekoek, G., Kohri, K., Kokeyama, K., Koley, S., Kolitsidou, P., Kolstein, M., Komori, K., Kong, A. K. H., Kontos, A., Korobko, M., Kossak, R. V., Kou, X., Koushik, A., Kouvatsos, N., Kovalam, M., Kozak, D. B., Kranzhoff, S. L., Kringel, V., Krishnendu, N. V., Królak, A., Kruska, K., Kuehn, G., Kuijer, P., Kulkarni, S., Ramamohan, A. Kulur, Kumar, A., Kumar, Praveen, Kumar, Prayush, Kumar, Rahul, Kumar, Rakesh, Kume, J., Kuns, K., Kuntimaddi, N., Kuroyanagi, S., Kurth, N. J., Kuwahara, S., Kwak, K., Kwan, K., Kwok, J., Lacaille, G., Lagabbe, P., Laghi, D., Lai, S., Laity, A. H., Lakkis, M. H., Lalande, E., Lalleman, M., Lalremruati, P. C., Landry, M., Lane, B. B., Lang, R. N., Lange, J., Lantz, B., La Rana, A., La Rosa, I., Lartaux-Vollard, A., Lasky, P. D., Lawrence, J., Lawrence, M. N., Laxen, M., Lazzarini, A., Lazzaro, C., Leaci, P., Lecoeuche, Y. K., Lee, H. M., Lee, H. W., Lee, K., Lee, R. -K., Lee, R., Lee, S., Lee, Y., Legred, I. N., Lehmann, J., Lehner, L., Jean, M. Le, Lemaître, A., Lenti, M., Leonardi, M., Lequime, M., Leroy, N., Lesovsky, M., Letendre, N., Lethuillier, M., Levin, S. E., Levin, Y., Leyde, K., Li, A. K. Y., Li, K. L., Li, T. G. F., Li, X., Li, Z., Lihos, A., Lin, C-Y., Lin, C. -Y., Lin, E. T., Lin, F., Lin, H., Lin, L. C. -C., Lin, Y. -C., Linde, F., Linker, S. D., Littenberg, T. B., Liu, A., Liu, G. C., Liu, Jian, Villarreal, F. Llamas, Llobera-Querol, J., Lo, R. K. L., Locquet, J. -P., London, L. T., Longo, A., Lopez, D., Portilla, M. Lopez, Lorenzini, M., Lorenzo-Medina, A., Loriette, V., Lormand, M., Losurdo, G., Lott IV, T. P., Lough, J. D., Loughlin, H. A., Lousto, C. O., Lowry, M. J., Lu, N., Lück, H., Lumaca, D., Lundgren, A. P., Lussier, A. W., Ma, L. -T., Ma, S., Ma'arif, M., Macas, R., Macedo, A., MacInnis, M., Maciy, R. R., Macleod, D. M., MacMillan, I. A. O., Macquet, A., Macri, D., Maeda, K., Maenaut, S., Hernandez, I. Magaña, Magare, S. S., Magazzù, C., Magee, R. M., Maggio, E., Maggiore, R., Magnozzi, M., Mahesh, M., Mahesh, S., Maini, M., Majhi, S., Majorana, E., Makarem, C. N., Makelele, E., Malaquias-Reis, J. A., Mali, U., Maliakal, S., Malik, A., Man, N., Mandic, V., Mangano, V., Mannix, B., Mansell, G. L., Mansingh, G., Manske, M., Mantovani, M., Mapelli, M., Marchesoni, F., Pina, D. Marín, Marion, F., Márka, S., Márka, Z., Markosyan, A. S., Markowitz, A., Maros, E., Marsat, S., Martelli, F., Martin, I. W., Martin, R. M., Martinez, B. B., Martinez, M., Martinez, V., Martini, A., Martinovic, K., Martins, J. C., Martynov, D. V., Marx, E. J., Massaro, L., Masserot, A., Masso-Reid, M., Mastrodicasa, M., Mastrogiovanni, S., Matcovich, T., Matiushechkina, M., Matsuyama, M., Mavalvala, N., Maxwell, N., McCarrol, G., McCarthy, R., McCormick, S., McCuller, L., McEachin, S., McElhenny, C., McGhee, G. I., McGinn, J., McGowan, K. B. M., McIver, J., McLeod, A., McRae, T., Meacher, D., Meijer, Q., Melatos, A., Mellaerts, S., Menendez-Vazquez, A., Menoni, C. S., Mera, F., Mercer, R. A., Mereni, L., Merfeld, K., Merilh, E. L., Mérou, J. R., Merritt, J. D., Merzougui, M., Messenger, C., Messick, C., Meyer-Conde, M., Meylahn, F., Mhaske, A., Miani, A., Miao, H., Michaloliakos, I., Michel, C., Michimura, Y., Middleton, H., Miller, A. L., Miller, S., Millhouse, M., Milotti, E., Milotti, V., Minenkov, Y., Mio, N., Mir, Ll. M., Mirasola, L., Miravet-Tenés, M., Miritescu, C. -A., Mishra, A. K., Mishra, A., Mishra, C., Mishra, T., Mitchell, A. L., Mitchell, J. G., Mitra, S., Mitrofanov, V. P., Mittleman, R., Miyakawa, O., Miyamoto, S., Miyoki, S., Mo, G., Mobilia, L., Mohapatra, S. R. P., Mohite, S. R., Molina-Ruiz, M., Mondal, C., Mondin, M., Montani, M., Moore, C. J., Moraru, D., More, A., More, S., Moreno, G., Morgan, C., Morisaki, S., Moriwaki, Y., Morras, G., Moscatello, A., Mourier, P., Mours, B., Mow-Lowry, C. M., Muciaccia, F., Mukherjee, Arunava, Mukherjee, D., Mukherjee, Samanwaya, Mukherjee, Soma, Mukherjee, Subroto, Mukherjee, Suvodip, Mukund, N., Mullavey, A., Munch, J., Mundi, J., Mungioli, C. L., Oberg, W. R. Munn, Murakami, Y., Murakoshi, M., Murray, P. G., Muusse, S., Nabari, D., Nadji, S. L., Nagar, A., Nagarajan, N., Nagler, K. N., Nakagaki, K., Nakamura, K., Nakano, H., Nakano, M., Nandi, D., Napolano, V., Narayan, P., Nardecchia, I., Narola, H., Naticchioni, L., Nayak, R. K., Neilson, J., Nelson, A., Nelson, T. J. N., Nery, M., Neunzert, A., Ng, S., Quynh, L. Nguyen, Nichols, S. A., Nielsen, A. B., Nieradka, G., Niko, A., Nishino, Y., Nishizawa, A., Nissanke, S., Nitoglia, E., Niu, W., Nocera, F., Norman, M., North, C., Novak, J., Siles, J. F. Nuño, Nuttall, L. K., Obayashi, K., Oberling, J., O'Dell, J., Oertel, M., Offermans, A., Oganesyan, G., Oh, J. J., Oh, K., O'Hanlon, T., Ohashi, M., Ohkawa, M., Ohme, F., Oliveira, A. S., Oliveri, R., O'Neal, B., Oohara, K., O'Reilly, B., Ormsby, N. D., Orselli, M., O'Shaughnessy, R., O'Shea, S., Oshima, Y., Oshino, S., Ossokine, S., Osthelder, C., Ota, I., Ottaway, D. J., Ouzriat, A., Overmier, H., Owen, B. J., Pace, A. E., Pagano, R., Page, M. A., Pai, A., Pal, A., Pal, S., Palaia, M. A., Pálfi, M., Palma, P. P., Palomba, C., Palud, P., Pan, H., Pan, J., Pan, K. C., Panai, R., Panda, P. K., Pandey, S., Panebianco, L., Pang, P. T. H., Pannarale, F., Pannone, K. A., Pant, B. C., Panther, F. H., Paoletti, F., Paolone, A., Papalexakis, E. E., Papalini, L., Papigkiotis, G., Paquis, A., Parisi, A., Park, B. -J., Park, J., Parker, W., Pascale, G., Pascucci, D., Pasqualetti, A., Passaquieti, R., Passenger, L., Passuello, D., Patane, O., Pathak, D., Pathak, M., Patra, A., Patricelli, B., Patron, A. S., Paul, K., Paul, S., Payne, E., Pearce, T., Pedraza, M., Pegna, R., Pele, A., Arellano, F. E. Peña, Penn, S., Penuliar, M. D., Perego, A., Pereira, Z., Perez, J. J., Périgois, C., Perna, G., Perreca, A., Perret, J., Perriès, S., Perry, J. W., Pesios, D., Petracca, S., Petrillo, C., Pfeiffer, H. P., Pham, H., Pham, K. A., Phukon, K. S., Phurailatpam, H., Piarulli, M., Piccari, L., Piccinni, O. J., Pichot, M., Piendibene, M., Piergiovanni, F., Pierini, L., Pierra, G., Pierro, V., Pietrzak, M., Pillas, M., Pilo, F., Pinard, L., Pinto, I. M., Pinto, M., Piotrzkowski, B. J., Pirello, M., Pitkin, M. D., Placidi, A., Placidi, E., Planas, M. L., Plastino, W., Poggiani, R., Polini, E., Pompili, L., Poon, J., Porcelli, E., Porter, E. K., Posnansky, C., Poulton, R., Powell, J., Pracchia, M., Pradhan, B. K., Pradier, T., Prajapati, A. K., Prasai, K., Prasanna, R., Prasia, P., Pratten, G., Principe, G., Principe, M., Prodi, G. A., Prokhorov, L., Prosposito, P., Puecher, A., Pullin, J., Punturo, M., Puppo, P., Pürrer, M., Qi, H., Qin, J., Quéméner, G., Quetschke, V., Quigley, C., Quinonez, P. J., Quitzow-James, R., Raab, F. J., Raabith, S. S., Raaijmakers, G., Raja, S., Rajan, C., Rajbhandari, B., Ramirez, K. E., Vidal, F. A. Ramis, Ramos-Buades, A., Rana, D., Ranjan, S., Ransom, K., Rapagnani, P., Ratto, B., Rawat, S., Ray, A., Raymond, V., Razzano, M., Read, J., Payo, M. Recaman, Regimbau, T., Rei, L., Reid, S., Reitze, D. H., Relton, P., Renzini, A. I., Rettegno, P., Revenu, B., Reyes, R., Rezaei, A. S., Ricci, F., Ricci, M., Ricciardone, A., Richardson, J. W., Richardson, M., Rijal, A., Riles, K., Riley, H. K., Rinaldi, S., Rittmeyer, J., Robertson, C., Robinet, F., Robinson, M., Rocchi, A., Rolland, L., Rollins, J. G., Romano, A. E., Romano, R., Romero, A., Romero-Shaw, I. M., Romie, J. H., Ronchini, S., Roocke, T. J., Rosa, L., Rosauer, T. J., Rose, C. A., Rosińska, D., Ross, M. P., Rossello, M., Rowan, S., Roy, S. K., Roy, S., Rozza, D., Ruggi, P., Ruhama, N., Morales, E. Ruiz, Ruiz-Rocha, K., Sachdev, S., Sadecki, T., Sadiq, J., Saffarieh, P., Sah, M. R., Saha, S. S., Saha, S., Sainrat, T., Menon, S. Sajith, Sakai, K., Sakellariadou, M., Sakon, S., Salafia, O. S., Salces-Carcoba, F., Salconi, L., Saleem, M., Salemi, F., Sallé, M., Salvador, S., Sanchez, A., Sanchez, E. J., Sanchez, J. H., Sanchez, L. E., Sanchis-Gual, N., Sanders, J. R., Sänger, E. M., Santoliquido, F., Saravanan, T. R., Sarin, N., Sasaoka, S., Sasli, A., Sassi, P., Sassolas, B., Satari, H., Sato, R., Sato, Y., Sauter, O., Savage, R. L., Sawada, T., Sawant, H. L., Sayah, S., Scacco, V., Schaetzl, D., Scheel, M., Schiebelbein, A., Schiworski, M. G., Schmidt, P., Schmidt, S., Schnabel, R., Schneewind, M., Schofield, R. M. S., Schouteden, K., Schulte, B. W., Schutz, B. F., Schwartz, E., Scialpi, M., Scott, J., Scott, S. M., Seetharamu, T. C., Seglar-Arroyo, M., Sekiguchi, Y., Sellers, D., Sengupta, A. S., Sentenac, D., Seo, E. G., Seo, J. W., Sequino, V., Serra, M., Servignat, G., Sevrin, A., Shaffer, T., Shah, U. S., Shaikh, M. A., Shao, L., Sharma, A. K., Sharma, P., Sharma-Chaudhary, S., Shaw, M. R., Shawhan, P., Shcheblanov, N. S., Sheridan, E., Shikano, Y., Shikauchi, M., Shimode, K., Shinkai, H., Shiota, J., Shoemaker, D. H., Shoemaker, D. M., Short, R. W., ShyamSundar, S., Sider, A., Siegel, H., Sieniawska, M., Sigg, D., Silenzi, L., Simmonds, M., Singer, L. P., Singh, A., Singh, D., Singh, M. K., Singh, S., Singha, A., Sintes, A. M., Sipala, V., Skliris, V., Slagmolen, B. J. J., Slaven-Blair, T. J., Smetana, J., Smith, J. R., Smith, L., Smith, R. J. E., Smith, W. J., Soldateschi, J., Somiya, K., Song, I., Soni, K., Soni, S., Sordini, V., Sorrentino, F., Sorrentino, N., Sotani, H., Soulard, R., Southgate, A., Spagnuolo, V., Spencer, A. P., Spera, M., Spinicelli, P., Spoon, J. B., Sprague, C. A., Srivastava, A. K., Stachurski, F., Steer, D. A., Steinlechner, J., Steinlechner, S., Stergioulas, N., Stevens, P., StPierre, M., Stratta, G., Strong, M. D., Strunk, A., Sturani, R., Stuver, A. L., Suchenek, M., Sudhagar, S., Sueltmann, N., Suleiman, L., Sullivan, K. D., Sun, L., Sunil, S., Suresh, J., Sutton, P. J., Suzuki, T., Suzuki, Y., Swinkels, B. L., Syx, A., Szczepańczyk, M. J., Szewczyk, P., Tacca, M., Tagoshi, H., Tait, S. C., Takahashi, H., Takahashi, R., Takamori, A., Takase, T., Takatani, K., Takeda, H., Takeshita, K., Talbot, C., Tamaki, M., Tamanini, N., Tanabe, D., Tanaka, K., Tanaka, S. J., Tanaka, T., Tang, D., Tanioka, S., Tanner, D. B., Tao, L., Tapia, R. D., Martín, E. N. Tapia San, Tarafder, R., Taranto, C., Taruya, A., Tasson, J. D., Teloi, M., Tenorio, R., Themann, H., Theodoropoulos, A., Thirugnanasambandam, M. P., Thomas, L. M., Thomas, M., Thomas, P., Thompson, J. E., Thondapu, S. R., Thorne, K. A., Thrane, E., Tissino, J., Tiwari, A., Tiwari, P., Tiwari, S., Tiwari, V., Todd, M. R., Toivonen, A. M., Toland, K., Tolley, A. E., Tomaru, T., Tomita, K., Tomura, T., Tong-Yu, C., Toriyama, A., Toropov, N., Torres-Forné, A., Torrie, C. I., Toscani, M., Melo, I. Tosta e, Tournefier, E., Trapananti, A., Travasso, F., Traylor, G., Trevor, M., Tringali, M. C., Tripathee, A., Troian, G., Troiano, L., Trovato, A., Trozzo, L., Trudeau, R. J., Tsang, T. T. L., Tso, R., Tsuchida, S., Tsukada, L., Tsutsui, T., Turbang, K., Turconi, M., Turski, C., Ubach, H., Uchiyama, T., Udall, R. P., Uehara, T., Uematsu, M., Ueno, K., Ueno, S., Undheim, V., Ushiba, T., Vacatello, M., Vahlbruch, H., Vaidya, N., Vajente, G., Vajpeyi, A., Valdes, G., Valencia, J., Valentini, M., Vallejo-Peña, S. A., Vallero, S., Valsan, V., van Bakel, N., van Beuzekom, M., van Dael, M., Brand, J. F. J. van den, Broeck, C. Van Den, Vander-Hyde, D. C., van der Sluys, M., Van de Walle, A., van Dongen, J., Vandra, K., van Haevermaet, H., van Heijningen, J. V., Van Hove, P., VanKeuren, M., Vanosky, J., van Putten, M. H. P. M., van Ranst, Z., van Remortel, N., Vardaro, M., Vargas, A. F., Varghese, J. J., Varma, V., Vasúth, M., Vecchio, A., Vedovato, G., Veitch, J., Veitch, P. J., Venikoudis, S., Venneberg, J., Verdier, P., Verkindt, D., Verma, B., Verma, P., Verma, Y., Vermeulen, S. M., Vetrano, F., Veutro, A., Vibhute, A. M., Viceré, A., Vidyant, S., Viets, A. D., Vijaykumar, A., Vilkha, A., Villa-Ortega, V., Vincent, E. T., Vinet, J. -Y., Viret, S., Virtuoso, A., Vitale, S., Vives, A., Vocca, H., Voigt, D., von Reis, E. R. G., von Wrangel, J. S. A., Vyatchanin, S. P., Wade, L. E., Wade, M., Wagner, K. J., Wajid, A., Walker, M., Wallace, G. S., Wallace, L., Wang, H., Wang, J. Z., Wang, W. H., Wang, Z., Waratkar, G., Warner, J., Was, M., Washimi, T., Washington, N. Y., Watarai, D., Wayt, K. E., Weaver, B. R., Weaver, B., Weaving, C. R., Webster, S. A., Weinert, M., Weinstein, A. J., Weiss, R., Wellmann, F., Wen, L., Weßels, P., Wette, K., Whelan, J. T., Whiting, B. F., Whittle, C., Wildberger, J. B., Wilk, O. S., Wilken, D., Wilkin, A. T., Willadsen, D. J., Willetts, K., Williams, D., Williams, M. J., Williams, N. S., Willis, J. L., Willke, B., Wils, M., Winterflood, J., Wipf, C. C., Woan, G., Woehler, J., Wofford, J. K., Wolfe, N. E., Wong, H. T., Wong, H. W. Y., Wong, I. C. F., Wright, J. L., Wright, M., Wu, C., Wu, D. S., Wu, H., Wuchner, E., Wysocki, D. M., Xu, V. A., Xu, Y., Yadav, N., Yamamoto, H., Yamamoto, K., Yamamoto, T. S., Yamamoto, T., Yamamura, S., Yamazaki, R., Yan, S., Yan, T., Yang, F. W., Yang, F., Yang, K. Z., Yang, Y., Yarbrough, Z., Yasui, H., Yeh, S. -W., Yelikar, A. B., Yin, X., Yokoyama, J., Yokozawa, T., Yoo, J., Yu, H., Yuan, S., Yuzurihara, H., Zadrożny, A., Zanolin, M., Zeeshan, M., Zelenova, T., Zendri, J. -P., Zeoli, M., Zerrad, M., Zevin, M., Zhang, A. C., Zhang, L., Zhang, R., Zhang, T., Zhang, Y., Zhao, C., Zhao, Yue, Zhao, Yuhang, Zheng, Y., Zhong, H., Zhou, R., Zhu, X. -J., Zhu, Z. -H., Zucker, M. E., and Zweizig, J.
- Subjects
Astrophysics - High Energy Astrophysical Phenomena - Abstract
The magnetar SGR 1935+2154 is the only known Galactic source of fast radio bursts (FRBs). FRBs from SGR 1935+2154 were first detected by CHIME/FRB and STARE2 in 2020 April, after the conclusion of the LIGO, Virgo, and KAGRA Collaborations' O3 observing run. Here we analyze four periods of gravitational wave (GW) data from the GEO600 detector coincident with four periods of FRB activity detected by CHIME/FRB, as well as X-ray glitches and X-ray bursts detected by NICER and NuSTAR close to the time of one of the FRBs. We do not detect any significant GW emission from any of the events. Instead, using a short-duration GW search (for bursts $\leq$ 1 s) we derive 50\% (90\%) upper limits of $10^{48}$ ($10^{49}$) erg for GWs at 300 Hz and $10^{49}$ ($10^{50}$) erg at 2 kHz, and constrain the GW-to-radio energy ratio to $\leq 10^{14} - 10^{16}$. We also derive upper limits from a long-duration search for bursts with durations between 1 and 10 s. These represent the strictest upper limits on concurrent GW emission from FRBs., Comment: 15 pages of text including references, 4 figures, 5 tables
- Published
- 2024
9. Mixture of Multicenter Experts in Multimodal Generative AI for Advanced Radiotherapy Target Delineation
- Author
-
Oh, Yujin, Park, Sangjoon, Li, Xiang, Yi, Wang, Paly, Jonathan, Efstathiou, Jason, Chan, Annie, Kim, Jun Won, Byun, Hwa Kyung, Lee, Ik Jae, Cho, Jaeho, Wee, Chan Woo, Shu, Peng, Wang, Peilong, Yu, Nathan, Holmes, Jason, Ye, Jong Chul, Li, Quanzheng, Liu, Wei, Koom, Woong Sub, Kim, Jin Sung, and Kim, Kyungsang
- Subjects
Electrical Engineering and Systems Science - Image and Video Processing ,Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Machine Learning - Abstract
Clinical experts employ diverse philosophies and strategies in patient care, influenced by regional patient populations. However, existing medical artificial intelligence (AI) models are often trained on data distributions that disproportionately reflect highly prevalent patterns, reinforcing biases and overlooking the diverse expertise of clinicians. To overcome this limitation, we introduce the Mixture of Multicenter Experts (MoME) approach. This method strategically integrates specialized expertise from diverse clinical strategies, enhancing the AI model's ability to generalize and adapt across multiple medical centers. The MoME-based multimodal target volume delineation model, trained with few-shot samples including images and clinical notes from each medical center, outperformed baseline methods in prostate cancer radiotherapy target delineation. The advantages of MoME were most pronounced when data characteristics varied across centers or when data availability was limited, demonstrating its potential for broader clinical applications. Therefore, the MoME framework enables the deployment of AI-based target volume delineation models in resource-constrained medical facilities by adapting to specific preferences of each medical center only using a few sample data, without the need for data sharing between institutions. Expanding the number of multicenter experts within the MoME framework will significantly enhance the generalizability, while also improving the usability and adaptability of clinical AI applications in the field of precision radiation oncology., Comment: 39 pages
- Published
- 2024
10. Online Learning Problems, Academic Worries, Social Interaction, and Psychological Well-Being among Secondary School Students in Hong Kong during the COVID-19 Pandemic: The Socioeconomic and Gender Differences
- Author
-
Siu-Ming Chan, Gary Ka-Ki Chung, Yat-Hang Chan, Thomas Sze-Kit Lee, Ji-Kang Chen, Hung Wong, Roger Yat-Nork Chung, Yikang Chen, and Esther Sui-Chu Ho
- Abstract
During the COVID-19 pandemic, numerous schools halted face-to-face teaching and instead resorted to online classes. The impact of online learning on students' academic worries, social relationships, and psychological well-being has received growing attention. Based on a sample of 1095 students aged 14-16 in 12 secondary schools in Hong Kong, structural equation modeling was used to investigate the association between online learning problems, academic worries and performance, social interaction, and psychosocial well-being during the pandemic. The path relationships among variables were compared by gender groups and socio-economic status (SES). Online learning problems were found to be negatively associated with psychosocial well-being through academic worries and social interaction as mediators. The effect of mediation among female students was larger than that of males, whereas the impact of academic worries on both psychosocial well-being and social interaction was stronger and significant only in the low SES group. As the COVID-19 pandemic significantly impacted the learning and well-being of students, attention should be given to gender and SES difference among students.
- Published
- 2024
- Full Text
- View/download PDF
11. The Burial Record of Prehistoric Liangshan in Southwest China: Graves as Composite Objects by Anke Hein (review)
- Author
-
Chan, Annie
- Published
- 2021
- Full Text
- View/download PDF
12. Recursive and iterative approaches to generate rotation Gray codes for stamp foldings and semi-meanders
- Author
-
Liu, Bowie, Wong, Dennis, Lam, Chan-Tong, and Im, Marcus
- Subjects
Computer Science - Data Structures and Algorithms - Abstract
We first present a simple recursive algorithm that generates cyclic rotation Gray codes for stamp foldings and semi-meanders, where consecutive strings differ by a stamp rotation. These are the first known Gray codes for stamp foldings and semi-meanders, and we thus solve an open problem posted by Sawada and Li in [Electron. J. Comb. 19(2), 2012]. We then introduce an iterative algorithm that generates the same rotation Gray codes for stamp foldings and semi-meanders. Both the recursive and iterative algorithms generate stamp foldings and semi-meanders in constant amortized time and $O(n)$-amortized time per string respectively, using a linear amount of memory.
- Published
- 2024
13. What talking you?: Translating Code-Mixed Messaging Texts to English
- Author
-
Ng, Lynnette Hui Xian and Chan, Luo Qi
- Subjects
Computer Science - Computation and Language - Abstract
Translation of code-mixed texts to formal English allow a wider audience to understand these code-mixed languages, and facilitate downstream analysis applications such as sentiment analysis. In this work, we look at translating Singlish, which is colloquial Singaporean English, to formal standard English. Singlish is formed through the code-mixing of multiple Asian languages and dialects. We analysed the presence of other Asian languages and variants which can facilitate translation. Our dataset is short message texts, written as informal communication between Singlish speakers. We use a multi-step prompting scheme on five Large Language Models (LLMs) for language detection and translation. Our analysis show that LLMs do not perform well in this task, and we describe the challenges involved in translation of code-mixed languages. We also release our dataset in this link https://github.com/luoqichan/singlish.
- Published
- 2024
14. Analyzing The Language of Visual Tokens
- Author
-
Chan, David M., Corona, Rodolfo, Park, Joonyong, Cho, Cheol Jun, Bai, Yutong, and Darrell, Trevor
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence ,Computer Science - Computation and Language ,Computer Science - Machine Learning - Abstract
With the introduction of transformer-based models for vision and language tasks, such as LLaVA and Chameleon, there has been renewed interest in the discrete tokenized representation of images. These models often treat image patches as discrete tokens, analogous to words in natural language, learning joint alignments between visual and human languages. However, little is known about the statistical behavior of these visual languages - whether they follow similar frequency distributions, grammatical structures, or topologies as natural languages. In this paper, we take a natural-language-centric approach to analyzing discrete visual languages and uncover striking similarities and fundamental differences. We demonstrate that, although visual languages adhere to Zipfian distributions, higher token innovation drives greater entropy and lower compression, with tokens predominantly representing object parts, indicating intermediate granularity. We also show that visual languages lack cohesive grammatical structures, leading to higher perplexity and weaker hierarchical organization compared to natural languages. Finally, we demonstrate that, while vision models align more closely with natural languages than other models, this alignment remains significantly weaker than the cohesion found within natural languages. Through these experiments, we demonstrate how understanding the statistical properties of discrete visual languages can inform the design of more effective computer vision models.
- Published
- 2024
15. Field tunable BKT and quantum phase transitions in spin-1/2 triangular lattice antiferromagnet
- Author
-
Zhang, Dechen, Zhu, Yuan, Zheng, Guoxin, Chen, Kuan-Wen, Huang, Qing, Zhou, Lingxiao, Liu, Yujie, Jenkins, Kaila, Chan, Aaron, Zhou, Haidong, and Li, Lu
- Subjects
Condensed Matter - Strongly Correlated Electrons - Abstract
Quantum magnetism is one of the most active fields for exploring exotic phases and phase transitions. The recently synthesized Na2BaCo(PO4)2 (NBCP) is an ideal material incarnation of the spin-1/2 easy axis triangular lattice antiferromagnet (TLAF). Experimental evidence shows that NBCP hosts the spin supersolid state with a giant magnetocaloric effect. It was also proposed that the applied magnetic field B can drive the system through Berezinskii-Kosterlitz-Thouless (BKT) and other richer quantum phase transitions. However, the detection of these transitions is challenging because they onset at extremely low temperature T at around 60 mK, and the measurement of the magnetic susceptibility of these transitions requires high sensitivity. With the help of our newly developed gradient force magnetometer in a dilution refrigerator, we constructed the contour diagram of the magnetic susceptibility in the B-T phase diagram in T as cold as 30 mK. These results provide a more comprehensive and accurate understanding of the several field-tunable quantum phase transitions and BKT melting of the spin supersolidity, which are especially significant when their giant magnetocaloric effects highlight potential applications for sub-Kelvin refrigeration under concerns about global helium shortages., Comment: 19 pages, 5 main figures, and 4 figures in the supplement
- Published
- 2024
16. Topological Singularities in Metasurface Scattering Matrices: From Nodal Lines to Exceptional Lines
- Author
-
Chen, Jingguang, Liu, Wenzhe, Wang, Jiajun, Zhang, Ruo-Yang, Cui, Xiaohan, Guan, Fang, Shi, Lei, Zi, Jian, and Chan, C. T.
- Subjects
Physics - Optics - Abstract
Topological properties of photonic structures described by Hamiltonian matrices have been extensively studied in recent years. Photonic systems are often open systems, and their coupling with the environment is characterized by scattering matrices, which can exhibit topological features as well. In this work, we uncover that topological singularities can be manifested in the scattering matrices of two-dimensional periodic photonic systems with open boundaries in the third dimension, introducing a new topological approach to describe scattering. We elaborate the importance of symmetry and demonstrate that mirror symmetry gives rise to the formation of diabolic points and nodal lines in the three-dimensional frequency-momentum space, which transform into exceptional points and lines in the presence of material loss. These topological features in the eigenvalue structure of the scattering matrix manifest as vortex lines in the cross-polarization scattering phase, providing a direct link between the eigen-problem and observable scattering phenomena in the frequency-momentum space. We demonstrate these phenomena numerically and experimentally using a reflective non-local metasurface. These findings extend the concept of topological singularities to scattering matrices and pave the way for novel photonic devices and wavefront engineering techniques., Comment: 19 pages, 4 figures
- Published
- 2024
17. A robust first order meshfree method for time-dependent nonlinear conservation laws
- Author
-
Kwan, Samuel and Chan, Jesse
- Subjects
Mathematics - Numerical Analysis - Abstract
We introduce a robust first order accurate meshfree method to numerically solve time-dependent nonlinear conservation laws. The main contribution of this work is the meshfree construction of first order consistent summation by parts differentiations. We describe how to efficiently construct such operators on a point cloud. We then study the performance of such differentiations, and then combine these operators with a numerical flux-based formulation to approximate the solution of nonlinear conservation laws, with focus on the advection equation and the compressible Euler equations. We observe numerically that, while the resulting mesh-free differentiation operators are only $O(h^\frac{1}{2})$ accurate in the $L^2$ norm, they achieve $O(h)$ rates of convergence when applied to the numerical solution of PDEs.
- Published
- 2024
18. Benchmarking Accuracy in an Emulated Memory Experiment
- Author
-
Chan, Tim
- Subjects
Computer Science - Distributed, Parallel, and Cluster Computing ,Quantum Physics - Abstract
This note proposes a simpler method to extract the logical error rate from an emulated surface code memory experiment., Comment: 2+1 pages, 2 figures
- Published
- 2024
19. Is This the Same Code? A Comprehensive Study of Decompilation Techniques for WebAssembly Binaries
- Author
-
Wu, Wei-Cheng, Yan, Yutian, Egilsson, Hallgrimur David, Park, David, Chan, Steven, Hauser, Christophe, and Wang, Weihang
- Subjects
Computer Science - Software Engineering - Abstract
WebAssembly is a low-level bytecode language designed for client-side execution in web browsers. The need for decompilation techniques that recover high-level source code from WASM binaries has grown as WASM continues to gain widespread adoption and its security concerns. However little research has been done to assess the quality of decompiled code from WASM. This paper aims to fill this gap by conducting a comprehensive comparative analysis between decompiled C code from WASM binaries and state-of-the-art native binary decompilers. We presented a novel framework for empirically evaluating C-based decompilers from various aspects including correctness/ readability/ and structural similarity. The proposed metrics are validated practicality in decompiler assessment and provided insightful observations regarding the characteristics and constraints of existing decompiled code. This in turn contributes to bolstering the security and reliability of software systems that rely on WASM and native binaries., Comment: SecureComm'24: Proceedings of the 20th EAI International Conference on Security and Privacy in Communication Networks
- Published
- 2024
20. New applications of Bayesian optimization based on element mapping to design high-capacity NASICON-type cathode of sodium-ion battery
- Author
-
Park, Sanghyeon, Shim, Yoonsu, Hur, Junpyo, Jeon, Dongmin, Yuk, Jong Min, and Lee, Chan-Woo
- Subjects
Condensed Matter - Materials Science - Abstract
Sodium-ion batteries are emerging as promising alternatives to lithium-ion batteries due to the abundance of sodium resources. Na3V2(PO4)2F3 (NVPF), a cathode material for sodium ion batteries, is attracting attention from its rate capability and high working voltage, but its low discharge capacity is one of the challenges. In this work, we aim to design a high-capacity NASICON-type cathode of sodium-ion battery by discovering element combinations that can stabilize the sodium excess phase in NVPFs. For the efficient discovery of element combinations, we propose a Bayesian optimization-based algorithm for chemical composition discovery. Specifically, we propose an element mapping technique to solve the limitation of Bayesian optimization in discovering chemical composition. By constructing a chemical space applicable to Bayesian optimization through element mapping and optimizing the constructed chemical space, we found optimal binary element combinations. This work not only offers insights into designing high-capacity cathodes, but also demonstrates the efficacy of the proposed algorithm in data-driven materials design., Comment: 20 pages, 5 figures
- Published
- 2024
21. Unveiling Normative Trajectories of Lifespan Brain Maturation Using Quantitative MRI
- Author
-
Chen, Xinjie, Ocampo-Pineda, Mario, Lu, Po-Jui, Ekerdt, Clara, Weigel, Matthias, Jansen, Michelle G., Cagol, Alessandro, Chan, Kwok-Shing, Schädelin, Sabine, Zwiers, Marcel, Oosterman, Joukje M., Norris, David G., Bayer, Johanna M. M., Marquand, Andre F., Menks, Willeke M., Kuhle, Jens, Kappos, Ludwig, Melie-Garcia, Lester, Granziera, Cristina, and Marques, José P.
- Subjects
Physics - Medical Physics ,Quantitative Biology - Neurons and Cognition - Abstract
Background: Brain maturation and aging involve significant microstructural changes, resulting in functional and cognitive alterations. Quantitative MRI (qMRI) can measure this evolution, distinguishing the physiological effects of normal aging from pathological deviations. Methods: We conducted a multicentre study using qMRI metrics (R1, R2*, and Quantitative Susceptibility Mapping) to model age trajectories across brain structures, including tractography-based white matter bundles (TWMB), superficial white matter (SWM), and cortical grey matter (CGM). MRI data from 537 healthy subjects, aged 8 to 79 years, were harmonized using two independent methods. We modeled age trajectories and performed regional analyses to capture maturation patterns and aging effects across the lifespan. Findings: Our findings revealed a distinct brain maturation gradient, with early qMRI peak values in TWMB, followed by SWM, and culminating in CGM regions. This gradient was observed as a posterior-to-anterior maturation pattern in the cortex and an inferior-to-superior maturation pattern in white matter tracts. R1 demonstrated the most robust age trajectories, while R2* and susceptibility exhibited greater variability and different patterns. The normative modeling framework confirmed the reliability of our age-modelled trajectories across datasets. Interpretation: Our study highlights the potential of multiparametric qMRI to capture complex, region-specific brain development patterns, addressing the need for comprehensive, age-spanning studies across multiple brain structures. Various harmonization strategies can merge qMRI cohorts, improving the robustness of qMRI-based age models and facilitating the understanding of normal patterns and disease-associated deviations.
- Published
- 2024
22. Quantum linear algebra for disordered electrons
- Author
-
Chen, Jielun and Chan, Garnet Kin-Lic
- Subjects
Quantum Physics ,Condensed Matter - Disordered Systems and Neural Networks ,Condensed Matter - Mesoscale and Nanoscale Physics - Abstract
We describe how to use quantum linear algebra to simulate a physically realistic model of disordered non-interacting electrons on exponentially many lattice sites. The physics of disordered electrons outside of one dimension challenges classical computation due to the critical nature of the Anderson localization transition or exponential localization lengths, while the atypical distribution of the local density of states limits the power of disorder averaged approaches. We overcome this by simulating an exponentially large disorder instance using a block-encoded hopping matrix of physical form where disorder is introduced by pseudorandom functions. Key physical quantities, including the reduced density matrix, Green's function, and local density of states, as well as bulk-averaged observables such as the linear conductivity, can then be computed using quantum singular value transformation, quantum amplitude estimation, and trace estimation., Comment: 12 pages, 2 figures
- Published
- 2024
23. Computational Imaging Through Atmospheric Turbulence
- Author
-
Chimitt, Nicholas and Chan, Stanley H.
- Subjects
Electrical Engineering and Systems Science - Image and Video Processing - Abstract
Since the seminal work of Andrey Kolmogorov in the early 1940's, imaging through atmospheric turbulence has grown from a pure scientific pursuit to an important subject across a multitude of civilian, space-mission, and national security applications. Fueled by the recent advancement of deep learning, the field is further experiencing a new wave of momentum. However, for these deep learning methods to perform well, new efforts are needed to build faster and more accurate computational models while at the same time maximizing the performance of image reconstruction. The book is written primarily for image processing engineers, computer vision scientists, and engineering students who are interested in the field of atmospheric turbulence, statistical optics, and image processing. The book can be used as a graduate text, or advanced topic classes for undergraduates.
- Published
- 2024
- Full Text
- View/download PDF
24. A Demonstration of Adaptive Collaboration of Large Language Models for Medical Decision-Making
- Author
-
Kim, Yubin, Park, Chanwoo, Jeong, Hyewon, Grau-Vilchez, Cristina, Chan, Yik Siu, Xu, Xuhai, McDuff, Daniel, Lee, Hyeonhoon, Ghassemi, Marzyeh, Breazeal, Cynthia, and Park, Hae Won
- Subjects
Computer Science - Computation and Language - Abstract
Medical Decision-Making (MDM) is a multi-faceted process that requires clinicians to assess complex multi-modal patient data patient, often collaboratively. Large Language Models (LLMs) promise to streamline this process by synthesizing vast medical knowledge and multi-modal health data. However, single-agent are often ill-suited for nuanced medical contexts requiring adaptable, collaborative problem-solving. Our MDAgents addresses this need by dynamically assigning collaboration structures to LLMs based on task complexity, mimicking real-world clinical collaboration and decision-making. This framework improves diagnostic accuracy and supports adaptive responses in complex, real-world medical scenarios, making it a valuable tool for clinicians in various healthcare settings, and at the same time, being more efficient in terms of computing cost than static multi-agent decision making methods.
- Published
- 2024
25. Evaluating the Evolution of YOLO (You Only Look Once) Models: A Comprehensive Benchmark Study of YOLO11 and Its Predecessors
- Author
-
Jegham, Nidhal, Koh, Chan Young, Abdelatti, Marwan, and Hendawi, Abdeltawab
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
This study presents a comprehensive benchmark analysis of various YOLO (You Only Look Once) algorithms, from YOLOv3 to the newest addition. It represents the first research to comprehensively evaluate the performance of YOLO11, the latest addition to the YOLO family. It evaluates their performance on three diverse datasets: Traffic Signs (with varying object sizes), African Wildlife (with diverse aspect ratios and at least one instance of the object per image), and Ships and Vessels (with small-sized objects of a single class), ensuring a comprehensive assessment across datasets with distinct challenges. To ensure a robust evaluation, we employ a comprehensive set of metrics, including Precision, Recall, Mean Average Precision (mAP), Processing Time, GFLOPs count, and Model Size. Our analysis highlights the distinctive strengths and limitations of each YOLO version. For example: YOLOv9 demonstrates substantial accuracy but struggles with detecting small objects and efficiency whereas YOLOv10 exhibits relatively lower accuracy due to architectural choices that affect its performance in overlapping object detection but excels in speed and efficiency. Additionally, the YOLO11 family consistently shows superior performance in terms of accuracy, speed, computational efficiency, and model size. YOLO11m achieved a remarkable balance of accuracy and efficiency, scoring mAP50-95 scores of 0.795, 0.81, and 0.325 on the Traffic Signs, African Wildlife, and Ships datasets, respectively, while maintaining an average inference time of 2.4ms, a model size of 38.8Mb, and around 67.6 GFLOPs on average. These results provide critical insights for both industry and academia, facilitating the selection of the most suitable YOLO algorithm for diverse applications and guiding future enhancements., Comment: 20 pages
- Published
- 2024
26. Pressure-dependent magnetotransport measurement in Kagome metal Yb$_{0.5}$Co_3Ge$_3$
- Author
-
Cheng, Zhiyuan, Wang, Yaojia, Wu, Heng, Ali, Mazhar N., Chan, Julia Y., and Bhattacharyya, Semonti
- Subjects
Condensed Matter - Strongly Correlated Electrons ,Condensed Matter - Materials Science - Abstract
Kagome materials are known to be an ideal platform that hosts a plethora of interesting phases such as topological states, electronic correlation, and magnetism, owing to their unique band structure and geometry. We report magnetotransport measurement in Kagome metal Yb$_{0.5}$Co_3Ge$_3$ as a function of pressure. Below $\sim25^\circ$ K the temperature dependence of resistance shows an upturn that is accompanied by a strong negative magnetoresistance, which could be attributed to Kondo effect. Upon pressurization above 1 GPa the resistance shows a reduction as a function of temperature below $4^\circ$ K, along with a further enhanced negative magnetoresistance. This might indicate an onset of a pressure-induced Kondo coherence effect.
- Published
- 2024
27. Why Fine-grained Labels in Pretraining Benefit Generalization?
- Author
-
Hong, Guan Zhe, Cui, Yin, Fuxman, Ariel, Chan, Stanely, and Luo, Enming
- Subjects
Computer Science - Machine Learning ,Computer Science - Computer Vision and Pattern Recognition ,Statistics - Machine Learning - Abstract
Recent studies show that pretraining a deep neural network with fine-grained labeled data, followed by fine-tuning on coarse-labeled data for downstream tasks, often yields better generalization than pretraining with coarse-labeled data. While there is ample empirical evidence supporting this, the theoretical justification remains an open problem. This paper addresses this gap by introducing a "hierarchical multi-view" structure to confine the input data distribution. Under this framework, we prove that: 1) coarse-grained pretraining only allows a neural network to learn the common features well, while 2) fine-grained pretraining helps the network learn the rare features in addition to the common ones, leading to improved accuracy on hard downstream test samples., Comment: arXiv admin note: substantial text overlap with arXiv:2303.16887
- Published
- 2024
28. On the Tambara Affine Line
- Author
-
Chan, David, Mehrle, David, Quigley, J. D., Spitz, Ben, and Van Niel, Danika
- Subjects
Mathematics - Algebraic Topology ,Mathematics - Commutative Algebra ,Mathematics - Algebraic Geometry ,Mathematics - Group Theory ,55P91, 19A22, 13A50, 13B21 - Abstract
Tambara functors are the analogue of commutative rings in equivariant algebra. Nakaoka defined ideals in Tambara functors, leading to the definition of the Nakaoka spectrum of prime ideals in a Tambara functor. In this work, we continue the study of the Nakoaka spectra of Tambara functors. We describe, in terms of the Zariski spectra of ordinary commutative rings, the Nakaoka spectra of many Tambara functors. In particular: we identify the Nakaoka spectrum of the fixed point Tambara functor of any $G$-ring with the GIT quotient of its classical Zariski spectrum; we describe the Nakaoka spectrum of the complex representation ring Tambara functor over a cyclic group of prime order $p$; we describe the affine line (the Nakaoka spectra of free Tambara functors on one generator) over a cyclic group of prime order $p$ in terms of the Zariski spectra of $\mathbb{Z}[x]$, $\mathbb{Z}[x,y]$, and the ring of cyclic polynomials $\mathbb{Z}[x_0,\ldots,x_{p-1}]^{C_p}$. To obtain these results, we introduce a "ghost construction" which produces an integral extension of any $C_p$-Tambara functor, the Nakaoka spectrum of which is describable. To relate the Nakaoka spectrum of a Tambara functor to that of its ghost, we prove several new results in equivariant commutative algebra, including a weak form of the Hilbert basis theorem, going up, lying over, and levelwise radicality of prime ideals in Tambara functors. These results also allow us to compute the Krull dimensions of many Tambara functors., Comment: 52 pages. Comments welcome!
- Published
- 2024
29. Toward Understanding In-context vs. In-weight Learning
- Author
-
Chan, Bryan, Chen, Xinyi, György, András, and Schuurmans, Dale
- Subjects
Computer Science - Machine Learning - Abstract
It has recently been demonstrated empirically that in-context learning emerges in transformers when certain distributional properties are present in the training data, but this ability can also diminish upon further training. We provide a new theoretical understanding of these phenomena by identifying simplified distributional properties that give rise to the emergence and eventual disappearance of in-context learning. We do so by first analyzing a simplified model that uses a gating mechanism to choose between an in-weight and an in-context predictor. Through a combination of a generalization error and regret analysis we identify conditions where in-context and in-weight learning emerge. These theoretical findings are then corroborated experimentally by comparing the behaviour of a full transformer on the simplified distributions to that of the stylized model, demonstrating aligned results. We then extend the study to a full large language model, showing how fine-tuning on various collections of natural language prompts can elicit similar in-context and in-weight learning behaviour.
- Published
- 2024
30. VisAidMath: Benchmarking Visual-Aided Mathematical Reasoning
- Author
-
Ma, Jingkun, Zhan, Runzhe, Wong, Derek F., Li, Yang, Sun, Di, Chan, Hou Pong, and Chao, Lidia S.
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence ,Computer Science - Computation and Language ,Computer Science - Machine Learning - Abstract
Although previous research on large language models (LLMs) and large multi-modal models (LMMs) has systematically explored mathematical problem-solving (MPS) within visual contexts, the analysis of how these models process visual information during problem-solving remains insufficient. To address this gap, we present VisAidMath, a benchmark for evaluating the MPS process related to visual information. We follow a rigorous data curation pipeline involving both automated processes and manual annotations to ensure data quality and reliability. Consequently, this benchmark includes 1,200 challenging problems from various mathematical branches, vision-aid formulations, and difficulty levels, collected from diverse sources such as textbooks, examination papers, and Olympiad problems. Based on the proposed benchmark, we conduct comprehensive evaluations on ten mainstream LLMs and LMMs, highlighting deficiencies in the visual-aided reasoning process. For example, GPT-4V only achieves 45.33% accuracy in the visual-aided reasoning task, even with a drop of 2 points when provided with golden visual aids. In-depth analysis reveals that the main cause of deficiencies lies in hallucination regarding the implicit visual reasoning process, shedding light on future research directions in the visual-aided MPS process., Comment: 58 pages, 28 figures
- Published
- 2024
31. Physical Meaning of Principal Component Analysis for Lattice Systems with Translational Invariance
- Author
-
Park, Su-Chan
- Subjects
Condensed Matter - Statistical Mechanics ,Physics - Data Analysis, Statistics and Probability - Abstract
We seek for the physical implication of principal component analysis (PCA) applied to lattice systems with phase transitions, especially when the system is translationally invariant. We present a general approximate formula for a principal component as well as all other eigenvalues and argue that the approximation becomes exact if the size of data is infinite. The formula explains the connection between the principal component and the corresponding order parameter and, therefore, the reason why PCA is successful. Our result can also be used to estimate a principal component without performing matrix diagonalization.
- Published
- 2024
32. Neural Hamilton: Can A.I. Understand Hamiltonian Mechanics?
- Author
-
Kim, Tae-Geun and Park, Seong Chan
- Subjects
Computer Science - Machine Learning ,Mathematical Physics ,Physics - Classical Physics ,Physics - Computational Physics - Abstract
We propose a novel framework based on neural network that reformulates classical mechanics as an operator learning problem. A machine directly maps a potential function to its corresponding trajectory in phase space without solving the Hamilton equations. Most notably, while conventional methods tend to accumulate errors over time through iterative time integration, our approach prevents error propagation. Two newly developed neural network architectures, namely VaRONet and MambONet, are introduced to adapt the Variational LSTM sequence-to-sequence model and leverage the Mamba model for efficient temporal dynamics processing. We tested our approach with various 1D physics problems: harmonic oscillation, double-well potentials, Morse potential, and other potential models outside the training data. Compared to traditional numerical methods based on the fourth-order Runge-Kutta (RK4) algorithm, our model demonstrates improved computational efficiency and accuracy. Code is available at: https://github.com/Axect/Neural_Hamilton, Comment: 33 pages, 8 figures, 9 tables
- Published
- 2024
33. Multi-modal AI for comprehensive breast cancer prognostication
- Author
-
Witowski, Jan, Zeng, Ken, Cappadona, Joseph, Elayoubi, Jailan, Chiru, Elena Diana, Chan, Nancy, Kang, Young-Joon, Howard, Frederick, Ostrovnaya, Irina, Fernandez-Granda, Carlos, Schnabel, Freya, Ozerdem, Ugur, Liu, Kangning, Steinsnyder, Zoe, Thakore, Nitya, Sadic, Mohammad, Yeung, Frank, Liu, Elisa, Hill, Theodore, Swett, Benjamin, Rigau, Danielle, Clayburn, Andrew, Speirs, Valerie, Vetter, Marcus, Sojak, Lina, Soysal, Simone Muenst, Baumhoer, Daniel, Choucair, Khalil, Zong, Yu, Daoud, Lina, Saad, Anas, Abdulsattar, Waleed, Beydoun, Rafic, Pan, Jia-Wern, Makmur, Haslina, Teo, Soo-Hwang, Pak, Linda Ma, Angel, Victor, Zilenaite-Petrulaitiene, Dovile, Laurinavicius, Arvydas, Klar, Natalie, Piening, Brian D., Bifulco, Carlo, Jun, Sun-Young, Yi, Jae Pak, Lim, Su Hyun, Brufsky, Adam, Esteva, Francisco J., Pusztai, Lajos, LeCun, Yann, and Geras, Krzysztof J.
- Subjects
Computer Science - Artificial Intelligence ,Computer Science - Computer Vision and Pattern Recognition ,Electrical Engineering and Systems Science - Image and Video Processing - Abstract
Treatment selection in breast cancer is guided by molecular subtypes and clinical characteristics. Recurrence risk assessment plays a crucial role in personalizing treatment. Current methods, including genomic assays, have limited accuracy and clinical utility, leading to suboptimal decisions for many patients. We developed a test for breast cancer patient stratification based on digital pathology and clinical characteristics using novel AI methods. Specifically, we utilized a vision transformer-based pan-cancer foundation model trained with self-supervised learning to extract features from digitized H&E-stained slides. These features were integrated with clinical data to form a multi-modal AI test predicting cancer recurrence and death. The test was developed and evaluated using data from a total of 8,161 breast cancer patients across 15 cohorts originating from seven countries. Of these, 3,502 patients from five cohorts were used exclusively for evaluation, while the remaining patients were used for training. Our test accurately predicted our primary endpoint, disease-free interval, in the five external cohorts (C-index: 0.71 [0.68-0.75], HR: 3.63 [3.02-4.37, p<0.01]). In a direct comparison (N=858), the AI test was more accurate than Oncotype DX, the standard-of-care 21-gene assay, with a C-index of 0.67 [0.61-0.74] versus 0.61 [0.49-0.73], respectively. Additionally, the AI test added independent information to Oncotype DX in a multivariate analysis (HR: 3.11 [1.91-5.09, p<0.01)]). The test demonstrated robust accuracy across all major breast cancer subtypes, including TNBC (C-index: 0.71 [0.62-0.81], HR: 3.81 [2.35-6.17, p=0.02]), where no diagnostic tools are currently recommended by clinical guidelines. These results suggest that our AI test can improve accuracy, extend applicability to a wider range of patients, and enhance access to treatment selection tools.
- Published
- 2024
34. Finite quotients of Fuchsian groups
- Author
-
Chan, Frankie and Styron, Lindsey
- Subjects
Mathematics - Group Theory ,Mathematics - Geometric Topology - Abstract
This work provides an effective algorithm for distinguishing finite quotients between two non-isomorphic finitely generated Fuchsian groups $\Gamma$ and $\Lambda$. It will suffice to take a finite quotient which is abelian, dihedral, a subgroup of $\mathrm{PSL}(2,\mathbf{F}_q)$, or an abelian extension of one of these 3. We will develop an approach for creating group extensions upon a shared finite quotient of $\Gamma$ and $\Lambda$ which between them have differing degrees of smoothness. Regarding the order of a finite quotient that distinguishes between $\Gamma$ and $\Lambda$, we establish an upperbound as a function of the genera, the number of punctures, and the cone orders arising in $\Gamma$ and $\Lambda$., Comment: v1. 23 pages
- Published
- 2024
35. GPT-4o System Card
- Author
-
OpenAI, Hurst, Aaron, Lerer, Adam, Goucher, Adam P., Perelman, Adam, Ramesh, Aditya, Clark, Aidan, Ostrow, AJ, Welihinda, Akila, Hayes, Alan, Radford, Alec, Mądry, Aleksander, Baker-Whitcomb, Alex, Beutel, Alex, Borzunov, Alex, Carney, Alex, Chow, Alex, Kirillov, Alex, Nichol, Alex, Paino, Alex, Renzin, Alex, Passos, Alex Tachard, Kirillov, Alexander, Christakis, Alexi, Conneau, Alexis, Kamali, Ali, Jabri, Allan, Moyer, Allison, Tam, Allison, Crookes, Amadou, Tootoochian, Amin, Tootoonchian, Amin, Kumar, Ananya, Vallone, Andrea, Karpathy, Andrej, Braunstein, Andrew, Cann, Andrew, Codispoti, Andrew, Galu, Andrew, Kondrich, Andrew, Tulloch, Andrew, Mishchenko, Andrey, Baek, Angela, Jiang, Angela, Pelisse, Antoine, Woodford, Antonia, Gosalia, Anuj, Dhar, Arka, Pantuliano, Ashley, Nayak, Avi, Oliver, Avital, Zoph, Barret, Ghorbani, Behrooz, Leimberger, Ben, Rossen, Ben, Sokolowsky, Ben, Wang, Ben, Zweig, Benjamin, Hoover, Beth, Samic, Blake, McGrew, Bob, Spero, Bobby, Giertler, Bogo, Cheng, Bowen, Lightcap, Brad, Walkin, Brandon, Quinn, Brendan, Guarraci, Brian, Hsu, Brian, Kellogg, Bright, Eastman, Brydon, Lugaresi, Camillo, Wainwright, Carroll, Bassin, Cary, Hudson, Cary, Chu, Casey, Nelson, Chad, Li, Chak, Shern, Chan Jun, Conger, Channing, Barette, Charlotte, Voss, Chelsea, Ding, Chen, Lu, Cheng, Zhang, Chong, Beaumont, Chris, Hallacy, Chris, Koch, Chris, Gibson, Christian, Kim, Christina, Choi, Christine, McLeavey, Christine, Hesse, Christopher, Fischer, Claudia, Winter, Clemens, Czarnecki, Coley, Jarvis, Colin, Wei, Colin, Koumouzelis, Constantin, Sherburn, Dane, Kappler, Daniel, Levin, Daniel, Levy, Daniel, Carr, David, Farhi, David, Mely, David, Robinson, David, Sasaki, David, Jin, Denny, Valladares, Dev, Tsipras, Dimitris, Li, Doug, Nguyen, Duc Phong, Findlay, Duncan, Oiwoh, Edede, Wong, Edmund, Asdar, Ehsan, Proehl, Elizabeth, Yang, Elizabeth, Antonow, Eric, Kramer, Eric, Peterson, Eric, Sigler, Eric, Wallace, Eric, Brevdo, Eugene, Mays, Evan, Khorasani, Farzad, Such, Felipe Petroski, Raso, Filippo, Zhang, Francis, von Lohmann, Fred, Sulit, Freddie, Goh, Gabriel, Oden, Gene, Salmon, Geoff, Starace, Giulio, Brockman, Greg, Salman, Hadi, Bao, Haiming, Hu, Haitang, Wong, Hannah, Wang, Haoyu, Schmidt, Heather, Whitney, Heather, Jun, Heewoo, Kirchner, Hendrik, Pinto, Henrique Ponde de Oliveira, Ren, Hongyu, Chang, Huiwen, Chung, Hyung Won, Kivlichan, Ian, O'Connell, Ian, Osband, Ian, Silber, Ian, Sohl, Ian, Okuyucu, Ibrahim, Lan, Ikai, Kostrikov, Ilya, Sutskever, Ilya, Kanitscheider, Ingmar, Gulrajani, Ishaan, Coxon, Jacob, Menick, Jacob, Pachocki, Jakub, Aung, James, Betker, James, Crooks, James, Lennon, James, Kiros, Jamie, Leike, Jan, Park, Jane, Kwon, Jason, Phang, Jason, Teplitz, Jason, Wei, Jason, Wolfe, Jason, Chen, Jay, Harris, Jeff, Varavva, Jenia, Lee, Jessica Gan, Shieh, Jessica, Lin, Ji, Yu, Jiahui, Weng, Jiayi, Tang, Jie, Yu, Jieqi, Jang, Joanne, Candela, Joaquin Quinonero, Beutler, Joe, Landers, Joe, Parish, Joel, Heidecke, Johannes, Schulman, John, Lachman, Jonathan, McKay, Jonathan, Uesato, Jonathan, Ward, Jonathan, Kim, Jong Wook, Huizinga, Joost, Sitkin, Jordan, Kraaijeveld, Jos, Gross, Josh, Kaplan, Josh, Snyder, Josh, Achiam, Joshua, Jiao, Joy, Lee, Joyce, Zhuang, Juntang, Harriman, Justyn, Fricke, Kai, Hayashi, Kai, Singhal, Karan, Shi, Katy, Karthik, Kavin, Wood, Kayla, Rimbach, Kendra, Hsu, Kenny, Nguyen, Kenny, Gu-Lemberg, Keren, Button, Kevin, Liu, Kevin, Howe, Kiel, Muthukumar, Krithika, Luther, Kyle, Ahmad, Lama, Kai, Larry, Itow, Lauren, Workman, Lauren, Pathak, Leher, Chen, Leo, Jing, Li, Guy, Lia, Fedus, Liam, Zhou, Liang, Mamitsuka, Lien, Weng, Lilian, McCallum, Lindsay, Held, Lindsey, Ouyang, Long, Feuvrier, Louis, Zhang, Lu, Kondraciuk, Lukas, Kaiser, Lukasz, Hewitt, Luke, Metz, Luke, Doshi, Lyric, Aflak, Mada, Simens, Maddie, Boyd, Madelaine, Thompson, Madeleine, Dukhan, Marat, Chen, Mark, Gray, Mark, Hudnall, Mark, Zhang, Marvin, Aljubeh, Marwan, Litwin, Mateusz, Zeng, Matthew, Johnson, Max, Shetty, Maya, Gupta, Mayank, Shah, Meghan, Yatbaz, Mehmet, Yang, Meng Jia, Zhong, Mengchao, Glaese, Mia, Chen, Mianna, Janner, Michael, Lampe, Michael, Petrov, Michael, Wu, Michael, Wang, Michele, Fradin, Michelle, Pokrass, Michelle, Castro, Miguel, de Castro, Miguel Oom Temudo, Pavlov, Mikhail, Brundage, Miles, Wang, Miles, Khan, Minal, Murati, Mira, Bavarian, Mo, Lin, Molly, Yesildal, Murat, Soto, Nacho, Gimelshein, Natalia, Cone, Natalie, Staudacher, Natalie, Summers, Natalie, LaFontaine, Natan, Chowdhury, Neil, Ryder, Nick, Stathas, Nick, Turley, Nick, Tezak, Nik, Felix, Niko, Kudige, Nithanth, Keskar, Nitish, Deutsch, Noah, Bundick, Noel, Puckett, Nora, Nachum, Ofir, Okelola, Ola, Boiko, Oleg, Murk, Oleg, Jaffe, Oliver, Watkins, Olivia, Godement, Olivier, Campbell-Moore, Owen, Chao, Patrick, McMillan, Paul, Belov, Pavel, Su, Peng, Bak, Peter, Bakkum, Peter, Deng, Peter, Dolan, Peter, Hoeschele, Peter, Welinder, Peter, Tillet, Phil, Pronin, Philip, Tillet, Philippe, Dhariwal, Prafulla, Yuan, Qiming, Dias, Rachel, Lim, Rachel, Arora, Rahul, Troll, Rajan, Lin, Randall, Lopes, Rapha Gontijo, Puri, Raul, Miyara, Reah, Leike, Reimar, Gaubert, Renaud, Zamani, Reza, Wang, Ricky, Donnelly, Rob, Honsby, Rob, Smith, Rocky, Sahai, Rohan, Ramchandani, Rohit, Huet, Romain, Carmichael, Rory, Zellers, Rowan, Chen, Roy, Chen, Ruby, Nigmatullin, Ruslan, Cheu, Ryan, Jain, Saachi, Altman, Sam, Schoenholz, Sam, Toizer, Sam, Miserendino, Samuel, Agarwal, Sandhini, Culver, Sara, Ethersmith, Scott, Gray, Scott, Grove, Sean, Metzger, Sean, Hermani, Shamez, Jain, Shantanu, Zhao, Shengjia, Wu, Sherwin, Jomoto, Shino, Wu, Shirong, Shuaiqi, Xia, Phene, Sonia, Papay, Spencer, Narayanan, Srinivas, Coffey, Steve, Lee, Steve, Hall, Stewart, Balaji, Suchir, Broda, Tal, Stramer, Tal, Xu, Tao, Gogineni, Tarun, Christianson, Taya, Sanders, Ted, Patwardhan, Tejal, Cunninghman, Thomas, Degry, Thomas, Dimson, Thomas, Raoux, Thomas, Shadwell, Thomas, Zheng, Tianhao, Underwood, Todd, Markov, Todor, Sherbakov, Toki, Rubin, Tom, Stasi, Tom, Kaftan, Tomer, Heywood, Tristan, Peterson, Troy, Walters, Tyce, Eloundou, Tyna, Qi, Valerie, Moeller, Veit, Monaco, Vinnie, Kuo, Vishal, Fomenko, Vlad, Chang, Wayne, Zheng, Weiyi, Zhou, Wenda, Manassra, Wesam, Sheu, Will, Zaremba, Wojciech, Patil, Yash, Qian, Yilei, Kim, Yongjik, Cheng, Youlong, Zhang, Yu, He, Yuchen, Zhang, Yuchen, Jin, Yujia, Dai, Yunxing, and Malkov, Yury
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Computers and Society ,Computer Science - Machine Learning ,Computer Science - Sound ,Electrical Engineering and Systems Science - Audio and Speech Processing - Abstract
GPT-4o is an autoregressive omni model that accepts as input any combination of text, audio, image, and video, and generates any combination of text, audio, and image outputs. It's trained end-to-end across text, vision, and audio, meaning all inputs and outputs are processed by the same neural network. GPT-4o can respond to audio inputs in as little as 232 milliseconds, with an average of 320 milliseconds, which is similar to human response time in conversation. It matches GPT-4 Turbo performance on text in English and code, with significant improvement on text in non-English languages, while also being much faster and 50\% cheaper in the API. GPT-4o is especially better at vision and audio understanding compared to existing models. In line with our commitment to building AI safely and consistent with our voluntary commitments to the White House, we are sharing the GPT-4o System Card, which includes our Preparedness Framework evaluations. In this System Card, we provide a detailed look at GPT-4o's capabilities, limitations, and safety evaluations across multiple categories, focusing on speech-to-speech while also evaluating text and image capabilities, and measures we've implemented to ensure the model is safe and aligned. We also include third-party assessments on dangerous capabilities, as well as discussion of potential societal impacts of GPT-4o's text and vision capabilities.
- Published
- 2024
36. DQRM: Deep Quantized Recommendation Models
- Author
-
Zhou, Yang, Dong, Zhen, Chan, Ellick, Kalamkar, Dhiraj, Marculescu, Diana, and Keutzer, Kurt
- Subjects
Computer Science - Information Retrieval ,Computer Science - Artificial Intelligence - Abstract
Large-scale recommendation models are currently the dominant workload for many large Internet companies. These recommenders are characterized by massive embedding tables that are sparsely accessed by the index for user and item features. The size of these 1TB+ tables imposes a severe memory bottleneck for the training and inference of recommendation models. In this work, we propose a novel recommendation framework that is small, powerful, and efficient to run and train, based on the state-of-the-art Deep Learning Recommendation Model (DLRM). The proposed framework makes inference more efficient on the cloud servers, explores the possibility of deploying powerful recommenders on smaller edge devices, and optimizes the workload of the communication overhead in distributed training under the data parallelism settings. Specifically, we show that quantization-aware training (QAT) can impose a strong regularization effect to mitigate the severe overfitting issues suffered by DLRMs. Consequently, we achieved INT4 quantization of DLRM models without any accuracy drop. We further propose two techniques that improve and accelerate the conventional QAT workload specifically for the embedding tables in the recommendation models. Furthermore, to achieve efficient training, we quantize the gradients of the embedding tables into INT8 on top of the well-supported specified sparsification. We show that combining gradient sparsification and quantization together significantly reduces the amount of communication. Briefly, DQRM models with INT4 can achieve 79.07% accuracy on Kaggle with 0.27 GB model size, and 81.21% accuracy on the Terabyte dataset with 1.57 GB, which even outperform FP32 DLRMs that have much larger model sizes (2.16 GB on Kaggle and 12.58 on Terabyte).
- Published
- 2024
37. MetaTrading: An Immersion-Aware Model Trading Framework for Vehicular Metaverse Services
- Author
-
Wu, Hongjia, Zeng, Hui, Xiong, Zehui, Kang, Jiawen, Cai, Zhiping, Chan, Tse-Tin, Niyato, Dusit, and Han, Zhu
- Subjects
Computer Science - Machine Learning ,Computer Science - Cryptography and Security ,Computer Science - Computer Science and Game Theory - Abstract
Updates of extensive Internet of Things (IoT) data are critical to the immersion of vehicular metaverse services. However, providing high-quality and sustainable data in unstable and resource-constrained vehicular networks remains a significant challenge. To address this problem, we put forth a novel immersion-aware model trading framework that incentivizes metaverse users (MUs) to contribute learning models trained by their latest local data for augmented reality (AR) services in the vehicular metaverse, while preserving their privacy through federated learning. To comprehensively evaluate the contribution of locally trained learning models provided by MUs to AR services, we design a new immersion metric that captures service immersion by considering the freshness and accuracy of learning models, as well as the amount and potential value of raw data used for training. We model the trading interactions between metaverse service providers (MSPs) and MUs as an equilibrium problem with equilibrium constraints (EPEC) to analyze and balance their costs and gains. Moreover, considering dynamic network conditions and privacy concerns, we formulate the reward decisions of MSPs as a multi-agent Markov decision process. Then, a fully distributed dynamic reward method based on deep reinforcement learning is presented, which operates without any private information about MUs and other MSPs. Experimental results demonstrate that the proposed framework can effectively provide higher-value models for object detection and classification in AR services on real AR-related vehicle datasets compared to benchmark schemes.
- Published
- 2024
38. Amoeba: An AGN Model of Optical Emissions Beyond steady-state Accretion discs
- Author
-
Best, Henry, O'Dowd, Matthew, Fagin, Joshua, Chan, James H. H., and Ierace, Bridget
- Subjects
Astrophysics - Astrophysics of Galaxies ,Astrophysics - High Energy Astrophysical Phenomena - Abstract
Active Galactic Nuclei (AGN) are objects located in the heart of galaxies which emit powerful and complex radiation across the electromagnetic spectrum. Understanding AGN has become a topic of interest due to their importance in galactic evolution and their ability to act as a probe to the distant Universe. Within the next few years, wide-field surveys such as the Legacy Survey of Space and Time (LSST) at the Rubin Vera Observatory are expected to increase the number of known AGN to $\mathcal{O} (10^{7})$ and the number of strongly lensed AGN to $\mathcal{O} (10^{4})$. In this paper we introduce \texttt{Amoeba}: an AGN Model of Optical Emission Beyond steady-state Accretion discs. The goal of \texttt{Amoeba} is to provide a modular and flexible modelling environment for AGN, in which all components can interact with each other. Through this work we describe the framework for major AGN components to vary self-consistently and keep flux distributions to connect these components to spatial dependent processes. We model properties beyond traditional single-component models, such as the reverberation of the corona's bending power law power spectrum through the accretion disc and broad line region (BLR). We simulate obscuration by the dusty torus and differential magnification of the disc and BLR due to microlensing. These features are joined together to create some of the most realistic light curve simulations to date. \texttt{Amoeba} takes a step forward in AGN modelling by joining the accretion disc, BLR, torus, intrinsic signal, and microlensing into a coherent model.
- Published
- 2024
39. Joint Modeling of Quasar Variability and Accretion Disk Reprocessing using Latent Stochastic Differential Equations
- Author
-
Fagin, Joshua, Chan, James Hung-Hsu, Best, Henry, O'Dowd, Matthew, Ford, K. E. Saavik, Graham, Matthew J., Park, Ji Won, and Villar, V. Ashley
- Subjects
Astrophysics - Astrophysics of Galaxies ,Astrophysics - Instrumentation and Methods for Astrophysics - Abstract
Quasars are bright active galactic nuclei powered by the accretion of matter around supermassive black holes at the center of galaxies. Their stochastic brightness variability depends on the physical properties of the accretion disk and black hole. The upcoming Rubin Observatory Legacy Survey of Space and Time (LSST) is expected to observe tens of millions of quasars, so there is a need for efficient techniques like machine learning that can handle the large volume of data. Quasar variability is believed to be driven by an X-ray corona, which is reprocessed by the accretion disk and emitted as UV/optical variability. We are the first to introduce an auto-differentiable simulation of the accretion disk and reprocessing. We use the simulation as a direct component of our neural network to jointly model the driving variability and reprocessing to fit simulated LSST 10-year quasar light curves. The driving variability is reconstructed using a latent stochastic differential equation, a physically motivated, generative deep learning method that can model continuous-time stochastic dynamics. By embedding these physical processes into our network, we achieve a model that is more robust and interpretable. We also use transformers to scale our model to tens of millions of parameters. We demonstrate how our model outperforms a Gaussian process regression baseline and can infer accretion disk parameters and time delays between wavebands, even for out-of-distribution driving signals. Our approach provides a powerful and scalable framework that can be adapted to solve other inverse problems in multivariate time series with irregular sampling., Comment: 31 pages, 19 figures
- Published
- 2024
40. metasnf: Meta Clustering with Similarity Network Fusion in R
- Author
-
Velayudhan, Prashanth S, Xu, Xiaoqiao, Kallurkar, Prajkta, Balbon, Ana Patricia, Secara, Maria T, Taback, Adam, Sabac, Denise, Chan, Nicholas, Ma, Shihao, Wang, Bo, Felsky, Daniel, Ameis, Stephanie H, Cox, Brian, Hawco, Colin, Erdman, Lauren, and Wheeler, Anne L
- Subjects
Statistics - Computation ,Computer Science - Machine Learning - Abstract
metasnf is an R package that enables users to apply meta clustering, a method for efficiently searching a broad space of cluster solutions by clustering the solutions themselves, to clustering workflows based on similarity network fusion (SNF). SNF is a multi-modal data integration algorithm commonly used for biomedical subtype discovery. The package also contains functions to assist with cluster visualization, characterization, and validation. This package can help researchers identify SNF-derived cluster solutions that are guided by context-specific utility over context-agnostic measures of quality., Comment: 72 pages, 22 figures, submitted to Journal of Statistical Software
- Published
- 2024
41. Medical Imaging Complexity and its Effects on GAN Performance
- Author
-
Cagas, William, Ko, Chan, Hsiao, Blake, Grandhi, Shryuk, Bhattacharya, Rishi, Zhu, Kevin, and Lam, Michael
- Subjects
Electrical Engineering and Systems Science - Image and Video Processing ,Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Machine Learning - Abstract
The proliferation of machine learning models in diverse clinical applications has led to a growing need for high-fidelity, medical image training data. Such data is often scarce due to cost constraints and privacy concerns. Alleviating this burden, medical image synthesis via generative adversarial networks (GANs) emerged as a powerful method for synthetically generating photo-realistic images based on existing sets of real medical images. However, the exact image set size required to efficiently train such a GAN is unclear. In this work, we experimentally establish benchmarks that measure the relationship between a sample dataset size and the fidelity of the generated images, given the dataset's distribution of image complexities. We analyze statistical metrics based on delentropy, an image complexity measure rooted in Shannon's entropy in information theory. For our pipeline, we conduct experiments with two state-of-the-art GANs, StyleGAN 3 and SPADE-GAN, trained on multiple medical imaging datasets with variable sample sizes. Across both GANs, general performance improved with increasing training set size but suffered with increasing complexity., Comment: Accepted to ACCV, Workshop on Generative AI for Synthetic Medical Data
- Published
- 2024
42. Zeta functions of orders on surfaces
- Author
-
Chan, Daniel and Lynch, Sean B.
- Subjects
Mathematics - Number Theory ,11R54 - Abstract
In this manuscript, we give effective methods for computing the zeta function of maximal orders on surfaces., Comment: 23 pages
- Published
- 2024
43. Satori: Towards Proactive AR Assistant with Belief-Desire-Intention User Modeling
- Author
-
Li, Chenyi, Wu, Guande, Chan, Gromit Yeuk-Yin, Turakhia, Dishita G, Quispe, Sonia Castelo, Li, Dong, Welch, Leslie, Silva, Claudio, and Qian, Jing
- Subjects
Computer Science - Human-Computer Interaction ,Computer Science - Artificial Intelligence - Abstract
Augmented Reality assistance are increasingly popular for supporting users with tasks like assembly and cooking. However, current practice typically provide reactive responses initialized from user requests, lacking consideration of rich contextual and user-specific information. To address this limitation, we propose a novel AR assistance system, Satori, that models both user states and environmental contexts to deliver proactive guidance. Our system combines the Belief-Desire-Intention (BDI) model with a state-of-the-art multi-modal large language model (LLM) to infer contextually appropriate guidance. The design is informed by two formative studies involving twelve experts. A sixteen within-subject study find that Satori achieves performance comparable to an designer-created Wizard-of-Oz (WoZ) system without relying on manual configurations or heuristics, thereby enhancing generalizability, reusability and opening up new possibilities for AR assistance.
- Published
- 2024
44. Optical optimization of a multi-slit extreme ultraviolet spectrograph for global solar corona diagnostics
- Author
-
Feng, Yufei, Bai, Xianyong, Guo, Sifan, Tian, Hui, Chan, Lami, Deng, Yuanyong, Yang, Qi, Duan, Wei, Zhu, Xiaoming, Yang, Xiao, Feng, Zhiwei, and Zhang, Zhiyong
- Subjects
Astrophysics - Solar and Stellar Astrophysics - Abstract
The spatial-temporal evolution of coronal plasma parameters of the solar outer atmosphere at global scales, derived from solar full-disk imaging spectroscopic observation in the extreme-ultraviolet band, is critical for understanding and forecasting solar eruptions. We propose a multi-slits extreme ultraviolet imaging spectrograph for global coronal diagnostics with high cadence and present the preliminary instrument designs for the wavelength range from 18.3 to 19.8 nm. The instrument takes a comprehensive approach to obtain global coronal spatial and spectral information, improve the detected cadence and avoid overlapping. We first describe the relationship between optical properties and structural parameters, especially the relationship between the overlapping and the number of slits, and give a general multi-slits extreme-ultraviolet imaging spectrograph design process. Themultilayer structure is optimized to enhance the effective areas in the observation band. Five distantly-separated slits are set to divide the entire solar field of view, which increase the cadence for raster scanning the solar disk by 5 times relative to a single slit. The spectral resolving power of the optical system with an aperture diameter of 150 mm are optimized to be greater than 1461. The spatial resolution along the slits direction and the scanning direction are about 4.4''and 6.86'', respectively. The Al/Mo/B4C multilayer structure is optimized and the peak effective area is about 1.60 cm2 at 19.3 nm with a full width at half maximum of about 1.3 nm. The cadence to finish full-disk raster scan is about 5 minutes. Finally, the instrument performance is evaluated by an end-to-end calculation of the system photon budget and a simulation of the observational image and spectra. Our investigation shows that this approach is promising for global coronal plasma diagnostics., Comment: This version of the article has been accepted for publication, after peer review (when applicable) but is not the Version of Record and does not reflect post-acceptance improvements, or any corrections. The Version of Record is available online at: http://dx.doi.org/10.1007/s10686-024-09961-9
- Published
- 2024
- Full Text
- View/download PDF
45. Rulebreakers Challenge: Revealing a Blind Spot in Large Language Models' Reasoning with Formal Logic
- Author
-
Chan, Jason, Gaizauskas, Robert, and Zhao, Zhixue
- Subjects
Computer Science - Computation and Language - Abstract
Formal logic has long been applied to natural language reasoning, but this approach can sometimes lead to conclusions that, while logically entailed, are factually inconsistent with the premises or are not typically inferred by humans. This study introduces the concept of "rulebreakers", which refers to instances where logical entailment diverges from factually acceptable inference. We present RULEBREAKERS, a novel dataset for evaluating Large Language Models' (LLMs) ability to distinguish between rulebreakers and non-rulebreakers. Focusing on modus tollens and disjunctive syllogism, we assess six state-of-the-art LLMs using RULEBREAKERS, measuring their performance in terms of token-level exact accuracy and model confidence. Our findings reveal that while most models perform poorly to moderately in recognizing rulebreakers, they demonstrate a latent ability to distinguish rulebreakers when assessed by their confidence levels. Further analysis suggests that the failure to recognize rulebreakers is potentially associated with the models' world knowledge and their attention distribution patterns. This research highlights the limitation of LLMs' reasoning capabilities, and contributes to the ongoing discussion on reasoning in LLMs., Comment: Preprint
- Published
- 2024
46. Limpeh ga li gong: Challenges in Singlish Annotations
- Author
-
Chan, Luo Qi and Ng, Lynnette Hui Xian
- Subjects
Computer Science - Computation and Language ,Computer Science - Information Retrieval - Abstract
Singlish, or Colloquial Singapore English, is a language formed from oral and social communication within multicultural Singapore. In this work, we work on a fundamental Natural Language Processing (NLP) task: Parts-Of-Speech (POS) tagging of Singlish sentences. For our analysis, we build a parallel Singlish dataset containing direct English translations and POS tags, with translation and POS annotation done by native Singlish speakers. Our experiments show that automatic transition- and transformer- based taggers perform with only $\sim 80\%$ accuracy when evaluated against human-annotated POS labels, suggesting that there is indeed room for improvement on computation analysis of the language. We provide an exposition of challenges in Singlish annotation: its inconsistencies in form and semantics, the highly context-dependent particles of the language, its structural unique expressions, and the variation of the language on different mediums. Our task definition, resultant labels and results reflects the challenges in analysing colloquial languages formulated from a variety of dialects, and paves the way for future studies beyond POS tagging.
- Published
- 2024
47. ComPO: Community Preferences for Language Model Personalization
- Author
-
Kumar, Sachin, Park, Chan Young, Tsvetkov, Yulia, Smith, Noah A., and Hajishirzi, Hannaneh
- Subjects
Computer Science - Computation and Language - Abstract
Conventional algorithms for training language models (LMs) with human feedback rely on preferences that are assumed to account for an "average" user, disregarding subjectivity and finer-grained variations. Recent studies have raised concerns that aggregating such diverse and often contradictory human feedback to finetune models results in generic models that generate outputs not preferred by many user groups, as they tend to average out styles and norms. To address this issue, we draw inspiration from recommendation systems and propose ComPO, a method to personalize preference optimization in LMs by contextualizing the probability distribution of model outputs with the preference provider. Focusing on group-level preferences rather than individuals, we collect and release ComPRed, a question answering dataset with community-level preferences from Reddit. This dataset facilitates studying diversity in preferences without incurring privacy concerns associated with individual feedback. Our experiments reveal that conditioning language models on a community identifier (i.e., subreddit name) during preference tuning substantially enhances model performance. Conversely, replacing this context with random subreddit identifiers significantly diminishes performance, highlighting the effectiveness of our approach in tailoring responses to communities' preferences.
- Published
- 2024
48. A 300 mm foundry silicon spin qubit unit cell exceeding 99% fidelity in all operations
- Author
-
Steinacker, Paul, Stuyck, Nard Dumoulin, Lim, Wee Han, Tanttu, Tuomo, Feng, MengKe, Nickl, Andreas, Serrano, Santiago, Candido, Marco, Cifuentes, Jesus D., Hudson, Fay E., Chan, Kok Wai, Kubicek, Stefan, Jussot, Julien, Canvel, Yann, Beyne, Sofie, Shimura, Yosuke, Loo, Roger, Godfrin, Clement, Raes, Bart, Baudot, Sylvain, Wan, Danny, Laucht, Arne, Yang, Chih Hwan, Saraiva, Andre, Escott, Christopher C., De Greve, Kristiaan, and Dzurak, Andrew S.
- Subjects
Condensed Matter - Mesoscale and Nanoscale Physics ,Quantum Physics - Abstract
Fabrication of quantum processors in advanced 300 mm wafer-scale complementary metal-oxide-semiconductor (CMOS) foundries provides a unique scaling pathway towards commercially viable quantum computing with potentially millions of qubits on a single chip. Here, we show precise qubit operation of a silicon two-qubit device made in a 300 mm semiconductor processing line. The key metrics including single- and two-qubit control fidelities exceed 99% and state preparation and measurement fidelity exceeds 99.9%, as evidenced by gate set tomography (GST). We report coherence and lifetimes up to $T_\mathrm{2}^{\mathrm{*}} = 30.4$ $\mu$s, $T_\mathrm{2}^{\mathrm{Hahn}} = 803$ $\mu$s, and $T_1 = 6.3$ s. Crucially, the dominant operational errors originate from residual nuclear spin carrying isotopes, solvable with further isotopic purification, rather than charge noise arising from the dielectric environment. Our results answer the longstanding question whether the favourable properties including high-fidelity operation and long coherence times can be preserved when transitioning from a tailored academic to an industrial semiconductor fabrication technology., Comment: 10 pages, 4 figures, 4 extended data figures
- Published
- 2024
49. EVA: An Embodied World Model for Future Video Anticipation
- Author
-
Chi, Xiaowei, Zhang, Hengyuan, Fan, Chun-Kai, Qi, Xingqun, Zhang, Rongyu, Chen, Anthony, Chan, Chi-min, Xue, Wei, Luo, Wenhan, Zhang, Shanghang, and Guo, Yike
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Multimedia ,Computer Science - Robotics - Abstract
World models integrate raw data from various modalities, such as images and language to simulate comprehensive interactions in the world, thereby displaying crucial roles in fields like mixed reality and robotics. Yet, applying the world model for accurate video prediction is quite challenging due to the complex and dynamic intentions of the various scenes in practice. In this paper, inspired by the human rethinking process, we decompose the complex video prediction into four meta-tasks that enable the world model to handle this issue in a more fine-grained manner. Alongside these tasks, we introduce a new benchmark named Embodied Video Anticipation Benchmark (EVA-Bench) to provide a well-rounded evaluation. EVA-Bench focused on evaluating the video prediction ability of human and robot actions, presenting significant challenges for both the language model and the generation model. Targeting embodied video prediction, we propose the Embodied Video Anticipator (EVA), a unified framework aiming at video understanding and generation. EVA integrates a video generation model with a visual language model, effectively combining reasoning capabilities with high-quality generation. Moreover, to enhance the generalization of our framework, we tailor-designed a multi-stage pretraining paradigm that adaptatively ensembles LoRA to produce high-fidelity results. Extensive experiments on EVA-Bench highlight the potential of EVA to significantly improve performance in embodied scenes, paving the way for large-scale pre-trained models in real-world prediction tasks.
- Published
- 2024
50. Non-Hermitian Skin Effect Along Hyperbolic Geodesics
- Author
-
Shen, Ruizhe, Chan, Wei Jie, and Lee, Ching Hua
- Subjects
Condensed Matter - Other Condensed Matter ,Condensed Matter - Mesoscale and Nanoscale Physics - Abstract
Recently, it has been revealed that a variety of novel phenomena emerge in hyperbolic spaces, while non-Hermitian physics has significantly enriched the landscape of condensed matter physics. Building on these developments, we construct a geodesic-based method to study the non-Hermitian skin effect (NHSE) in non-reciprocal hyperbolic lattices. Additionally, we develop a geodesic-periodic boundary condition (geodesic-PBC), akin to the Euclidean periodic boundary condition (PBC), that complements its open boundary condition. Importantly, we find that the non-reciprocal directionality within a hyperbolic regular polygon and the geodesic-based boundary determine the spectral sensitivity, and hence, the NHSE. Unlike in Euclidean models, however, we must utilize boundary localization to distinguish non-trivial skin modes from their trivial boundary counterpart due to the extensive boundary volume of hyperbolic lattices. We also relate the spatial density profile with the finite-size scaling of hyperbolic lattices. These aspects highlight the profound impact of hyperbolic geometry on non-Hermitian systems and offer new perspectives on the intricate relationship between the geometric characteristics of hyperbolic lattices and non-Hermitian physics., Comment: 13 pages and 10 figures
- Published
- 2024
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.