1,214,474 results on '"Khan, A."'
Search Results
2. Evaluation of antianxiety effect of Nigella sativa Linn. seed oil in mice
- Author
-
Khan, A.K. Afzal, Ghanta, Mohan Krishna, Nayaka, Swapna R., and Usha, N.S.
- Published
- 2023
- Full Text
- View/download PDF
3. Efficient and Power-Aware Design of a Novel Sparse Kogge-Stone Adder using Hybrid Carry Prefix Generator Adder
- Author
-
KHAN, A. and WAIRYA, S.
- Subjects
circuit simulation ,circuit topology ,digital circuits ,parallel architectures ,very large scale integration ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 ,Computer engineering. Computer hardware ,TK7885-7895 - Abstract
This paper presents a novel Sparse Kogge-Stone adder architecture with a sparsity factor of 2, offering a compelling solution to the challenges faced by parallel prefix adders. The superior performance is achieved by including the hybrid carry prefix generator adder (HCPGA), which leads to the elimination of redundant components, and improvements in power consumption and circuit area without compromising computation speed. The proposed hybrid architecture efficiently generates carry prefixes that negates the need for the conventional generate and propagate block, resulting in reduced computational complexity. The effectiveness of the proposed architecture has been extensively validated using Cadence Virtuoso in the 45nm technology node. In addition to evaluating standard performance parameters such as power, delay, and area, comprehensive Monte Carlo simulations and process corner analyses have been performed to ensure the robustness and reliability of the design. Furthermore, the practical application of the proposed architecture has been demonstrated by integrating it into a digital multiplier architecture, showcasing its potential to enhance the computational capabilities of complex arithmetic circuits. This research contributes to the advancement of efficient adder designs for high-performance computing applications, making it highly beneficial and relevant for modern digital circuit designs.
- Published
- 2024
- Full Text
- View/download PDF
4. Utilization of mushroom waste as non-conventional feed additive in broiler chicken
- Author
-
Banday, M.T., Adil, S., Wani, M.A., Khan, A.A., Sheikh, I.U., and Shubeena, S.
- Published
- 2023
- Full Text
- View/download PDF
5. Constraints faced by yak and yak cross rearing communities of Ladakh Region
- Author
-
Wani, Akeel Yousf, Khan, A. A., Hamadani, H., Sheikh, I.U., Banday, M.T., Akand, A. H., Shahnaz, S., and Baba, S.H.
- Published
- 2023
- Full Text
- View/download PDF
6. Indian agricultural sector present and post pandemic condition
- Author
-
Upendra, R.S., Ahmed, Mohammed Riyaz, Kumar, T. Nitesh, Prithviraj, S.R., and Khan, A. Shahid
- Published
- 2023
- Full Text
- View/download PDF
7. Productive performance and economics of broiler chicken fed heat treated sheep manure based diets supplemented with enzyme
- Author
-
Sheikh, I.U., Banday, M.T., Khan, A.A., Adil, S., Baba, I.A., Hamadani, H., Patoo, R.A., Zaffer, B., and Nissa, S.S.
- Published
- 2022
- Full Text
- View/download PDF
8. Epidemiological Pattern of Neonatal Calf Diarrhea and a Randomized On-Field Trial to Evaluate Effectiveness of Zinc
- Author
-
Bhat, I.A., Ain, Q.U., Bashir, S., Nazir, T., Sheikh, G.N., Khan, A.A., and Dar, A.A.
- Published
- 2022
- Full Text
- View/download PDF
9. VideoGLaMM: A Large Multimodal Model for Pixel-Level Visual Grounding in Videos
- Author
-
Munasinghe, Shehan, Gani, Hanan, Zhu, Wenqi, Cao, Jiale, Xing, Eric, Khan, Fahad Shahbaz, and Khan, Salman
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Fine-grained alignment between videos and text is challenging due to complex spatial and temporal dynamics in videos. Existing video-based Large Multimodal Models (LMMs) handle basic conversations but struggle with precise pixel-level grounding in videos. To address this, we introduce VideoGLaMM, a LMM designed for fine-grained pixel-level grounding in videos based on user-provided textual inputs. Our design seamlessly connects three key components: a Large Language Model, a dual vision encoder that emphasizes both spatial and temporal details, and a spatio-temporal decoder for accurate mask generation. This connection is facilitated via tunable V-L and L-V adapters that enable close Vision-Language (VL) alignment. The architecture is trained to synchronize both spatial and temporal elements of video content with textual instructions. To enable fine-grained grounding, we curate a multimodal dataset featuring detailed visually-grounded conversations using a semiautomatic annotation pipeline, resulting in a diverse set of 38k video-QA triplets along with 83k objects and 671k masks. We evaluate VideoGLaMM on three challenging tasks: Grounded Conversation Generation, Visual Grounding, and Referring Video Segmentation. Experimental results show that our model consistently outperforms existing approaches across all three tasks., Comment: Technical Report of VideoGLaMM
- Published
- 2024
10. Auto-assessment of assessment: A conceptual framework towards fulfilling the policy gaps in academic assessment practices
- Author
-
Khan, Wasiq, Topham, Luke K., Atherton, Peter, Al-Shabandar, Raghad, Kolivand, Hoshang, Khan, Iftikhar, and Hussain, Abir
- Subjects
Computer Science - Computers and Society ,Computer Science - Artificial Intelligence ,I.2 ,K.3 - Abstract
Education is being transformed by rapid advances in Artificial Intelligence (AI), including emerging Generative Artificial Intelligence (GAI). Such technology can significantly support academics and students by automating monotonous tasks and making personalised suggestions. However, despite the potential of the technology, there are significant concerns regarding AI misuse, particularly by students in assessments. There are two schools of thought: one advocates for a complete ban on it, while the other views it as a valuable educational tool, provided it is governed by a robust usage policy. This contradiction clearly indicates a major policy gap in academic practices, and new policies are required to uphold academic standards while enabling staff and students to benefit from technological advancements. We surveyed 117 academics from three countries (UK, UAE, and Iraq), and identified that most academics retain positive opinions regarding AI in education. For example, the majority of experienced academics do not favour complete bans, and they see the potential benefits of AI for students, teaching staff, and academic institutions. Importantly, academics specifically identified the particular benefits of AI for autonomous assessment (71.79% of respondents agreed). Therefore, for the first time, we propose a novel AI framework for autonomously evaluating students' work (e.g., reports, coursework, etc.) and automatically assigning grades based on their knowledge and in-depth understanding of the submitted content. The survey results further highlight a significant lack of awareness of modern AI-based tools (e.g., ChatGPT) among experienced academics, a gap that must be addressed to uphold educational standards., Comment: 20 Pages, 5 Figures, submitted for journal peer-review
- Published
- 2024
11. CAMEL-Bench: A Comprehensive Arabic LMM Benchmark
- Author
-
Ghaboura, Sara, Heakl, Ahmed, Thawakar, Omkar, Alharthi, Ali, Riahi, Ines, Saif, Abduljalil, Laaksonen, Jorma, Khan, Fahad S., Khan, Salman, and Anwer, Rao M.
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence ,Computer Science - Computation and Language ,Computer Science - Computers and Society ,Computer Science - Machine Learning - Abstract
Recent years have witnessed a significant interest in developing large multimodal models (LMMs) capable of performing various visual reasoning and understanding tasks. This has led to the introduction of multiple LMM benchmarks to evaluate LMMs on different tasks. However, most existing LMM evaluation benchmarks are predominantly English-centric. In this work, we develop a comprehensive LMM evaluation benchmark for the Arabic language to represent a large population of over 400 million speakers. The proposed benchmark, named CAMEL-Bench, comprises eight diverse domains and 38 sub-domains including, multi-image understanding, complex visual perception, handwritten document understanding, video understanding, medical imaging, plant diseases, and remote sensing-based land use understanding to evaluate broad scenario generalizability. Our CAMEL-Bench comprises around 29,036 questions that are filtered from a larger pool of samples, where the quality is manually verified by native speakers to ensure reliable model assessment. We conduct evaluations of both closed-source, including GPT-4 series, and open-source LMMs. Our analysis reveals the need for substantial improvement, especially among the best open-source models, with even the closed-source GPT-4o achieving an overall score of 62%. Our benchmark and evaluation scripts are open-sourced., Comment: 10 pages, 5 figures, NAACL
- Published
- 2024
12. How to Continually Adapt Text-to-Image Diffusion Models for Flexible Customization?
- Author
-
Dong, Jiahua, Liang, Wenqi, Li, Hongliu, Zhang, Duzhen, Cao, Meng, Ding, Henghui, Khan, Salman, and Khan, Fahad Shahbaz
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Custom diffusion models (CDMs) have attracted widespread attention due to their astonishing generative ability for personalized concepts. However, most existing CDMs unreasonably assume that personalized concepts are fixed and cannot change over time. Moreover, they heavily suffer from catastrophic forgetting and concept neglect on old personalized concepts when continually learning a series of new concepts. To address these challenges, we propose a novel Concept-Incremental text-to-image Diffusion Model (CIDM), which can resolve catastrophic forgetting and concept neglect to learn new customization tasks in a concept-incremental manner. Specifically, to surmount the catastrophic forgetting of old concepts, we develop a concept consolidation loss and an elastic weight aggregation module. They can explore task-specific and task-shared knowledge during training, and aggregate all low-rank weights of old concepts based on their contributions during inference. Moreover, in order to address concept neglect, we devise a context-controllable synthesis strategy that leverages expressive region features and noise estimation to control the contexts of generated images according to user conditions. Experiments validate that our CIDM surpasses existing custom diffusion models. The source codes are available at https://github.com/JiahuaDong/CIFC., Comment: Accepted to NeurIPS2024
- Published
- 2024
13. Frontiers in Intelligent Colonoscopy
- Author
-
Ji, Ge-Peng, Liu, Jingyi, Xu, Peng, Barnes, Nick, Khan, Fahad Shahbaz, Khan, Salman, and Fan, Deng-Ping
- Subjects
Electrical Engineering and Systems Science - Image and Video Processing ,Computer Science - Computer Vision and Pattern Recognition - Abstract
Colonoscopy is currently one of the most sensitive screening methods for colorectal cancer. This study investigates the frontiers of intelligent colonoscopy techniques and their prospective implications for multimodal medical applications. With this goal, we begin by assessing the current data-centric and model-centric landscapes through four tasks for colonoscopic scene perception, including classification, detection, segmentation, and vision-language understanding. This assessment enables us to identify domain-specific challenges and reveals that multimodal research in colonoscopy remains open for further exploration. To embrace the coming multimodal era, we establish three foundational initiatives: a large-scale multimodal instruction tuning dataset ColonINST, a colonoscopy-designed multimodal language model ColonGPT, and a multimodal benchmark. To facilitate ongoing monitoring of this rapidly evolving field, we provide a public website for the latest updates: https://github.com/ai4colonoscopy/IntelliScope., Comment: [work in progress] A comprehensive survey of intelligent colonoscopy in the multimodal era
- Published
- 2024
14. Search for gravitational waves emitted from SN 2023ixf
- Author
-
The LIGO Scientific Collaboration, the Virgo Collaboration, the KAGRA Collaboration, Abac, A. G., Abbott, R., Abouelfettouh, I., Acernese, F., Ackley, K., Adhicary, S., Adhikari, N., Adhikari, R. X., Adkins, V. K., Agarwal, D., Agathos, M., Abchouyeh, M. Aghaei, Aguiar, O. D., Aguilar, I., Aiello, L., Ain, A., Akutsu, T., Albanesi, S., Alfaidi, R. A., Al-Jodah, A., Alléné, C., Allocca, A., Al-Shammari, S., Altin, P. A., Alvarez-Lopez, S., Amato, A., Amez-Droz, L., Amorosi, A., Amra, C., Ananyeva, A., Anderson, S. B., Anderson, W. G., Andia, M., Ando, M., Andrade, T., Andres, N., Andrés-Carcasona, M., Andrić, T., Anglin, J., Ansoldi, S., Antelis, J. M., Antier, S., Aoumi, M., Appavuravther, E. Z., Appert, S., Apple, S. K., Arai, K., Araya, A., Araya, M. C., Areeda, J. S., Argianas, L., Aritomi, N., Armato, F., Arnaud, N., Arogeti, M., Aronson, S. M., Ashton, G., Aso, Y., Assiduo, M., Melo, S. Assis de Souza, Aston, S. M., Astone, P., Attadio, F., Aubin, F., AultONeal, K., Avallone, G., Babak, S., Badaracco, F., Badger, C., Bae, S., Bagnasco, S., Bagui, E., Baier, J. G., Baiotti, L., Bajpai, R., Baka, T., Ball, M., Ballardin, G., Ballmer, S. W., Banagiri, S., Banerjee, B., Bankar, D., Baral, P., Barayoga, J. C., Barish, B. C., Barker, D., Barneo, P., Barone, F., Barr, B., Barsotti, L., Barsuglia, M., Barta, D., Bartoletti, A. M., Barton, M. A., Bartos, I., Basak, S., Basalaev, A., Bassiri, R., Basti, A., Bates, D. E., Bawaj, M., Baxi, P., Bayley, J. C., Baylor, A. C., Baynard II, P. A., Bazzan, M., Bedakihale, V. M., Beirnaert, F., Bejger, M., Belardinelli, D., Bell, A. S., Benedetto, V., Benoit, W., Bentley, J. D., Yaala, M. Ben, Bera, S., Berbel, M., Bergamin, F., Berger, B. K., Bernuzzi, S., Beroiz, M., Bersanetti, D., Bertolini, A., Betzwieser, J., Beveridge, D., Bevins, N., Bhandare, R., Bhardwaj, U., Bhatt, R., Bhattacharjee, D., Bhaumik, S., Bhowmick, S., Bianchi, A., Bilenko, I. A., Billingsley, G., Binetti, A., Bini, S., Birnholtz, O., Biscoveanu, S., Bisht, A., Bitossi, M., Bizouard, M. -A., Blackburn, J. K., Blagg, L. A., Blair, C. D., Blair, D. G., Bobba, F., Bode, N., Boileau, G., Boldrini, M., Bolingbroke, G. N., Bolliand, A., Bonavena, L. D., Bondarescu, R., Bondu, F., Bonilla, E., Bonilla, M. S., Bonino, A., Bonnand, R., Booker, P., Borchers, A., Boschi, V., Bose, S., Bossilkov, V., Boudart, V., Boudon, A., Bozzi, A., Bradaschia, C., Brady, P. R., Braglia, M., Branch, A., Branchesi, M., Brandt, J., Braun, I., Breschi, M., Briant, T., Brillet, A., Brinkmann, M., Brockill, P., Brockmueller, E., Brooks, A. F., Brown, B. C., Brown, D. D., Brozzetti, M. L., Brunett, S., Bruno, G., Bruntz, R., Bryant, J., Bucci, F., Buchanan, J., Bulashenko, O., Bulik, T., Bulten, H. J., Buonanno, A., Burtnyk, K., Buscicchio, R., Buskulic, D., Buy, C., Byer, R. L., Davies, G. S. Cabourn, Cabras, G., Cabrita, R., Cáceres-Barbosa, V., Cadonati, L., Cagnoli, G., Cahillane, C., Bustillo, J. Calderón, Callister, T. A., Calloni, E., Camp, J. B., Canepa, M., Santoro, G. Caneva, Cannon, K. C., Cao, H., Capistran, L. A., Capocasa, E., Capote, E., Carapella, G., Carbognani, F., Carlassara, M., Carlin, J. B., Carpinelli, M., Carrillo, G., Carter, J. J., Carullo, G., Diaz, J. Casanueva, Casentini, C., Castro-Lucas, S. Y., Caudill, S., Cavaglià, M., Cavalieri, R., Cella, G., Cerdá-Durán, P., Cesarini, E., Chaibi, W., Chakraborty, P., Subrahmanya, S. Chalathadka, Chan, J. C. L., Chan, M., Chandra, K., Chang, R. -J., Chao, S., Charlton, E. L., Charlton, P., Chassande-Mottin, E., Chatterjee, C., Chatterjee, Debarati, Chatterjee, Deep, Chaturvedi, M., Chaty, S., Chen, A., Chen, A. H. -Y., Chen, D., Chen, H., Chen, H. Y., Chen, J., Chen, K. H., Chen, Y., Chen, Yanbei, Chen, Yitian, Cheng, H. P., Chessa, P., Cheung, H. T., Cheung, S. Y., Chiadini, F., Chiarini, G., Chierici, R., Chincarini, A., Chiofalo, M. L., Chiummo, A., Chou, C., Choudhary, S., Christensen, N., Chua, S. S. Y., Chugh, P., Ciani, G., Ciecielag, P., Cieślar, M., Cifaldi, M., Ciolfi, R., Clara, F., Clark, J. A., Clarke, J., Clarke, T. A., Clearwater, P., Clesse, S., Coccia, E., Codazzo, E., Cohadon, P. -F., Colace, S., Colleoni, M., Collette, C. G., Collins, J., Colloms, S., Colombo, A., Colpi, M., Compton, C. M., Connolly, G., Conti, L., Corbitt, T. R., Cordero-Carrión, I., Corezzi, S., Cornish, N. J., Corsi, A., Cortese, S., Costa, C. A., Cottingham, R., Coughlin, M. W., Couineaux, A., Coulon, J. -P., Countryman, S. T., Coupechoux, J. -F., Couvares, P., Coward, D. M., Cowart, M. J., Coyne, R., Craig, K., Creed, R., Creighton, J. D. E., Creighton, T. D., Cremonese, P., Criswell, A. W., Crockett-Gray, J. C. G., Crook, S., Crouch, R., Csizmazia, J., Cudell, J. R., Cullen, T. J., Cumming, A., Cuoco, E., Cusinato, M., Dabadie, P., Canton, T. Dal, Dall'Osso, S., Pra, S. Dal, Dálya, G., D'Angelo, B., Danilishin, S., D'Antonio, S., Danzmann, K., Darroch, K. E., Dartez, L. P., Dasgupta, A., Datta, S., Dattilo, V., Daumas, A., Davari, N., Dave, I., Davenport, A., Davier, M., Davies, T. F., Davis, D., Davis, L., Davis, M. C., Davis, P. J., Dax, M., De Bolle, J., Deenadayalan, M., Degallaix, J., De Laurentis, M., Deléglise, S., De Lillo, F., Dell'Aquila, D., Del Pozzo, W., De Marco, F., De Matteis, F., D'Emilio, V., Demos, N., Dent, T., Depasse, A., DePergola, N., De Pietri, R., De Rosa, R., De Rossi, C., DeSalvo, R., De Simone, R., Dhani, A., Diab, R., Díaz, M. C., Di Cesare, M., Dideron, G., Didio, N. A., Dietrich, T., Di Fiore, L., Di Fronzo, C., Di Giovanni, M., Di Girolamo, T., Diksha, D., Di Michele, A., Ding, J., Di Pace, S., Di Palma, I., Di Renzo, F., Divyajyoti, Dmitriev, A., Doctor, Z., Dohmen, E., Doleva, P. P., Dominguez, D., D'Onofrio, L., Donovan, F., Dooley, K. L., Dooney, T., Doravari, S., Dorosh, O., Drago, M., Driggers, J. C., Ducoin, J. -G., Dunn, L., Dupletsa, U., D'Urso, D., Duval, H., Duverne, P. -A., Dwyer, S. E., Eassa, C., Ebersold, M., Eckhardt, T., Eddolls, G., Edelman, B., Edo, T. B., Edy, O., Effler, A., Eichholz, J., Einsle, H., Eisenmann, M., Eisenstein, R. A., Ejlli, A., Eleveld, R. M., Emma, M., Endo, K., Engl, A. J., Enloe, E., Errico, L., Essick, R. C., Estellés, H., Estevez, D., Etzel, T., Evans, M., Evstafyeva, T., Ewing, B. E., Ezquiaga, J. M., Fabrizi, F., Faedi, F., Fafone, V., Fairhurst, S., Farah, A. M., Farr, B., Farr, W. M., Favaro, G., Favata, M., Fays, M., Fazio, M., Feicht, J., Fejer, M. M., Felicetti, R., Fenyvesi, E., Ferguson, D. L., Ferraiuolo, S., Ferrante, I., Ferreira, T. A., Fidecaro, F., Figura, P., Fiori, A., Fiori, I., Fishbach, M., Fisher, R. P., Fittipaldi, R., Fiumara, V., Flaminio, R., Fleischer, S. M., Fleming, L. S., Floden, E., Foley, E. M., Fong, H., Font, J. A., Fornal, B., Forsyth, P. W. F., Franceschetti, K., Franchini, N., Frasca, S., Frasconi, F., Mascioli, A. Frattale, Frei, Z., Freise, A., Freitas, O., Frey, R., Frischhertz, W., Fritschel, P., Frolov, V. V., Fronzé, G. G., Fuentes-Garcia, M., Fujii, S., Fujimori, T., Fulda, P., Fyffe, M., Gadre, B., Gair, J. R., Galaudage, S., Galdi, V., Gallagher, H., Gallardo, S., Gallego, B., Gamba, R., Gamboa, A., Ganapathy, D., Ganguly, A., Garaventa, B., García-Bellido, J., Núñez, C. García, García-Quirós, C., Gardner, J. W., Gardner, K. A., Gargiulo, J., Garron, A., Garufi, F., Gasbarra, C., Gateley, B., Gayathri, V., Gemme, G., Gennai, A., Gennari, V., George, J., George, R., Gerberding, O., Gergely, L., Ghosh, Archisman, Ghosh, Sayantan, Ghosh, Shaon, Ghosh, Shrobana, Ghosh, Suprovo, Ghosh, Tathagata, Giacoppo, L., Giaime, J. A., Giardina, K. D., Gibson, D. R., Gibson, D. T., Gier, C., Giri, P., Gissi, F., Gkaitatzis, S., Glanzer, J., Glotin, F., Godfrey, J., Godwin, P., Goebbels, N. L., Goetz, E., Golomb, J., Lopez, S. Gomez, Goncharov, B., Gong, Y., González, G., Goodarzi, P., Goode, S., Goodwin-Jones, A. W., Gosselin, M., Göttel, A. S., Gouaty, R., Gould, D. W., Govorkova, K., Goyal, S., Grace, B., Grado, A., Graham, V., Granados, A. E., Granata, M., Granata, V., Gras, S., Grassia, P., Gray, A., Gray, C., Gray, R., Greco, G., Green, A. C., Green, S. M., Green, S. R., Gretarsson, A. M., Gretarsson, E. M., Griffith, D., Griffiths, W. L., Griggs, H. L., Grignani, G., Grimaldi, A., Grimaud, C., Grote, H., Guerra, D., Guetta, D., Guidi, G. M., Guimaraes, A. R., Gulati, H. K., Gulminelli, F., Gunny, A. M., Guo, H., Guo, W., Guo, Y., Gupta, Anchal, Gupta, Anuradha, Gupta, Ish, Gupta, N. C., Gupta, P., Gupta, S. K., Gupta, T., Gupte, N., Gurs, J., Gutierrez, N., Guzman, F., H, H. -Y., Haba, D., Haberland, M., Haino, S., Hall, E. D., Hamilton, E. Z., Hammond, G., Han, W. -B., Haney, M., Hanks, J., Hanna, C., Hannam, M. D., Hannuksela, O. A., Hanselman, A. G., Hansen, H., Hanson, J., Harada, R., Hardison, A. R., Haris, K., Harmark, T., Harms, J., Harry, G. M., Harry, I. W., Hart, J., Haskell, B., Haster, C. -J., Hathaway, J. S., Haughian, K., Hayakawa, H., Hayama, K., Hayes, R., Heffernan, A., Heidmann, A., Heintze, M. C., Heinze, J., Heinzel, J., Heitmann, H., Hellman, F., Hello, P., Helmling-Cornell, A. F., Hemming, G., Henderson-Sapir, O., Hendry, M., Heng, I. S., Hennes, E., Henshaw, C., Hertog, T., Heurs, M., Hewitt, A. L., Heyns, J., Higginbotham, S., Hild, S., Hill, S., Himemoto, Y., Hirata, N., Hirose, C., Hoang, S., Hochheim, S., Hofman, D., Holland, N. A., Holley-Bockelmann, K., Holmes, Z. J., Holz, D. E., Honet, L., Hong, C., Hornung, J., Hoshino, S., Hough, J., Hourihane, S., Howell, E. J., Hoy, C. G., Hrishikesh, C. A., Hsieh, H. -F., Hsiung, C., Hsu, H. C., Hsu, W. -F., Hu, P., Hu, Q., Huang, H. Y., Huang, Y. -J., Huddart, A. D., Hughey, B., Hui, D. C. Y., Hui, V., Husa, S., Huxford, R., Huynh-Dinh, T., Iampieri, L., Iandolo, G. A., Ianni, M., Iess, A., Imafuku, H., Inayoshi, K., Inoue, Y., Iorio, G., Iqbal, M. H., Irwin, J., Ishikawa, R., Isi, M., Ismail, M. A., Itoh, Y., Iwanaga, H., Iwaya, M., Iyer, B. R., JaberianHamedan, V., Jacquet, C., Jacquet, P. -E., Jadhav, S. J., Jadhav, S. P., Jain, T., James, A. L., James, P. A., Jamshidi, R., Janquart, J., Janssens, K., Janthalur, N. N., Jaraba, S., Jaranowski, P., Jaume, R., Javed, W., Jennings, A., Jia, W., Jiang, J., Kubisz, J., Johanson, C., Johns, G. R., Johnson, N. A., Johnston, M. C., Johnston, R., Johny, N., Jones, D. H., Jones, D. I., Jones, R., Jose, S., Joshi, P., Ju, L., Jung, K., Junker, J., Juste, V., Kajita, T., Kaku, I., Kalaghatgi, C., Kalogera, V., Kamiizumi, M., Kanda, N., Kandhasamy, S., Kang, G., Kanner, J. B., Kapadia, S. J., Kapasi, D. P., Karat, S., Karathanasis, C., Kashyap, R., Kasprzack, M., Kastaun, W., Kato, T., Katsavounidis, E., Katzman, W., Kaushik, R., Kawabe, K., Kawamoto, R., Kazemi, A., Keitel, D., Kelley-Derzon, J., Kennington, J., Kesharwani, R., Key, J. S., Khadela, R., Khadka, S., Khalili, F. Y., Khan, F., Khan, I., Khanam, T., Khursheed, M., Khusid, N. M., Kiendrebeogo, W., Kijbunchoo, N., Kim, C., Kim, J. C., Kim, K., Kim, M. H., Kim, S., Kim, Y. -M., Kimball, C., Kinley-Hanlon, M., Kinnear, M., Kissel, J. S., Klimenko, S., Knee, A. M., Knust, N., Kobayashi, K., Obergaulinger, M., Koch, P., Koehlenbeck, S. M., Koekoek, G., Kohri, K., Kokeyama, K., Koley, S., Kolitsidou, P., Kolstein, M., Komori, K., Kong, A. K. H., Kontos, A., Korobko, M., Kossak, R. V., Kou, X., Koushik, A., Kouvatsos, N., Kovalam, M., Kozak, D. B., Kranzhoff, S. L., Kringel, V., Krishnendu, N. V., Królak, A., Kruska, K., Kuehn, G., Kuijer, P., Kulkarni, S., Ramamohan, A. Kulur, Kumar, A., Kumar, Praveen, Kumar, Prayush, Kumar, Rahul, Kumar, Rakesh, Kume, J., Kuns, K., Kuntimaddi, N., Kuroyanagi, S., Kurth, N. J., Kuwahara, S., Kwak, K., Kwan, K., Kwok, J., Lacaille, G., Lagabbe, P., Laghi, D., Lai, S., Laity, A. H., Lakkis, M. H., Lalande, E., Lalleman, M., Lalremruati, P. C., Landry, M., Lane, B. B., Lang, R. N., Lange, J., Lantz, B., La Rana, A., La Rosa, I., Lartaux-Vollard, A., Lasky, P. D., Lawrence, J., Lawrence, M. N., Laxen, M., Lazzarini, A., Lazzaro, C., Leaci, P., Lecoeuche, Y. K., Lee, H. M., Lee, H. W., Lee, K., Lee, R. -K., Lee, R., Lee, S., Lee, Y., Legred, I. N., Lehmann, J., Lehner, L., Jean, M. Le, Lemaître, A., Lenti, M., Leonardi, M., Lequime, M., Leroy, N., Lesovsky, M., Letendre, N., Lethuillier, M., Levin, S. E., Levin, Y., Leyde, K., Li, A. K. Y., Li, K. L., Li, T. G. F., Li, X., Li, Z., Lihos, A., Lin, C-Y., Lin, C. -Y., Lin, E. T., Lin, F., Lin, H., Lin, L. C. -C., Lin, Y. -C., Linde, F., Linker, S. D., Littenberg, T. B., Liu, A., Liu, G. C., Liu, Jian, Villarreal, F. Llamas, Llobera-Querol, J., Lo, R. K. L., Locquet, J. -P., London, L. T., Longo, A., Lopez, D., Portilla, M. Lopez, Lorenzini, M., Lorenzo-Medina, A., Loriette, V., Lormand, M., Losurdo, G., Lott IV, T. P., Lough, J. D., Loughlin, H. A., Lousto, C. O., Lowry, M. J., Lu, N., Lück, H., Lumaca, D., Lundgren, A. P., Lussier, A. W., Ma, L. -T., Ma, S., Ma'arif, M., Macas, R., Macedo, A., MacInnis, M., Maciy, R. R., Macleod, D. M., MacMillan, I. A. O., Macquet, A., Macri, D., Maeda, K., Maenaut, S., Hernandez, I. Magaña, Magare, S. S., Magazzù, C., Magee, R. M., Maggio, E., Maggiore, R., Magnozzi, M., Mahesh, M., Mahesh, S., Maini, M., Majhi, S., Majorana, E., Makarem, C. N., Makelele, E., Malaquias-Reis, J. A., Mali, U., Maliakal, S., Malik, A., Man, N., Mandic, V., Mangano, V., Mannix, B., Mansell, G. L., Mansingh, G., Manske, M., Mantovani, M., Mapelli, M., Marchesoni, F., Pina, D. Marín, Marion, F., Márka, S., Márka, Z., Markosyan, A. S., Markowitz, A., Maros, E., Marsat, S., Martelli, F., Martin, I. W., Martin, R. M., Martinez, B. B., Martinez, M., Martinez, V., Martini, A., Martinovic, K., Martins, J. C., Martynov, D. V., Marx, E. J., Massaro, L., Masserot, A., Masso-Reid, M., Mastrodicasa, M., Mastrogiovanni, S., Matcovich, T., Matiushechkina, M., Matsuyama, M., Mavalvala, N., Maxwell, N., McCarrol, G., McCarthy, R., McClelland, D. E., McCormick, S., McCuller, L., McEachin, S., McElhenny, C., McGhee, G. I., McGinn, J., McGowan, K. B. M., McIver, J., McLeod, A., McRae, T., Meacher, D., Meijer, Q., Melatos, A., Mellaerts, S., Menendez-Vazquez, A., Menoni, C. S., Mera, F., Mercer, R. A., Mereni, L., Merfeld, K., Merilh, E. L., Mérou, J. R., Merritt, J. D., Merzougui, M., Messenger, C., Messick, C., Meyer-Conde, M., Meylahn, F., Mhaske, A., Miani, A., Miao, H., Michaloliakos, I., Michel, C., Michimura, Y., Middleton, H., Miller, A. L., Miller, S., Millhouse, M., Milotti, E., Milotti, V., Minenkov, Y., Mio, N., Mir, Ll. M., Mirasola, L., Miravet-Tenés, M., Miritescu, C. -A., Mishra, A. K., Mishra, A., Mishra, C., Mishra, T., Mitchell, A. L., Mitchell, J. G., Mitra, S., Mitrofanov, V. P., Mittleman, R., Miyakawa, O., Miyamoto, S., Miyoki, S., Mo, G., Mobilia, L., Mohapatra, S. R. P., Mohite, S. R., Molina-Ruiz, M., Mondal, C., Mondin, M., Montani, M., Moore, C. J., Moraru, D., More, A., More, S., Moreno, G., Morgan, C., Morisaki, S., Moriwaki, Y., Morras, G., Moscatello, A., Mourier, P., Mours, B., Mow-Lowry, C. M., Muciaccia, F., Mukherjee, Arunava, Mukherjee, D., Mukherjee, Samanwaya, Mukherjee, Soma, Mukherjee, Subroto, Mukherjee, Suvodip, Mukund, N., Mullavey, A., Munch, J., Mundi, J., Mungioli, C. L., Oberg, W. R. Munn, Murakami, Y., Murakoshi, M., Murray, P. G., Muusse, S., Nabari, D., Nadji, S. L., Nagar, A., Nagarajan, N., Nagler, K. N., Nakagaki, K., Nakamura, K., Nakano, H., Nakano, M., Nandi, D., Napolano, V., Narayan, P., Nardecchia, I., Narikawa, T., Narola, H., Naticchioni, L., Nayak, R. K., Neilson, J., Nelson, A., Nelson, T. J. N., Nery, M., Neunzert, A., Ng, S., Quynh, L. Nguyen, Nichols, S. A., Nielsen, A. B., Nieradka, G., Niko, A., Nishino, Y., Nishizawa, A., Nissanke, S., Nitoglia, E., Niu, W., Nocera, F., Norman, M., North, C., Novak, J., Siles, J. F. Nuño, Nuttall, L. K., Obayashi, K., Oberling, J., O'Dell, J., Oertel, M., Offermans, A., Oganesyan, G., Oh, J. J., Oh, K., O'Hanlon, T., Ohashi, M., Ohkawa, M., Ohme, F., Oliveira, A. S., Oliveri, R., O'Neal, B., Oohara, K., O'Reilly, B., Ormsby, N. D., Orselli, M., O'Shaughnessy, R., O'Shea, S., Oshima, Y., Oshino, S., Ossokine, S., Osthelder, C., Ota, I., Ottaway, D. J., Ouzriat, A., Overmier, H., Owen, B. J., Pace, A. E., Pagano, R., Page, M. A., Pai, A., Pal, A., Pal, S., Palaia, M. A., Pálfi, M., Palma, P. P., Palomba, C., Palud, P., Pan, H., Pan, J., Pan, K. C., Panai, R., Panda, P. K., Pandey, S., Panebianco, L., Pang, P. T. H., Pannarale, F., Pannone, K. A., Pant, B. C., Panther, F. H., Paoletti, F., Paolone, A., Papalexakis, E. E., Papalini, L., Papigkiotis, G., Paquis, A., Parisi, A., Park, B. -J., Park, J., Parker, W., Pascale, G., Pascucci, D., Pasqualetti, A., Passaquieti, R., Passenger, L., Passuello, D., Patane, O., Pathak, D., Pathak, M., Patra, A., Patricelli, B., Patron, A. S., Paul, K., Paul, S., Payne, E., Pearce, T., Pedraza, M., Pegna, R., Pele, A., Arellano, F. E. Peña, Penn, S., Penuliar, M. D., Perego, A., Pereira, Z., Perez, J. J., Périgois, C., Perna, G., Perreca, A., Perret, J., Perriès, S., Perry, J. W., Pesios, D., Petracca, S., Petrillo, C., Pfeiffer, H. P., Pham, H., Pham, K. A., Phukon, K. S., Phurailatpam, H., Piarulli, M., Piccari, L., Piccinni, O. J., Pichot, M., Piendibene, M., Piergiovanni, F., Pierini, L., Pierra, G., Pierro, V., Pietrzak, M., Pillas, M., Pilo, F., Pinard, L., Pinto, I. M., Pinto, M., Piotrzkowski, B. J., Pirello, M., Pitkin, M. D., Placidi, A., Placidi, E., Planas, M. L., Plastino, W., Poggiani, R., Polini, E., Pompili, L., Poon, J., Porcelli, E., Porter, E. K., Posnansky, C., Poulton, R., Powell, J., Pracchia, M., Pradhan, B. K., Pradier, T., Prajapati, A. K., Prasai, K., Prasanna, R., Prasia, P., Pratten, G., Principe, G., Principe, M., Prodi, G. A., Prokhorov, L., Prosposito, P., Puecher, A., Pullin, J., Punturo, M., Puppo, P., Pürrer, M., Qi, H., Qin, J., Quéméner, G., Quetschke, V., Quigley, C., Quinonez, P. J., Raab, F. J., Raabith, S. S., Raaijmakers, G., Raja, S., Rajan, C., Rajbhandari, B., Ramirez, K. E., Vidal, F. A. Ramis, Ramos-Buades, A., Rana, D., Ranjan, S., Ransom, K., Rapagnani, P., Ratto, B., Rawat, S., Ray, A., Raymond, V., Razzano, M., Read, J., Payo, M. Recaman, Regimbau, T., Rei, L., Reid, S., Reitze, D. H., Relton, P., Renzini, A. I., Rettegno, P., Revenu, B., Reyes, R., Rezaei, A. S., Ricci, F., Ricci, M., Ricciardone, A., Richardson, J. W., Richardson, M., Rijal, A., Riles, K., Riley, H. K., Rinaldi, S., Rittmeyer, J., Robertson, C., Robinet, F., Robinson, M., Rocchi, A., Rolland, L., Rollins, J. G., Romano, A. E., Romano, R., Romero, A., Romero-Shaw, I. M., Romie, J. H., Ronchini, S., Roocke, T. J., Rosa, L., Rosauer, T. J., Rose, C. A., Rosińska, D., Ross, M. P., Rossello, M., Rowan, S., Roy, S. K., Roy, S., Rozza, D., Ruggi, P., Ruhama, N., Morales, E. Ruiz, Ruiz-Rocha, K., Sachdev, S., Sadecki, T., Sadiq, J., Saffarieh, P., Sah, M. R., Saha, S. S., Saha, S., Sainrat, T., Menon, S. Sajith, Sakai, K., Sakellariadou, M., Sakon, S., Salafia, O. S., Salces-Carcoba, F., Salconi, L., Saleem, M., Salemi, F., Sallé, M., Salvador, S., Sanchez, A., Sanchez, E. J., Sanchez, J. H., Sanchez, L. E., Sanchis-Gual, N., Sanders, J. R., Sänger, E. M., Santoliquido, F., Saravanan, T. R., Sarin, N., Sasaoka, S., Sasli, A., Sassi, P., Sassolas, B., Satari, H., Sato, R., Sato, Y., Sauter, O., Savage, R. L., Sawada, T., Sawant, H. L., Sayah, S., Scacco, V., Schaetzl, D., Scheel, M., Schiebelbein, A., Schiworski, M. G., Schmidt, P., Schmidt, S., Schnabel, R., Schneewind, M., Schofield, R. M. S., Schouteden, K., Schulte, B. W., Schutz, B. F., Schwartz, E., Scialpi, M., Scott, J., Scott, S. M., Seetharamu, T. C., Seglar-Arroyo, M., Sekiguchi, Y., Sellers, D., Sengupta, A. S., Sentenac, D., Seo, E. G., Seo, J. W., Sequino, V., Serra, M., Servignat, G., Sevrin, A., Shaffer, T., Shah, U. S., Shaikh, M. A., Shao, L., Sharma, A. K., Sharma, P., Sharma-Chaudhary, S., Shaw, M. R., Shawhan, P., Shcheblanov, N. S., Sheridan, E., Shikano, Y., Shikauchi, M., Shimode, K., Shinkai, H., Shiota, J., Shoemaker, D. H., Shoemaker, D. M., Short, R. W., ShyamSundar, S., Sider, A., Siegel, H., Sieniawska, M., Sigg, D., Silenzi, L., Simmonds, M., Singer, L. P., Singh, A., Singh, D., Singh, M. K., Singh, S., Singha, A., Sintes, A. M., Sipala, V., Skliris, V., Slagmolen, B. J. J., Slaven-Blair, T. J., Smetana, J., Smith, J. R., Smith, L., Smith, R. J. E., Smith, W. J., Soldateschi, J., Somiya, K., Song, I., Soni, K., Soni, S., Sordini, V., Sorrentino, F., Sorrentino, N., Sotani, H., Soulard, R., Southgate, A., Spagnuolo, V., Spencer, A. P., Spera, M., Spinicelli, P., Spoon, J. B., Sprague, C. A., Srivastava, A. K., Stachurski, F., Steer, D. A., Steinlechner, J., Steinlechner, S., Stergioulas, N., Stevens, P., StPierre, M., Stratta, G., Strong, M. D., Strunk, A., Sturani, R., Stuver, A. L., Suchenek, M., Sudhagar, S., Sueltmann, N., Suleiman, L., Sullivan, K. D., Sun, L., Sunil, S., Suresh, J., Sutton, P. J., Suzuki, T., Suzuki, Y., Swinkels, B. L., Syx, A., Szczepańczyk, M. J., Szewczyk, P., Tacca, M., Tagoshi, H., Tait, S. C., Takahashi, H., Takahashi, R., Takamori, A., Takase, T., Takatani, K., Takeda, H., Takeshita, K., Talbot, C., Tamaki, M., Tamanini, N., Tanabe, D., Tanaka, K., Tanaka, S. J., Tanaka, T., Tang, D., Tanioka, S., Tanner, D. B., Tao, L., Tapia, R. D., Martín, E. N. Tapia San, Tarafder, R., Taranto, C., Taruya, A., Tasson, J. D., Teloi, M., Tenorio, R., Themann, H., Theodoropoulos, A., Thirugnanasambandam, M. P., Thomas, L. M., Thomas, M., Thomas, P., Thompson, J. E., Thondapu, S. R., Thorne, K. A., Thrane, E., Tissino, J., Tiwari, A., Tiwari, P., Tiwari, S., Tiwari, V., Todd, M. R., Toivonen, A. M., Toland, K., Tolley, A. E., Tomaru, T., Tomita, K., Tomura, T., Tong-Yu, C., Toriyama, A., Toropov, N., Torres-Forné, A., Torrie, C. I., Toscani, M., Melo, I. Tosta e, Tournefier, E., Trapananti, A., Travasso, F., Traylor, G., Trevor, M., Tringali, M. C., Tripathee, A., Troian, G., Troiano, L., Trovato, A., Trozzo, L., Trudeau, R. J., Tsang, T. T. L., Tso, R., Tsuchida, S., Tsukada, L., Tsutsui, T., Turbang, K., Turconi, M., Turski, C., Ubach, H., Uchikata, N., Uchiyama, T., Udall, R. P., Uehara, T., Uematsu, M., Ueno, K., Ueno, S., Undheim, V., Ushiba, T., Vacatello, M., Vahlbruch, H., Vaidya, N., Vajente, G., Vajpeyi, A., Valdes, G., Valencia, J., Valentini, M., Vallejo-Peña, S. A., Vallero, S., Valsan, V., van Bakel, N., van Beuzekom, M., van Dael, M., Brand, J. F. J. van den, Broeck, C. Van Den, Vander-Hyde, D. C., van der Sluys, M., Van de Walle, A., van Dongen, J., Vandra, K., van Haevermaet, H., van Heijningen, J. V., Van Hove, P., VanKeuren, M., Vanosky, J., van Putten, M. H. P. M., van Ranst, Z., van Remortel, N., Vardaro, M., Vargas, A. F., Varghese, J. J., Varma, V., Vasúth, M., Vecchio, A., Vedovato, G., Veitch, J., Veitch, P. J., Venikoudis, S., Venneberg, J., Verdier, P., Verkindt, D., Verma, B., Verma, P., Verma, Y., Vermeulen, S. M., Vetrano, F., Veutro, A., Vibhute, A. M., Viceré, A., Vidyant, S., Viets, A. D., Vijaykumar, A., Vilkha, A., Villa-Ortega, V., Vincent, E. T., Vinet, J. -Y., Viret, S., Virtuoso, A., Vitale, S., Vives, A., Vocca, H., Voigt, D., von Reis, E. R. G., von Wrangel, J. S. A., Vyatchanin, S. P., Wade, L. E., Wade, M., Wagner, K. J., Wajid, A., Walker, M., Wallace, G. S., Wallace, L., Wang, H., Wang, J. Z., Wang, W. H., Wang, Z., Waratkar, G., Warner, J., Was, M., Washimi, T., Washington, N. Y., Watarai, D., Wayt, K. E., Weaver, B. R., Weaver, B., Weaving, C. R., Webster, S. A., Weinert, M., Weinstein, A. J., Weiss, R., Wellmann, F., Wen, L., Weßels, P., Wette, K., Whelan, J. T., Whiting, B. F., Whittle, C., Wildberger, J. B., Wilk, O. S., Wilken, D., Wilkin, A. T., Willadsen, D. J., Willetts, K., Williams, D., Williams, M. J., Williams, N. S., Willis, J. L., Willke, B., Wils, M., Winterflood, J., Wipf, C. C., Woan, G., Woehler, J., Wofford, J. K., Wolfe, N. E., Wong, H. T., Wong, H. W. Y., Wong, I. C. F., Wright, J. L., Wright, M., Wu, C., Wu, D. S., Wu, H., Wuchner, E., Wysocki, D. M., Xu, V. A., Xu, Y., Yadav, N., Yamamoto, H., Yamamoto, K., Yamamoto, T. S., Yamamoto, T., Yamamura, S., Yamazaki, R., Yan, S., Yan, T., Yang, F. W., Yang, F., Yang, K. Z., Yang, Y., Yarbrough, Z., Yasui, H., Yeh, S. -W., Yelikar, A. B., Yin, X., Yokoyama, J., Yokozawa, T., Yoo, J., Yu, H., Yuan, S., Yuzurihara, H., Zadrożny, A., Zanolin, M., Zeeshan, M., Zelenova, T., Zendri, J. -P., Zeoli, M., Zerrad, M., Zevin, M., Zhang, A. C., Zhang, L., Zhang, R., Zhang, T., Zhang, Y., Zhao, C., Zhao, Yue, Zhao, Yuhang, Zheng, Y., Zhong, H., Zhou, R., Zhu, X. -J., Zhu, Z. -H., Zimmerman, A. B., Zucker, M. E., and Zweizig, J.
- Subjects
Astrophysics - High Energy Astrophysical Phenomena - Abstract
We present the results of a search for gravitational-wave transients associated with core-collapse supernova SN 2023ixf, which was observed in the galaxy Messier 101 via optical emission on 2023 May 19th, during the LIGO-Virgo-KAGRA 15th Engineering Run. We define a five-day on-source window during which an accompanying gravitational-wave signal may have occurred. No gravitational waves have been identified in data when at least two gravitational-wave observatories were operating, which covered $\sim 14\%$ of this five-day window. We report the search detection efficiency for various possible gravitational-wave emission models. Considering the distance to M101 (6.7 Mpc), we derive constraints on the gravitational-wave emission mechanism of core-collapse supernovae across a broad frequency spectrum, ranging from 50 Hz to 2 kHz where we assume the GW emission occurred when coincident data are available in the on-source window. Considering an ellipsoid model for a rotating proto-neutron star, our search is sensitive to gravitational-wave energy $1 \times 10^{-5} M_{\odot} c^2$ and luminosity $4 \times 10^{-5} M_{\odot} c^2/\text{s}$ for a source emitting at 50 Hz. These constraints are around an order of magnitude more stringent than those obtained so far with gravitational-wave data. The constraint on the ellipticity of the proto-neutron star that is formed is as low as $1.04$, at frequencies above $1200$ Hz, surpassing results from SN 2019ejj., Comment: Main paper: 6 pages, 4 figures and 1 table. Total with appendices: 20 pages, 4 figures, and 1 table
- Published
- 2024
15. A search using GEO600 for gravitational waves coincident with fast radio bursts from SGR 1935+2154
- Author
-
The LIGO Scientific Collaboration, the Virgo Collaboration, the KAGRA Collaboration, Abac, A. G., Abbott, R., Abouelfettouh, I., Acernese, F., Ackley, K., Adhicary, S., Adhikari, N., Adhikari, R. X., Adkins, V. K., Agarwal, D., Agathos, M., Abchouyeh, M. Aghaei, Aguiar, O. D., Aguilar, I., Aiello, L., Ain, A., Ajith, P., Akutsu, T., Albanesi, S., Alfaidi, R. A., Al-Jodah, A., Alléné, C., Allocca, A., Al-Shammari, S., Altin, P. A., Alvarez-Lopez, S., Amato, A., Amez-Droz, L., Amorosi, A., Amra, C., Ananyeva, A., Anderson, S. B., Anderson, W. G., Andia, M., Ando, M., Andrade, T., Andres, N., Andrés-Carcasona, M., Andrić, T., Anglin, J., Ansoldi, S., Antelis, J. M., Antier, S., Aoumi, M., Appavuravther, E. Z., Appert, S., Apple, S. K., Arai, K., Araya, A., Araya, M. C., Areeda, J. S., Argianas, L., Aritomi, N., Armato, F., Arnaud, N., Arogeti, M., Aronson, S. M., Ashton, G., Aso, Y., Assiduo, M., Melo, S. Assis de Souza, Aston, S. M., Astone, P., Attadio, F., Aubin, F., AultONeal, K., Avallone, G., Azrad, D., Babak, S., Badaracco, F., Badger, C., Bae, S., Bagnasco, S., Bagui, E., Baier, J. G., Baiotti, L., Bajpai, R., Baka, T., Ball, M., Ballardin, G., Ballmer, S. W., Banagiri, S., Banerjee, B., Bankar, D., Baral, P., Barayoga, J. C., Barish, B. C., Barker, D., Barneo, P., Barone, F., Barr, B., Barsotti, L., Barsuglia, M., Barta, D., Bartoletti, A. M., Barton, M. A., Bartos, I., Basak, S., Basalaev, A., Bassiri, R., Basti, A., Bates, D. E., Bawaj, M., Baxi, P., Bayley, J. C., Baylor, A. C., Baynard II, P. A., Bazzan, M., Bedakihale, V. M., Beirnaert, F., Bejger, M., Belardinelli, D., Bell, A. S., Benedetto, V., Benoit, W., Bentley, J. D., Yaala, M. Ben, Bera, S., Berbel, M., Bergamin, F., Berger, B. K., Bernuzzi, S., Beroiz, M., Bersanetti, D., Bertolini, A., Betzwieser, J., Beveridge, D., Bevins, N., Bhandare, R., Bhardwaj, U., Bhatt, R., Bhattacharjee, D., Bhaumik, S., Bhowmick, S., Bianchi, A., Bilenko, I. A., Billingsley, G., Binetti, A., Bini, S., Birnholtz, O., Biscoveanu, S., Bisht, A., Bitossi, M., Bizouard, M. -A., Blackburn, J. K., Blagg, L. A., Blair, C. D., Blair, D. G., Bobba, F., Bode, N., Boileau, G., Boldrini, M., Bolingbroke, G. N., Bolliand, A., Bonavena, L. D., Bondarescu, R., Bondu, F., Bonilla, E., Bonilla, M. S., Bonino, A., Bonnand, R., Booker, P., Borchers, A., Boschi, V., Bose, S., Bossilkov, V., Boudart, V., Boudon, A., Bozzi, A., Bradaschia, C., Brady, P. R., Braglia, M., Branch, A., Branchesi, M., Brandt, J., Braun, I., Breschi, M., Briant, T., Brillet, A., Brinkmann, M., Brockill, P., Brockmueller, E., Brooks, A. F., Brown, B. C., Brown, D. D., Brozzetti, M. L., Brunett, S., Bruno, G., Bruntz, R., Bryant, J., Bucci, F., Buchanan, J., Bulashenko, O., Bulik, T., Bulten, H. J., Buonanno, A., Burtnyk, K., Buscicchio, R., Buskulic, D., Buy, C., Byer, R. L., Davies, G. S. Cabourn, Cabras, G., Cabrita, R., Cáceres-Barbosa, V., Cadonati, L., Cagnoli, G., Cahillane, C., Bustillo, J. Calderón, Callister, T. A., Calloni, E., Camp, J. B., Canepa, M., Santoro, G. Caneva, Cannon, K. C., Cao, H., Capistran, L. A., Capocasa, E., Capote, E., Carapella, G., Carbognani, F., Carlassara, M., Carlin, J. B., Carpinelli, M., Carrillo, G., Carter, J. J., Carullo, G., Diaz, J. Casanueva, Casentini, C., Castro-Lucas, S. Y., Caudill, S., Cavaglià, M., Cavalieri, R., Cella, G., Cerdá-Durán, P., Cesarini, E., Chaibi, W., Chakraborty, P., Subrahmanya, S. Chalathadka, Chan, J. C. L., Chan, M., Chandra, K., Chang, R. -J., Chao, S., Charlton, E. L., Charlton, P., Chassande-Mottin, E., Chatterjee, C., Chatterjee, Debarati, Chatterjee, Deep, Chaturvedi, M., Chaty, S., Chen, A., Chen, A. H. -Y., Chen, D., Chen, H., Chen, H. Y., Chen, J., Chen, K. H., Chen, Y., Chen, Yanbei, Chen, Yitian, Cheng, H. P., Chessa, P., Cheung, H. T., Cheung, S. Y., Chiadini, F., Chiarini, G., Chierici, R., Chincarini, A., Chiofalo, M. L., Chiummo, A., Chou, C., Choudhary, S., Christensen, N., Chua, S. S. Y., Chugh, P., Ciani, G., Ciecielag, P., Cieślar, M., Cifaldi, M., Ciolfi, R., Clara, F., Clark, J. A., Clarke, J., Clarke, T. A., Clearwater, P., Clesse, S., Coccia, E., Codazzo, E., Cohadon, P. -F., Colace, S., Colleoni, M., Collette, C. G., Collins, J., Colloms, S., Colombo, A., Colpi, M., Compton, C. M., Connolly, G., Conti, L., Corbitt, T. R., Cordero-Carrión, I., Corezzi, S., Cornish, N. J., Corsi, A., Cortese, S., Costa, C. A., Cottingham, R., Coughlin, M. W., Couineaux, A., Coulon, J. -P., Countryman, S. T., Coupechoux, J. -F., Couvares, P., Coward, D. M., Cowart, M. J., Coyne, R., Craig, K., Creed, R., Creighton, J. D. E., Creighton, T. D., Cremonese, P., Criswell, A. W., Crockett-Gray, J. C. G., Crook, S., Crouch, R., Csizmazia, J., Cudell, J. R., Cullen, T. J., Cumming, A., Cuoco, E., Cusinato, M., Dabadie, P., Canton, T. Dal, Dall'Osso, S., Pra, S. Dal, Dálya, G., D'Angelo, B., Danilishin, S., D'Antonio, S., Danzmann, K., Darroch, K. E., Dartez, L. P., Dasgupta, A., Datta, S., Dattilo, V., Daumas, A., Davari, N., Dave, I., Davenport, A., Davier, M., Davies, T. F., Davis, D., Davis, L., Davis, M. C., Davis, P. J., Dax, M., De Bolle, J., Deenadayalan, M., Degallaix, J., De Laurentis, M., Deléglise, S., De Lillo, F., Dell'Aquila, D., Del Pozzo, W., De Marco, F., De Matteis, F., D'Emilio, V., Demos, N., Dent, T., Depasse, A., DePergola, N., De Pietri, R., De Rosa, R., De Rossi, C., DeSalvo, R., De Simone, R., Dhani, A., Diab, R., Díaz, M. C., Di Cesare, M., Dideron, G., Didio, N. A., Dietrich, T., Di Fiore, L., Di Fronzo, C., Di Giovanni, M., Di Girolamo, T., Diksha, D., Di Michele, A., Ding, J., Di Pace, S., Di Palma, I., Di Renzo, F., Divyajyoti, Dmitriev, A., Doctor, Z., Dohmen, E., Doleva, P. P., Dominguez, D., D'Onofrio, L., Donovan, F., Dooley, K. L., Dooney, T., Doravari, S., Dorosh, O., Drago, M., Driggers, J. C., Ducoin, J. -G., Dunn, L., Dupletsa, U., D'Urso, D., Duval, H., Duverne, P. -A., Dwyer, S. E., Eassa, C., Ebersold, M., Eckhardt, T., Eddolls, G., Edelman, B., Edo, T. B., Edy, O., Effler, A., Eichholz, J., Einsle, H., Eisenmann, M., Eisenstein, R. A., Ejlli, A., Eleveld, R. M., Emma, M., Endo, K., Engl, A. J., Enloe, E., Errico, L., Essick, R. C., Estellés, H., Estevez, D., Etzel, T., Evans, M., Evstafyeva, T., Ewing, B. E., Ezquiaga, J. M., Fabrizi, F., Faedi, F., Fafone, V., Fairhurst, S., Farah, A. M., Farr, B., Farr, W. M., Favaro, G., Favata, M., Fays, M., Fazio, M., Feicht, J., Fejer, M. M., Felicetti, R. ., Fenyvesi, E., Ferguson, D. L., Ferraiuolo, S., Ferrante, I., Ferreira, T. A., Fidecaro, F., Figura, P., Fiori, A., Fiori, I., Fishbach, M., Fisher, R. P., Fittipaldi, R., Fiumara, V., Flaminio, R., Fleischer, S. M., Fleming, L. S., Floden, E., Foley, E. M., Fong, H., Font, J. A., Fornal, B., Forsyth, P. W. F., Franceschetti, K., Franchini, N., Frasca, S., Frasconi, F., Mascioli, A. Frattale, Frei, Z., Freise, A., Freitas, O., Frey, R., Frischhertz, W., Fritschel, P., Frolov, V. V., Fronzé, G. G., Fuentes-Garcia, M., Fujii, S., Fujimori, T., Fulda, P., Fyffe, M., Gadre, B., Gair, J. R., Galaudage, S., Galdi, V., Gallagher, H., Gallardo, S., Gallego, B., Gamba, R., Gamboa, A., Ganapathy, D., Ganguly, A., Garaventa, B., García-Bellido, J., Núñez, C. García, García-Quirós, C., Gardner, J. W., Gardner, K. A., Gargiulo, J., Garron, A., Garufi, F., Gasbarra, C., Gateley, B., Gayathri, V., Gemme, G., Gennai, A., Gennari, V., George, J., George, R., Gerberding, O., Gergely, L., Ghonge, S., Ghosh, Archisman, Ghosh, Sayantan, Ghosh, Shaon, Ghosh, Shrobana, Ghosh, Suprovo, Ghosh, Tathagata, Giacoppo, L., Giaime, J. A., Giardina, K. D., Gibson, D. R., Gibson, D. T., Gier, C., Giri, P., Gissi, F., Gkaitatzis, S., Glanzer, J., Glotin, F., Godfrey, J., Godwin, P., Goebbels, N. L., Goetz, E., Golomb, J., Lopez, S. Gomez, Goncharov, B., Gong, Y., González, G., Goodarzi, P., Goode, S., Goodwin-Jones, A. W., Gosselin, M., Göttel, A. S., Gouaty, R., Gould, D. W., Govorkova, K., Goyal, S., Grace, B., Grado, A., Graham, V., Granados, A. E., Granata, M., Granata, V., Gras, S., Grassia, P., Gray, A., Gray, C., Gray, R., Greco, G., Green, A. C., Green, S. M., Green, S. R., Gretarsson, A. M., Gretarsson, E. M., Griffith, D., Griffiths, W. L., Griggs, H. L., Grignani, G., Grimaldi, A., Grimaud, C., Grote, H., Guerra, D., Guetta, D., Guidi, G. M., Guimaraes, A. R., Gulati, H. K., Gulminelli, F., Gunny, A. M., Guo, H., Guo, W., Guo, Y., Gupta, Anchal, Gupta, Anuradha, Gupta, Ish, Gupta, N. C., Gupta, P., Gupta, S. K., Gupta, T., Gupte, N., Gurs, J., Gutierrez, N., Guzman, F., H, H. -Y., Haba, D., Haberland, M., Haino, S., Hall, E. D., Hamilton, E. Z., Hammond, G., Han, W. -B., Haney, M., Hanks, J., Hanna, C., Hannam, M. D., Hannuksela, O. A., Hanselman, A. G., Hansen, H., Hanson, J., Harada, R., Hardison, A. R., Haris, K., Harmark, T., Harms, J., Harry, G. M., Harry, I. W., Hart, J., Haskell, B., Haster, C. -J., Hathaway, J. S., Haughian, K., Hayakawa, H., Hayama, K., Hayes, R., Heffernan, A., Heidmann, A., Heintze, M. C., Heinze, J., Heinzel, J., Heitmann, H., Hellman, F., Hello, P., Helmling-Cornell, A. F., Hemming, G., Henderson-Sapir, O., Hendry, M., Heng, I. S., Hennes, E., Henshaw, C., Hertog, T., Heurs, M., Hewitt, A. L., Heyns, J., Higginbotham, S., Hild, S., Hill, S., Himemoto, Y., Hirata, N., Hirose, C., Ho, W. C. G., Hoang, S., Hochheim, S., Hofman, D., Holland, N. A., Holley-Bockelmann, K., Holmes, Z. J., Holz, D. E., Honet, L., Hong, C., Hornung, J., Hoshino, S., Hough, J., Hourihane, S., Howell, E. J., Hoy, C. G., Hrishikesh, C. A., Hsieh, H. -F., Hsiung, C., Hsu, H. C., Hsu, W. -F., Hu, P., Hu, Q., Huang, H. Y., Huang, Y. -J., Huddart, A. D., Hughey, B., Hui, D. C. Y., Hui, V., Husa, S., Huxford, R., Huynh-Dinh, T., Iampieri, L., Iandolo, G. A., Ianni, M., Iess, A., Imafuku, H., Inayoshi, K., Inoue, Y., Iorio, G., Iqbal, M. H., Irwin, J., Ishikawa, R., Isi, M., Ismail, M. A., Itoh, Y., Iwanaga, H., Iwaya, M., Iyer, B. R., JaberianHamedan, V., Jacquet, C., Jacquet, P. -E., Jadhav, S. J., Jadhav, S. P., Jain, T., James, A. L., James, P. A., Jamshidi, R., Janquart, J., Janssens, K., Janthalur, N. N., Jaraba, S., Jaranowski, P., Jaume, R., Javed, W., Jennings, A., Jia, W., Jiang, J., Kubisz, J., Johanson, C., Johns, G. R., Johnson, N. A., Johnston, M. C., Johnston, R., Johny, N., Jones, D. H., Jones, D. I., Jones, R., Jose, S., Joshi, P., Ju, L., Jung, K., Junker, J., Juste, V., Kajita, T., Kaku, I., Kalaghatgi, C., Kalogera, V., Kamiizumi, M., Kanda, N., Kandhasamy, S., Kang, G., Kanner, J. B., Kapadia, S. J., Kapasi, D. P., Karat, S., Karathanasis, C., Kashyap, R., Kasprzack, M., Kastaun, W., Kato, T., Katsavounidis, E., Katzman, W., Kaushik, R., Kawabe, K., Kawamoto, R., Kazemi, A., Keitel, D., Kelley-Derzon, J., Kennington, J., Kesharwani, R., Key, J. S., Khadela, R., Khadka, S., Khalili, F. Y., Khan, F., Khan, I., Khanam, T., Khursheed, M., Khusid, N. M., Kiendrebeogo, W., Kijbunchoo, N., Kim, C., Kim, J. C., Kim, K., Kim, M. H., Kim, S., Kim, Y. -M., Kimball, C., Kinley-Hanlon, M., Kinnear, M., Kissel, J. S., Klimenko, S., Knee, A. M., Knust, N., Kobayashi, K., Koch, P., Koehlenbeck, S. M., Koekoek, G., Kohri, K., Kokeyama, K., Koley, S., Kolitsidou, P., Kolstein, M., Komori, K., Kong, A. K. H., Kontos, A., Korobko, M., Kossak, R. V., Kou, X., Koushik, A., Kouvatsos, N., Kovalam, M., Kozak, D. B., Kranzhoff, S. L., Kringel, V., Krishnendu, N. V., Królak, A., Kruska, K., Kuehn, G., Kuijer, P., Kulkarni, S., Ramamohan, A. Kulur, Kumar, A., Kumar, Praveen, Kumar, Prayush, Kumar, Rahul, Kumar, Rakesh, Kume, J., Kuns, K., Kuntimaddi, N., Kuroyanagi, S., Kurth, N. J., Kuwahara, S., Kwak, K., Kwan, K., Kwok, J., Lacaille, G., Lagabbe, P., Laghi, D., Lai, S., Laity, A. H., Lakkis, M. H., Lalande, E., Lalleman, M., Lalremruati, P. C., Landry, M., Lane, B. B., Lang, R. N., Lange, J., Lantz, B., La Rana, A., La Rosa, I., Lartaux-Vollard, A., Lasky, P. D., Lawrence, J., Lawrence, M. N., Laxen, M., Lazzarini, A., Lazzaro, C., Leaci, P., Lecoeuche, Y. K., Lee, H. M., Lee, H. W., Lee, K., Lee, R. -K., Lee, R., Lee, S., Lee, Y., Legred, I. N., Lehmann, J., Lehner, L., Jean, M. Le, Lemaître, A., Lenti, M., Leonardi, M., Lequime, M., Leroy, N., Lesovsky, M., Letendre, N., Lethuillier, M., Levin, S. E., Levin, Y., Leyde, K., Li, A. K. Y., Li, K. L., Li, T. G. F., Li, X., Li, Z., Lihos, A., Lin, C-Y., Lin, C. -Y., Lin, E. T., Lin, F., Lin, H., Lin, L. C. -C., Lin, Y. -C., Linde, F., Linker, S. D., Littenberg, T. B., Liu, A., Liu, G. C., Liu, Jian, Villarreal, F. Llamas, Llobera-Querol, J., Lo, R. K. L., Locquet, J. -P., London, L. T., Longo, A., Lopez, D., Portilla, M. Lopez, Lorenzini, M., Lorenzo-Medina, A., Loriette, V., Lormand, M., Losurdo, G., Lott IV, T. P., Lough, J. D., Loughlin, H. A., Lousto, C. O., Lowry, M. J., Lu, N., Lück, H., Lumaca, D., Lundgren, A. P., Lussier, A. W., Ma, L. -T., Ma, S., Ma'arif, M., Macas, R., Macedo, A., MacInnis, M., Maciy, R. R., Macleod, D. M., MacMillan, I. A. O., Macquet, A., Macri, D., Maeda, K., Maenaut, S., Hernandez, I. Magaña, Magare, S. S., Magazzù, C., Magee, R. M., Maggio, E., Maggiore, R., Magnozzi, M., Mahesh, M., Mahesh, S., Maini, M., Majhi, S., Majorana, E., Makarem, C. N., Makelele, E., Malaquias-Reis, J. A., Mali, U., Maliakal, S., Malik, A., Man, N., Mandic, V., Mangano, V., Mannix, B., Mansell, G. L., Mansingh, G., Manske, M., Mantovani, M., Mapelli, M., Marchesoni, F., Pina, D. Marín, Marion, F., Márka, S., Márka, Z., Markosyan, A. S., Markowitz, A., Maros, E., Marsat, S., Martelli, F., Martin, I. W., Martin, R. M., Martinez, B. B., Martinez, M., Martinez, V., Martini, A., Martinovic, K., Martins, J. C., Martynov, D. V., Marx, E. J., Massaro, L., Masserot, A., Masso-Reid, M., Mastrodicasa, M., Mastrogiovanni, S., Matcovich, T., Matiushechkina, M., Matsuyama, M., Mavalvala, N., Maxwell, N., McCarrol, G., McCarthy, R., McCormick, S., McCuller, L., McEachin, S., McElhenny, C., McGhee, G. I., McGinn, J., McGowan, K. B. M., McIver, J., McLeod, A., McRae, T., Meacher, D., Meijer, Q., Melatos, A., Mellaerts, S., Menendez-Vazquez, A., Menoni, C. S., Mera, F., Mercer, R. A., Mereni, L., Merfeld, K., Merilh, E. L., Mérou, J. R., Merritt, J. D., Merzougui, M., Messenger, C., Messick, C., Meyer-Conde, M., Meylahn, F., Mhaske, A., Miani, A., Miao, H., Michaloliakos, I., Michel, C., Michimura, Y., Middleton, H., Miller, A. L., Miller, S., Millhouse, M., Milotti, E., Milotti, V., Minenkov, Y., Mio, N., Mir, Ll. M., Mirasola, L., Miravet-Tenés, M., Miritescu, C. -A., Mishra, A. K., Mishra, A., Mishra, C., Mishra, T., Mitchell, A. L., Mitchell, J. G., Mitra, S., Mitrofanov, V. P., Mittleman, R., Miyakawa, O., Miyamoto, S., Miyoki, S., Mo, G., Mobilia, L., Mohapatra, S. R. P., Mohite, S. R., Molina-Ruiz, M., Mondal, C., Mondin, M., Montani, M., Moore, C. J., Moraru, D., More, A., More, S., Moreno, G., Morgan, C., Morisaki, S., Moriwaki, Y., Morras, G., Moscatello, A., Mourier, P., Mours, B., Mow-Lowry, C. M., Muciaccia, F., Mukherjee, Arunava, Mukherjee, D., Mukherjee, Samanwaya, Mukherjee, Soma, Mukherjee, Subroto, Mukherjee, Suvodip, Mukund, N., Mullavey, A., Munch, J., Mundi, J., Mungioli, C. L., Oberg, W. R. Munn, Murakami, Y., Murakoshi, M., Murray, P. G., Muusse, S., Nabari, D., Nadji, S. L., Nagar, A., Nagarajan, N., Nagler, K. N., Nakagaki, K., Nakamura, K., Nakano, H., Nakano, M., Nandi, D., Napolano, V., Narayan, P., Nardecchia, I., Narola, H., Naticchioni, L., Nayak, R. K., Neilson, J., Nelson, A., Nelson, T. J. N., Nery, M., Neunzert, A., Ng, S., Quynh, L. Nguyen, Nichols, S. A., Nielsen, A. B., Nieradka, G., Niko, A., Nishino, Y., Nishizawa, A., Nissanke, S., Nitoglia, E., Niu, W., Nocera, F., Norman, M., North, C., Novak, J., Siles, J. F. Nuño, Nuttall, L. K., Obayashi, K., Oberling, J., O'Dell, J., Oertel, M., Offermans, A., Oganesyan, G., Oh, J. J., Oh, K., O'Hanlon, T., Ohashi, M., Ohkawa, M., Ohme, F., Oliveira, A. S., Oliveri, R., O'Neal, B., Oohara, K., O'Reilly, B., Ormsby, N. D., Orselli, M., O'Shaughnessy, R., O'Shea, S., Oshima, Y., Oshino, S., Ossokine, S., Osthelder, C., Ota, I., Ottaway, D. J., Ouzriat, A., Overmier, H., Owen, B. J., Pace, A. E., Pagano, R., Page, M. A., Pai, A., Pal, A., Pal, S., Palaia, M. A., Pálfi, M., Palma, P. P., Palomba, C., Palud, P., Pan, H., Pan, J., Pan, K. C., Panai, R., Panda, P. K., Pandey, S., Panebianco, L., Pang, P. T. H., Pannarale, F., Pannone, K. A., Pant, B. C., Panther, F. H., Paoletti, F., Paolone, A., Papalexakis, E. E., Papalini, L., Papigkiotis, G., Paquis, A., Parisi, A., Park, B. -J., Park, J., Parker, W., Pascale, G., Pascucci, D., Pasqualetti, A., Passaquieti, R., Passenger, L., Passuello, D., Patane, O., Pathak, D., Pathak, M., Patra, A., Patricelli, B., Patron, A. S., Paul, K., Paul, S., Payne, E., Pearce, T., Pedraza, M., Pegna, R., Pele, A., Arellano, F. E. Peña, Penn, S., Penuliar, M. D., Perego, A., Pereira, Z., Perez, J. J., Périgois, C., Perna, G., Perreca, A., Perret, J., Perriès, S., Perry, J. W., Pesios, D., Petracca, S., Petrillo, C., Pfeiffer, H. P., Pham, H., Pham, K. A., Phukon, K. S., Phurailatpam, H., Piarulli, M., Piccari, L., Piccinni, O. J., Pichot, M., Piendibene, M., Piergiovanni, F., Pierini, L., Pierra, G., Pierro, V., Pietrzak, M., Pillas, M., Pilo, F., Pinard, L., Pinto, I. M., Pinto, M., Piotrzkowski, B. J., Pirello, M., Pitkin, M. D., Placidi, A., Placidi, E., Planas, M. L., Plastino, W., Poggiani, R., Polini, E., Pompili, L., Poon, J., Porcelli, E., Porter, E. K., Posnansky, C., Poulton, R., Powell, J., Pracchia, M., Pradhan, B. K., Pradier, T., Prajapati, A. K., Prasai, K., Prasanna, R., Prasia, P., Pratten, G., Principe, G., Principe, M., Prodi, G. A., Prokhorov, L., Prosposito, P., Puecher, A., Pullin, J., Punturo, M., Puppo, P., Pürrer, M., Qi, H., Qin, J., Quéméner, G., Quetschke, V., Quigley, C., Quinonez, P. J., Quitzow-James, R., Raab, F. J., Raabith, S. S., Raaijmakers, G., Raja, S., Rajan, C., Rajbhandari, B., Ramirez, K. E., Vidal, F. A. Ramis, Ramos-Buades, A., Rana, D., Ranjan, S., Ransom, K., Rapagnani, P., Ratto, B., Rawat, S., Ray, A., Raymond, V., Razzano, M., Read, J., Payo, M. Recaman, Regimbau, T., Rei, L., Reid, S., Reitze, D. H., Relton, P., Renzini, A. I., Rettegno, P., Revenu, B., Reyes, R., Rezaei, A. S., Ricci, F., Ricci, M., Ricciardone, A., Richardson, J. W., Richardson, M., Rijal, A., Riles, K., Riley, H. K., Rinaldi, S., Rittmeyer, J., Robertson, C., Robinet, F., Robinson, M., Rocchi, A., Rolland, L., Rollins, J. G., Romano, A. E., Romano, R., Romero, A., Romero-Shaw, I. M., Romie, J. H., Ronchini, S., Roocke, T. J., Rosa, L., Rosauer, T. J., Rose, C. A., Rosińska, D., Ross, M. P., Rossello, M., Rowan, S., Roy, S. K., Roy, S., Rozza, D., Ruggi, P., Ruhama, N., Morales, E. Ruiz, Ruiz-Rocha, K., Sachdev, S., Sadecki, T., Sadiq, J., Saffarieh, P., Sah, M. R., Saha, S. S., Saha, S., Sainrat, T., Menon, S. Sajith, Sakai, K., Sakellariadou, M., Sakon, S., Salafia, O. S., Salces-Carcoba, F., Salconi, L., Saleem, M., Salemi, F., Sallé, M., Salvador, S., Sanchez, A., Sanchez, E. J., Sanchez, J. H., Sanchez, L. E., Sanchis-Gual, N., Sanders, J. R., Sänger, E. M., Santoliquido, F., Saravanan, T. R., Sarin, N., Sasaoka, S., Sasli, A., Sassi, P., Sassolas, B., Satari, H., Sato, R., Sato, Y., Sauter, O., Savage, R. L., Sawada, T., Sawant, H. L., Sayah, S., Scacco, V., Schaetzl, D., Scheel, M., Schiebelbein, A., Schiworski, M. G., Schmidt, P., Schmidt, S., Schnabel, R., Schneewind, M., Schofield, R. M. S., Schouteden, K., Schulte, B. W., Schutz, B. F., Schwartz, E., Scialpi, M., Scott, J., Scott, S. M., Seetharamu, T. C., Seglar-Arroyo, M., Sekiguchi, Y., Sellers, D., Sengupta, A. S., Sentenac, D., Seo, E. G., Seo, J. W., Sequino, V., Serra, M., Servignat, G., Sevrin, A., Shaffer, T., Shah, U. S., Shaikh, M. A., Shao, L., Sharma, A. K., Sharma, P., Sharma-Chaudhary, S., Shaw, M. R., Shawhan, P., Shcheblanov, N. S., Sheridan, E., Shikano, Y., Shikauchi, M., Shimode, K., Shinkai, H., Shiota, J., Shoemaker, D. H., Shoemaker, D. M., Short, R. W., ShyamSundar, S., Sider, A., Siegel, H., Sieniawska, M., Sigg, D., Silenzi, L., Simmonds, M., Singer, L. P., Singh, A., Singh, D., Singh, M. K., Singh, S., Singha, A., Sintes, A. M., Sipala, V., Skliris, V., Slagmolen, B. J. J., Slaven-Blair, T. J., Smetana, J., Smith, J. R., Smith, L., Smith, R. J. E., Smith, W. J., Soldateschi, J., Somiya, K., Song, I., Soni, K., Soni, S., Sordini, V., Sorrentino, F., Sorrentino, N., Sotani, H., Soulard, R., Southgate, A., Spagnuolo, V., Spencer, A. P., Spera, M., Spinicelli, P., Spoon, J. B., Sprague, C. A., Srivastava, A. K., Stachurski, F., Steer, D. A., Steinlechner, J., Steinlechner, S., Stergioulas, N., Stevens, P., StPierre, M., Stratta, G., Strong, M. D., Strunk, A., Sturani, R., Stuver, A. L., Suchenek, M., Sudhagar, S., Sueltmann, N., Suleiman, L., Sullivan, K. D., Sun, L., Sunil, S., Suresh, J., Sutton, P. J., Suzuki, T., Suzuki, Y., Swinkels, B. L., Syx, A., Szczepańczyk, M. J., Szewczyk, P., Tacca, M., Tagoshi, H., Tait, S. C., Takahashi, H., Takahashi, R., Takamori, A., Takase, T., Takatani, K., Takeda, H., Takeshita, K., Talbot, C., Tamaki, M., Tamanini, N., Tanabe, D., Tanaka, K., Tanaka, S. J., Tanaka, T., Tang, D., Tanioka, S., Tanner, D. B., Tao, L., Tapia, R. D., Martín, E. N. Tapia San, Tarafder, R., Taranto, C., Taruya, A., Tasson, J. D., Teloi, M., Tenorio, R., Themann, H., Theodoropoulos, A., Thirugnanasambandam, M. P., Thomas, L. M., Thomas, M., Thomas, P., Thompson, J. E., Thondapu, S. R., Thorne, K. A., Thrane, E., Tissino, J., Tiwari, A., Tiwari, P., Tiwari, S., Tiwari, V., Todd, M. R., Toivonen, A. M., Toland, K., Tolley, A. E., Tomaru, T., Tomita, K., Tomura, T., Tong-Yu, C., Toriyama, A., Toropov, N., Torres-Forné, A., Torrie, C. I., Toscani, M., Melo, I. Tosta e, Tournefier, E., Trapananti, A., Travasso, F., Traylor, G., Trevor, M., Tringali, M. C., Tripathee, A., Troian, G., Troiano, L., Trovato, A., Trozzo, L., Trudeau, R. J., Tsang, T. T. L., Tso, R., Tsuchida, S., Tsukada, L., Tsutsui, T., Turbang, K., Turconi, M., Turski, C., Ubach, H., Uchiyama, T., Udall, R. P., Uehara, T., Uematsu, M., Ueno, K., Ueno, S., Undheim, V., Ushiba, T., Vacatello, M., Vahlbruch, H., Vaidya, N., Vajente, G., Vajpeyi, A., Valdes, G., Valencia, J., Valentini, M., Vallejo-Peña, S. A., Vallero, S., Valsan, V., van Bakel, N., van Beuzekom, M., van Dael, M., Brand, J. F. J. van den, Broeck, C. Van Den, Vander-Hyde, D. C., van der Sluys, M., Van de Walle, A., van Dongen, J., Vandra, K., van Haevermaet, H., van Heijningen, J. V., Van Hove, P., VanKeuren, M., Vanosky, J., van Putten, M. H. P. M., van Ranst, Z., van Remortel, N., Vardaro, M., Vargas, A. F., Varghese, J. J., Varma, V., Vasúth, M., Vecchio, A., Vedovato, G., Veitch, J., Veitch, P. J., Venikoudis, S., Venneberg, J., Verdier, P., Verkindt, D., Verma, B., Verma, P., Verma, Y., Vermeulen, S. M., Vetrano, F., Veutro, A., Vibhute, A. M., Viceré, A., Vidyant, S., Viets, A. D., Vijaykumar, A., Vilkha, A., Villa-Ortega, V., Vincent, E. T., Vinet, J. -Y., Viret, S., Virtuoso, A., Vitale, S., Vives, A., Vocca, H., Voigt, D., von Reis, E. R. G., von Wrangel, J. S. A., Vyatchanin, S. P., Wade, L. E., Wade, M., Wagner, K. J., Wajid, A., Walker, M., Wallace, G. S., Wallace, L., Wang, H., Wang, J. Z., Wang, W. H., Wang, Z., Waratkar, G., Warner, J., Was, M., Washimi, T., Washington, N. Y., Watarai, D., Wayt, K. E., Weaver, B. R., Weaver, B., Weaving, C. R., Webster, S. A., Weinert, M., Weinstein, A. J., Weiss, R., Wellmann, F., Wen, L., Weßels, P., Wette, K., Whelan, J. T., Whiting, B. F., Whittle, C., Wildberger, J. B., Wilk, O. S., Wilken, D., Wilkin, A. T., Willadsen, D. J., Willetts, K., Williams, D., Williams, M. J., Williams, N. S., Willis, J. L., Willke, B., Wils, M., Winterflood, J., Wipf, C. C., Woan, G., Woehler, J., Wofford, J. K., Wolfe, N. E., Wong, H. T., Wong, H. W. Y., Wong, I. C. F., Wright, J. L., Wright, M., Wu, C., Wu, D. S., Wu, H., Wuchner, E., Wysocki, D. M., Xu, V. A., Xu, Y., Yadav, N., Yamamoto, H., Yamamoto, K., Yamamoto, T. S., Yamamoto, T., Yamamura, S., Yamazaki, R., Yan, S., Yan, T., Yang, F. W., Yang, F., Yang, K. Z., Yang, Y., Yarbrough, Z., Yasui, H., Yeh, S. -W., Yelikar, A. B., Yin, X., Yokoyama, J., Yokozawa, T., Yoo, J., Yu, H., Yuan, S., Yuzurihara, H., Zadrożny, A., Zanolin, M., Zeeshan, M., Zelenova, T., Zendri, J. -P., Zeoli, M., Zerrad, M., Zevin, M., Zhang, A. C., Zhang, L., Zhang, R., Zhang, T., Zhang, Y., Zhao, C., Zhao, Yue, Zhao, Yuhang, Zheng, Y., Zhong, H., Zhou, R., Zhu, X. -J., Zhu, Z. -H., Zucker, M. E., and Zweizig, J.
- Subjects
Astrophysics - High Energy Astrophysical Phenomena - Abstract
The magnetar SGR 1935+2154 is the only known Galactic source of fast radio bursts (FRBs). FRBs from SGR 1935+2154 were first detected by CHIME/FRB and STARE2 in 2020 April, after the conclusion of the LIGO, Virgo, and KAGRA Collaborations' O3 observing run. Here we analyze four periods of gravitational wave (GW) data from the GEO600 detector coincident with four periods of FRB activity detected by CHIME/FRB, as well as X-ray glitches and X-ray bursts detected by NICER and NuSTAR close to the time of one of the FRBs. We do not detect any significant GW emission from any of the events. Instead, using a short-duration GW search (for bursts $\leq$ 1 s) we derive 50\% (90\%) upper limits of $10^{48}$ ($10^{49}$) erg for GWs at 300 Hz and $10^{49}$ ($10^{50}$) erg at 2 kHz, and constrain the GW-to-radio energy ratio to $\leq 10^{14} - 10^{16}$. We also derive upper limits from a long-duration search for bursts with durations between 1 and 10 s. These represent the strictest upper limits on concurrent GW emission from FRBs., Comment: 15 pages of text including references, 4 figures, 5 tables
- Published
- 2024
16. Open3DTrack: Towards Open-Vocabulary 3D Multi-Object Tracking
- Author
-
Ishaq, Ayesha, Boudjoghra, Mohamed El Amine, Lahoud, Jean, Khan, Fahad Shahbaz, Khan, Salman, Cholakkal, Hisham, and Anwer, Rao Muhammad
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Robotics - Abstract
3D multi-object tracking plays a critical role in autonomous driving by enabling the real-time monitoring and prediction of multiple objects' movements. Traditional 3D tracking systems are typically constrained by predefined object categories, limiting their adaptability to novel, unseen objects in dynamic environments. To address this limitation, we introduce open-vocabulary 3D tracking, which extends the scope of 3D tracking to include objects beyond predefined categories. We formulate the problem of open-vocabulary 3D tracking and introduce dataset splits designed to represent various open-vocabulary scenarios. We propose a novel approach that integrates open-vocabulary capabilities into a 3D tracking framework, allowing for generalization to unseen object classes. Our method effectively reduces the performance gap between tracking known and novel objects through strategic adaptation. Experimental results demonstrate the robustness and adaptability of our method in diverse outdoor driving scenarios. To the best of our knowledge, this work is the first to address open-vocabulary 3D tracking, presenting a significant advancement for autonomous systems in real-world settings. Code, trained models, and dataset splits are available publicly., Comment: 7 pages, 4 figures, 3 tables
- Published
- 2024
17. AgriCLIP: Adapting CLIP for Agriculture and Livestock via Domain-Specialized Cross-Model Alignment
- Author
-
Nawaz, Umair, Awais, Muhammad, Gani, Hanan, Naseer, Muzammal, Khan, Fahad, Khan, Salman, and Anwer, Rao Muhammad
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Capitalizing on vast amount of image-text data, large-scale vision-language pre-training has demonstrated remarkable zero-shot capabilities and has been utilized in several applications. However, models trained on general everyday web-crawled data often exhibit sub-optimal performance for specialized domains, likely due to domain shift. Recent works have tackled this problem for some domains (e.g., healthcare) by constructing domain-specialized image-text data. However, constructing a dedicated large-scale image-text dataset for sustainable area of agriculture and livestock is still open to research. Further, this domain desires fine-grained feature learning due to the subtle nature of the downstream tasks (e.g, nutrient deficiency detection, livestock breed classification). To address this we present AgriCLIP, a vision-language foundational model dedicated to the domain of agriculture and livestock. First, we propose a large-scale dataset, named ALive, that leverages customized prompt generation strategy to overcome the scarcity of expert annotations. Our ALive dataset covers crops, livestock, and fishery, with around 600,000 image-text pairs. Second, we propose a training pipeline that integrates both contrastive and self-supervised learning to learn both global semantic and local fine-grained domain-specialized features. Experiments on diverse set of 20 downstream tasks demonstrate the effectiveness of AgriCLIP framework, achieving an absolute gain of 7.8\% in terms of average zero-shot classification accuracy, over the standard CLIP adaptation via domain-specialized ALive dataset. Our ALive dataset and code can be accessible at \href{https://github.com/umair1221/AgriCLIP/tree/main}{Github}.
- Published
- 2024
18. Continual Human Pose Estimation for Incremental Integration of Keypoints and Pose Variations
- Author
-
Khan, Muhammad Saif Ullah, Khan, Muhammad Ahmed Ullah, Afzal, Muhammad Zeshan, and Stricker, Didier
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
This paper reformulates cross-dataset human pose estimation as a continual learning task, aiming to integrate new keypoints and pose variations into existing models without losing accuracy on previously learned datasets. We benchmark this formulation against established regularization-based methods for mitigating catastrophic forgetting, including EWC, LFL, and LwF. Moreover, we propose a novel regularization method called Importance-Weighted Distillation (IWD), which enhances conventional LwF by introducing a layer-wise distillation penalty and dynamic temperature adjustment based on layer importance for previously learned knowledge. This allows for a controlled adaptation to new tasks that respects the stability-plasticity balance critical in continual learning. Through extensive experiments across three datasets, we demonstrate that our approach outperforms existing regularization-based continual learning strategies. IWD shows an average improvement of 3.60\% over the state-of-the-art LwF method. The results highlight the potential of our method to serve as a robust framework for real-world applications where models must evolve with new data without forgetting past knowledge.
- Published
- 2024
19. CDChat: A Large Multimodal Model for Remote Sensing Change Description
- Author
-
Noman, Mubashir, Ahsan, Noor, Naseer, Muzammal, Cholakkal, Hisham, Anwer, Rao Muhammad, Khan, Salman, and Khan, Fahad Shahbaz
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Large multimodal models (LMMs) have shown encouraging performance in the natural image domain using visual instruction tuning. However, these LMMs struggle to describe the content of remote sensing images for tasks such as image or region grounding, classification, etc. Recently, GeoChat make an effort to describe the contents of the RS images. Although, GeoChat achieves promising performance for various RS tasks, it struggles to describe the changes between bi-temporal RS images which is a key RS task. This necessitates the development of an LMM that can describe the changes between the bi-temporal RS images. However, there is insufficiency of datasets that can be utilized to tune LMMs. In order to achieve this, we introduce a change description instruction dataset that can be utilized to finetune an LMM and provide better change descriptions for RS images. Furthermore, we show that the LLaVA-1.5 model, with slight modifications, can be finetuned on the change description instruction dataset and achieve favorably better performance.
- Published
- 2024
20. Efficient Localized Adaptation of Neural Weather Forecasting: A Case Study in the MENA Region
- Author
-
Munir, Muhammad Akhtar, Khan, Fahad Shahbaz, and Khan, Salman
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Physics - Atmospheric and Oceanic Physics - Abstract
Accurate weather and climate modeling is critical for both scientific advancement and safeguarding communities against environmental risks. Traditional approaches rely heavily on Numerical Weather Prediction (NWP) models, which simulate energy and matter flow across Earth's systems. However, heavy computational requirements and low efficiency restrict the suitability of NWP, leading to a pressing need for enhanced modeling techniques. Neural network-based models have emerged as promising alternatives, leveraging data-driven approaches to forecast atmospheric variables. In this work, we focus on limited-area modeling and train our model specifically for localized region-level downstream tasks. As a case study, we consider the MENA region due to its unique climatic challenges, where accurate localized weather forecasting is crucial for managing water resources, agriculture and mitigating the impacts of extreme weather events. This targeted approach allows us to tailor the model's capabilities to the unique conditions of the region of interest. Our study aims to validate the effectiveness of integrating parameter-efficient fine-tuning (PEFT) methodologies, specifically Low-Rank Adaptation (LoRA) and its variants, to enhance forecast accuracy, as well as training speed, computational resource utilization, and memory efficiency in weather and climate modeling for specific regions., Comment: Our codebase and pre-trained models can be accessed at: [this url](https://github.com/akhtarvision/weather-regional)
- Published
- 2024
21. Personalized Federated Learning Techniques: Empirical Analysis
- Author
-
Khan, Azal Ahmad, Khan, Ahmad Faraz, Ali, Haider, and Anwar, Ali
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Cryptography and Security - Abstract
Personalized Federated Learning (pFL) holds immense promise for tailoring machine learning models to individual users while preserving data privacy. However, achieving optimal performance in pFL often requires a careful balancing act between memory overhead costs and model accuracy. This paper delves into the trade-offs inherent in pFL, offering valuable insights for selecting the right algorithms for diverse real-world scenarios. We empirically evaluate ten prominent pFL techniques across various datasets and data splits, uncovering significant differences in their performance. Our study reveals interesting insights into how pFL methods that utilize personalized (local) aggregation exhibit the fastest convergence due to their efficiency in communication and computation. Conversely, fine-tuning methods face limitations in handling data heterogeneity and potential adversarial attacks while multi-objective learning methods achieve higher accuracy at the cost of additional training and resource consumption. Our study emphasizes the critical role of communication efficiency in scaling pFL, demonstrating how it can significantly affect resource usage in real-world deployments.
- Published
- 2024
22. Fault Tolerant Metric Dimensions of Leafless Cacti Graphs with Application in Supply Chain Management
- Author
-
Asif, Tauseef, Haidar, Ghulam, Yousafzai, Faisal, Khan, Murad Ul Islam, Khan, Qaisar, and Fatima, Rakea
- Subjects
Computer Science - Discrete Mathematics ,Mathematics - Combinatorics ,05C12, 05C90 - Abstract
A resolving set for a simple graph $G$ is a subset of vertex set of $G$ such that it distinguishes all vertices of $G$ using the shortest distance from this subset. This subset is a metric basis if it is the smallest set with this property. A resolving set is a fault tolerant resolving set if the removal of any vertex from the subset still leaves it a resolving set. The smallest set satisfying this property is the fault tolerant metric basis, and the cardinality of this set is termed as fault tolerant metric dimension of $G$, denoted by $\beta'(G)$. In this article, we determine the fault tolerant metric dimension of bicyclic graphs of type-I and II and show that it is always $4$ for both types of graphs. We then use these results to form our basis to consider leafless cacti graphs, and calculate their fault tolerant metric dimensions in terms of \textit{inner cycles} and \textit{outer cycles}. We then consider a detailed real world example of supply and distribution center management, and discuss the application of fault tolerant metric dimension in such a scenario. We also briefly discuss some other scenarios where leafless cacti graphs can be used to model real world problems.
- Published
- 2024
23. AD-Net: Attention-based dilated convolutional residual network with guided decoder for robust skin lesion segmentation
- Author
-
Naveed, Asim, Naqvi, Syed S., Khan, Tariq M., Iqbal, Shahzaib, Wani, M. Yaqoob, and Khan, Haroon Ahmed
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
In computer-aided diagnosis tools employed for skin cancer treatment and early diagnosis, skin lesion segmentation is important. However, achieving precise segmentation is challenging due to inherent variations in appearance, contrast, texture, and blurry lesion boundaries. This research presents a robust approach utilizing a dilated convolutional residual network, which incorporates an attention-based spatial feature enhancement block (ASFEB) and employs a guided decoder strategy. In each dilated convolutional residual block, dilated convolution is employed to broaden the receptive field with varying dilation rates. To improve the spatial feature information of the encoder, we employed an attention-based spatial feature enhancement block in the skip connections. The ASFEB in our proposed method combines feature maps obtained from average and maximum-pooling operations. These combined features are then weighted using the active outcome of global average pooling and convolution operations. Additionally, we have incorporated a guided decoder strategy, where each decoder block is optimized using an individual loss function to enhance the feature learning process in the proposed AD-Net. The proposed AD-Net presents a significant benefit by necessitating fewer model parameters compared to its peer methods. This reduction in parameters directly impacts the number of labeled data required for training, facilitating faster convergence during the training process. The effectiveness of the proposed AD-Net was evaluated using four public benchmark datasets. We conducted a Wilcoxon signed-rank test to verify the efficiency of the AD-Net. The outcomes suggest that our method surpasses other cutting-edge methods in performance, even without the implementation of data augmentation strategies.
- Published
- 2024
24. CONDA: Condensed Deep Association Learning for Co-Salient Object Detection
- Author
-
Li, Long, Liu, Nian, Zhang, Dingwen, Li, Zhongyu, Khan, Salman, Anwer, Rao, Cholakkal, Hisham, Han, Junwei, and Khan, Fahad Shahbaz
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Inter-image association modeling is crucial for co-salient object detection. Despite satisfactory performance, previous methods still have limitations on sufficient inter-image association modeling. Because most of them focus on image feature optimization under the guidance of heuristically calculated raw inter-image associations. They directly rely on raw associations which are not reliable in complex scenarios, and their image feature optimization approach is not explicit for inter-image association modeling. To alleviate these limitations, this paper proposes a deep association learning strategy that deploys deep networks on raw associations to explicitly transform them into deep association features. Specifically, we first create hyperassociations to collect dense pixel-pair-wise raw associations and then deploys deep aggregation networks on them. We design a progressive association generation module for this purpose with additional enhancement of the hyperassociation calculation. More importantly, we propose a correspondence-induced association condensation module that introduces a pretext task, i.e. semantic correspondence estimation, to condense the hyperassociations for computational burden reduction and noise elimination. We also design an object-aware cycle consistency loss for high-quality correspondence estimations. Experimental results in three benchmark datasets demonstrate the remarkable effectiveness of our proposed method with various training settings., Comment: There is an error. In Sec 4.1, the number of images in some dataset is incorrect and needs to be revised
- Published
- 2024
25. Current Practices and Pitfalls of ELT Syllabi for Developing Engineering Students' Communicative English in Bangladesh
- Author
-
Mohammad Ehsanul Islam Khan, Mohammad Shahazahan Seraj Bhuiyan, and Mohammad Ekramul Islam Khan
- Abstract
The current practices and shortcomings of engineering students' English language teaching (ELT) syllabi were examined from the perspectives of learners and teachers in English as a foreign language (EFL) context. The syllabi included content that had little impact on students' communicative competence in English (CC-E). Students were generally concerned about their professional communication abilities. In this study, the researchers collected data from ten engineering-focused universities in Bangladesh. These universities' existing ELT syllabi (ELT-S) were examined, seeking the current practices and pitfalls. The study sampled 152 participants from the selected universities. The study followed a mixed-method approach. In the qualitative technique, content analysis, focus group discussion (FGD), and interviews were employed for data collection, while survey questions were used in the quantitative approach. The study's findings revealed that the existing English syllabi of those selected universities required updating and modification to meet the identified professional requirements regarding the type, credit allotment, content, classroom practices, class size, policies, etc. The improvements included redesigning English syllabi, material, and teaching methods to improve engineering students' communicative abilities. A uniform curriculum with at least one English language sessional course per semester in all engineering majors was strongly recommended.
- Published
- 2024
26. Study of the effects of heating on the physical, optical, and electrical properties of NiO thin films synthesized using a low-cost sol-gel method
- Author
-
Khan, Muhammad Yasir, Akhtar, Muhammad Wasim, Khan, Muhammad Furqan Ali, Abbass, Zeeshan, ur-Rasheed, Fayyaz, Ali, Muhammad Saquib, Pirzada, Noman, and Shahbaz, Raja
- Published
- 2024
27. Effect of Inclusion of Different Levels of Duckweed (Lemna minor) on the Performance of Broiler Chicken
- Author
-
Zaffer, B., Sheikh, I.U., Banday, M.T., Adil, S., Ahmed, H.A., Khan, A.S., Nissa, S.S., and Mirza, U.
- Published
- 2021
- Full Text
- View/download PDF
28. Standardization and HPTLC Fingerprinting of Unani compound formulation Habb-e-Muqil Jadeed
- Author
-
Negi, R. K., Naikodi, M. A. Rasheed, Sajwan, S., Khan, Asim Ali, Khan, A. S., and Meena, R. P.
- Published
- 2021
- Full Text
- View/download PDF
29. Multi-Task Adversarial Variational Autoencoder for Estimating Biological Brain Age with Multimodal Neuroimaging
- Author
-
Usman, Muhammad, Rehman, Azka, Shahid, Abdullah, Rehman, Abd Ur, Gho, Sung-Min, Lee, Aleum, Khan, Tariq M., and Razzak, Imran
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
Despite advances in deep learning for estimating brain age from structural MRI data, incorporating functional MRI data is challenging due to its complex structure and the noisy nature of functional connectivity measurements. To address this, we present the Multitask Adversarial Variational Autoencoder, a custom deep learning framework designed to improve brain age predictions through multimodal MRI data integration. This model separates latent variables into generic and unique codes, isolating shared and modality-specific features. By integrating multitask learning with sex classification as an additional task, the model captures sex-specific aging patterns. Evaluated on the OpenBHB dataset, a large multisite brain MRI collection, the model achieves a mean absolute error of 2.77 years, outperforming traditional methods. This success positions M-AVAE as a powerful tool for metaverse-based healthcare applications in brain age estimation.
- Published
- 2024
30. Holographic MIMO for Next Generation Non-Terrestrial Networks: Motivation, Opportunities, and Challenges
- Author
-
Iacovelli, Giovanni, Sheemar, Chandan Kumar, Khan, Wali Ullah, Mahmood, Asad, Alexandropoulos, George C., Querol, Jorge, and Chatzinotas, Symeon
- Subjects
Electrical Engineering and Systems Science - Signal Processing - Abstract
In this article, we propose the integration of the Holographic Multiple Input Multiple Output (HMIMO) as a transformative solution for next generation Non-Terrestrial Networks (NTNs), addressing key challenges, such as high hardware costs, launch expenses, and energy inefficiency. Traditional NTNs are constrained by the financial and operational limitations posed by bulky, costly antenna systems, alongside the complexities of maintaining effective communications in space. HMIMO offers a novel approach utilizing compact and lightweight arrays of densely packed radiating elements with real-time reconfiguration capabilities, thus, capable of optimizing system performance under dynamic conditions such as varying orbital dynamics and Doppler shifts. By replacing conventional antenna systems with HMIMO, the complexity and cost of satellite manufacturing and launch can be substantially reduced, enabling more streamlined and cost-effective satellite designs. This advancement holds significant potential to democratize space communications, making them accessible to a broader range of stakeholders, including smaller nations and commercial enterprises. Moreover, the inherent capabilities of HMIMO in enhancing energy efficiency, scalability, and adaptability position this technology as a key enabler of new use cases and sustainable satellite operations.
- Published
- 2024
31. A Density Functional Theory Study of Magnetic Transition in MnO2 adsorbed Vanadium Carbide (V$_2$C) MXene
- Author
-
Fatima, Mahjabeen, Khan, Saleem Ayaz, and Rizwan, Syed
- Subjects
Condensed Matter - Mesoscale and Nanoscale Physics ,Condensed Matter - Materials Science ,Physics - Applied Physics - Abstract
The work reports nonmagnetic behavior (0.04 $\mu$B) in two-dimensional (2D) V2C-OF MXene and ferromagnetism in MnO$_2$ adsorbed V2C-OF MXene. The density functional theory (DFT) calculations were carried out to study the magnetic moments of V$_2$C-OF and MnO$_2$@V$_2$C-OF MXene. The MXene, which is derived from the exfoliation of its parent V$_2$AlC MAX phase, shows a good potential to be a ferromagnetic when MnO$_2$ is adsorbed on it. The V$_2$C MXene and MnO$_2$ adsorbed V$_2$C MXene were successfully synthesized, as characterized using X-ray diffraction, showing an increased c-lattice parameter from 22.6{\AA} to 27.2{\AA} after MnO$_2$ adsorption. The DFT study confirmed that MnO$_2$ adsorbed V$_2$C MXene changed from nonmagnetic (in V$_2$C MXene) to a strong ferromagnetic with a magnetic moment of 4.48$\mu$B for Mn adsorbed V$_2$C-OF MXene. The current work is a step-forward towards understanding of magnetism in two-dimensional materials for future 2D spintronics.
- Published
- 2024
32. TimeLess: A Vision for the Next Generation of Software Development
- Author
-
Rasheed, Zeeshan, Sami, Malik Abdul, Rasku, Jussi, Kemell, Kai-Kristian, Zhang, Zheying, Harjamaki, Janne, Siddeeq, Shahbaz, Lahti, Sami, Herda, Tomas, Nurminen, Mikko, Lavesson, Niklas, de Cerqueira, Jose Siqueira, Hasan, Toufique, Khan, Ayman, Hasan, Mahade, Saari, Mika, Rantanen, Petri, Soini, Jari, and Abrahamsson, Pekka
- Subjects
Computer Science - Software Engineering - Abstract
Present-day software development faces three major challenges: complexity, time consumption, and high costs. Developing large software systems often requires battalions of teams and considerable time for meetings, which end without any action, resulting in unproductive cycles, delayed progress, and increased cost. What if, instead of large meetings with no immediate results, the software product is completed by the end of the meeting? In response, we present a vision for a system called TimeLess, designed to reshape the software development process by enabling immediate action during meetings. The goal is to shift meetings from planning discussions to productive, action-oriented sessions. This approach minimizes the time and effort required for development, allowing teams to focus on critical decision-making while AI agents execute development tasks based on the meeting discussions. We will employ multiple AI agents that work collaboratively to capture human discussions and execute development tasks in real time. This represents a step toward next-generation software development environments, where human expertise drives strategy and AI accelerates task execution., Comment: 5 pages, 4 figure, and 1 table
- Published
- 2024
33. PatchCTG: Patch Cardiotocography Transformer for Antepartum Fetal Health Monitoring
- Author
-
Khan, M. Jaleed, Vatish, Manu, and Jones, Gabriel Davis
- Subjects
Computer Science - Artificial Intelligence ,Computer Science - Machine Learning - Abstract
Antepartum Cardiotocography (CTG) is vital for fetal health monitoring, but traditional methods like the Dawes-Redman system are often limited by high inter-observer variability, leading to inconsistent interpretations and potential misdiagnoses. This paper introduces PatchCTG, a transformer-based model specifically designed for CTG analysis, employing patch-based tokenisation, instance normalisation and channel-independent processing to capture essential local and global temporal dependencies within CTG signals. PatchCTG was evaluated on the Oxford Maternity (OXMAT) dataset, comprising over 20,000 CTG traces across diverse clinical outcomes after applying the inclusion and exclusion criteria. With extensive hyperparameter optimisation, PatchCTG achieved an AUC of 77%, with specificity of 88% and sensitivity of 57% at Youden's index threshold, demonstrating adaptability to various clinical needs. Testing across varying temporal thresholds showed robust predictive performance, particularly with finetuning on data closer to delivery, achieving a sensitivity of 52% and specificity of 88% for near-delivery cases. These findings suggest the potential of PatchCTG to enhance clinical decision-making in antepartum care by providing a reliable, objective tool for fetal health assessment. The source code is available at https://github.com/jaleedkhan/PatchCTG.
- Published
- 2024
34. ALOcc: Adaptive Lifting-based 3D Semantic Occupancy and Cost Volume-based Flow Prediction
- Author
-
Chen, Dubing, Fang, Jin, Han, Wencheng, Cheng, Xinjing, Yin, Junbo, Xu, Chenzhong, Khan, Fahad Shahbaz, and Shen, Jianbing
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Vision-based semantic occupancy and flow prediction plays a crucial role in providing spatiotemporal cues for real-world tasks, such as autonomous driving. Existing methods prioritize higher accuracy to cater to the demands of these tasks. In this work, we strive to improve performance by introducing a series of targeted improvements for 3D semantic occupancy prediction and flow estimation. First, we introduce an occlusion-aware adaptive lifting mechanism with a depth denoising technique to improve the robustness of 2D-to-3D feature transformation and reduce the reliance on depth priors. Second, we strengthen the semantic consistency between 3D features and their original 2D modalities by utilizing shared semantic prototypes to jointly constrain both 2D and 3D features. This is complemented by confidence- and category-based sampling strategies to tackle long-tail challenges in 3D space. To alleviate the feature encoding burden in the joint prediction of semantics and flow, we propose a BEV cost volume-based prediction method that links flow and semantic features through a cost volume and employs a classification-regression supervision scheme to address the varying flow scales in dynamic scenes. Our purely convolutional architecture framework, named ALOcc, achieves an optimal tradeoff between speed and accuracy achieving state-of-the-art results on multiple benchmarks. On Occ3D and training without the camera visible mask, our ALOcc achieves an absolute gain of 2.5\% in terms of RayIoU while operating at a comparable speed compared to the state-of-the-art, using the same input size (256$\times$704) and ResNet-50 backbone. Our method also achieves 2nd place in the CVPR24 Occupancy and Flow Prediction Competition.
- Published
- 2024
35. A Composite Hydrogel of Porous Gold Nanorods and Gelatin: Nanoscale Structure and Rheo-Mechanical Properties
- Author
-
Khan, Irfan, Panda, Snigdharani, Kumar, Sugam, and Srivastava, Sunita
- Subjects
Condensed Matter - Soft Condensed Matter ,Physics - Applied Physics - Abstract
Incorporating nanomaterials into hydrogels allows for the creation of versatile materials with properties that can be precisely tailored by manipulating their nanoscale structures, leading to a wide range of bulk properties. Investigating the structural and property characteristics of composite hydrogels is crucial in tailoring their performance for specific applications. This study focuses on investigating the correlation between the structural arrangement and properties of a composite hydrogel of thermoresponsive polymer, gelatin, and light-responsive antimicrobial porous gold nanorods, $PAuNR$. The rheo-mechanical properties of the composite hydrogels are correlated with their nanoscale structural characteristics, investigated using small-angle neutron scattering ($SANS$). Analysis of $SANS$ data reveals a decrease in the fractal dimension of $PAuNRs$ incorporated hydrogel matrix, as compared to pure gelatin. Incorporating $PAuNRs$ results in formation of softer composite hydrogel as evident from decrease in viscoelastic moduli, critical yield strain, denaturation temperature and swelling ratio. Our results demonstrates that the structural modulation at the nanoscale can be precisely controlled through adjusting $PAuNRs$ concentration and temperature providing an fabrication mechanism for hydrogels with desired elastic properties. The reduced elasticity of the composite hydrogel and light sensitive/antimicrobial property of the $PAuNRs$ makes this system suitable for specific biomedical applications, such as tissue engineering, device fabrication and stimuli based controlled drug delivery devices respectively., Comment: Under review: The Journal of Chemical Physics (JCP24-AR-04120)
- Published
- 2024
36. Large Language Model in Medical Informatics: Direct Classification and Enhanced Text Representations for Automatic ICD Coding
- Author
-
Boukhers, Zeyd, Khan, AmeerAli, Ramadan, Qusai, and Yang, Cong
- Subjects
Computer Science - Machine Learning ,Computer Science - Information Retrieval - Abstract
Addressing the complexity of accurately classifying International Classification of Diseases (ICD) codes from medical discharge summaries is challenging due to the intricate nature of medical documentation. This paper explores the use of Large Language Models (LLM), specifically the LLAMA architecture, to enhance ICD code classification through two methodologies: direct application as a classifier and as a generator of enriched text representations within a Multi-Filter Residual Convolutional Neural Network (MultiResCNN) framework. We evaluate these methods by comparing them against state-of-the-art approaches, revealing LLAMA's potential to significantly improve classification outcomes by providing deep contextual insights into medical texts., Comment: accepted at the 2024 IEEE International Conference on Bioinformatics and Biomedicine (BIBM 2024)
- Published
- 2024
37. Harnessing Smartphone Sensors for Enhanced Road Safety: A Comprehensive Dataset and Review
- Author
-
Khandakar, Amith, Michelson, David G., Naznine, Mansura, Salam, Abdus, Nahiduzzaman, Md., Khan, Khaled M., Suganthan, Ponnuthurai Nagaratnam, Ayari, Mohamed Arselene, Menouar, Hamid, and Haider, Julfikar
- Subjects
Computer Science - Robotics ,Computer Science - Artificial Intelligence - Abstract
Severe collisions can result from aggressive driving and poor road conditions, emphasizing the need for effective monitoring to ensure safety. Smartphones, with their array of built-in sensors, offer a practical and affordable solution for road-sensing. However, the lack of reliable, standardized datasets has hindered progress in assessing road conditions and driving patterns. This study addresses this gap by introducing a comprehensive dataset derived from smartphone sensors, which surpasses existing datasets by incorporating a diverse range of sensors including accelerometer, gyroscope, magnetometer, GPS, gravity, orientation, and uncalibrated sensors. These sensors capture extensive parameters such as acceleration force, gravitation, rotation rate, magnetic field strength, and vehicle speed, providing a detailed understanding of road conditions and driving behaviors. The dataset is designed to enhance road safety, infrastructure maintenance, traffic management, and urban planning. By making this dataset available to the community, the study aims to foster collaboration, inspire further research, and facilitate the development of innovative solutions in intelligent transportation systems., Comment: 29 pages, 14 Figures, journal paper, submitted into Scientific Data Journal
- Published
- 2024
38. Token Merging for Training-Free Semantic Binding in Text-to-Image Synthesis
- Author
-
Hu, Taihang, Li, Linxuan, van de Weijer, Joost, Gao, Hongcheng, Khan, Fahad Shahbaz, Yang, Jian, Cheng, Ming-Ming, Wang, Kai, and Wang, Yaxing
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
Although text-to-image (T2I) models exhibit remarkable generation capabilities, they frequently fail to accurately bind semantically related objects or attributes in the input prompts; a challenge termed semantic binding. Previous approaches either involve intensive fine-tuning of the entire T2I model or require users or large language models to specify generation layouts, adding complexity. In this paper, we define semantic binding as the task of associating a given object with its attribute, termed attribute binding, or linking it to other related sub-objects, referred to as object binding. We introduce a novel method called Token Merging (ToMe), which enhances semantic binding by aggregating relevant tokens into a single composite token. This ensures that the object, its attributes and sub-objects all share the same cross-attention map. Additionally, to address potential confusion among main objects with complex textual prompts, we propose end token substitution as a complementary strategy. To further refine our approach in the initial stages of T2I generation, where layouts are determined, we incorporate two auxiliary losses, an entropy loss and a semantic binding loss, to iteratively update the composite token to improve the generation integrity. We conducted extensive experiments to validate the effectiveness of ToMe, comparing it against various existing methods on the T2I-CompBench and our proposed GPT-4o object binding benchmark. Our method is particularly effective in complex scenarios that involve multiple objects and attributes, which previous methods often fail to address. The code will be publicly available at \url{https://github.com/hutaihang/ToMe}., Comment: Accepted by Neurips2024
- Published
- 2024
39. Combining Entangled and Non-Entangled Based Quantum Key Distribution Protocol With GHZ State
- Author
-
Sykot, Arman, Rahman, Mohammad Hasibur, Anannya, Rifat Tasnim, Upoma, Khan Shariya Hasan, and Mahdy, M. R. C.
- Subjects
Quantum Physics ,Computer Science - Cryptography and Security ,Physics - Applied Physics - Abstract
This paper presents a novel hybrid Quantum Key Distribution ,QKD, protocol that combines entanglement based and non entanglement based approaches to optimize security and the number of generated keys. We introduce a dynamic system that integrates a three particle GHZ state method with the two state B92 protocol, using a quantum superposition state to probabilistically switch between them. The GHZ state component leverages strong three particle entanglement correlations for enhanced security, while the B92 component offers simplicity and potentially higher key generation rates. Implemented and simulated using Qiskit, our approach demonstrates higher number of generated keys compared to standalone protocols while maintaining robust security. We present a comprehensive analysis of the security properties and performance characteristics of the proposed protocol. The results show that this combined method effectively balances the trade offs inherent in QKD systems, offering a flexible framework adaptable to varying channel conditions and security requirements.This research contributes to ongoing efforts to make QKD more practical and efficient, potentially advancing the development of large scale, secured quantum networks., Comment: 14 pages, 24 equations, 9 figures
- Published
- 2024
40. Extended multi-stream temporal-attention module for skeleton-based human action recognition (HAR)
- Author
-
Mehmood, Faisal, Guo, Xin, Chen, Enqing, Akbar, Muhammad Azeem, Khan, Arif Ali, and Ullah, Sami
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Graph convolutional networks (GCNs) are an effective skeleton-based human action recognition (HAR) technique. GCNs enable the specification of CNNs to a non-Euclidean frame that is more flexible. The previous GCN-based models still have a lot of issues: (I) The graph structure is the same for all model layers and input data., Comment: This paper accepted in Computers in Human Behavior Journal
- Published
- 2024
41. Mitigating covariate shift in non-colocated data with learned parameter priors
- Author
-
Khan, Behraj, Mirza, Behroz, Durrani, Nouman, and Syed, Tahir
- Subjects
Computer Science - Machine Learning ,Computer Science - Computer Vision and Pattern Recognition - Abstract
When training data are distributed across{ time or space,} covariate shift across fragments of training data biases cross-validation, compromising model selection and assessment. We present \textit{Fragmentation-Induced covariate-shift Remediation} ($FIcsR$), which minimizes an $f$-divergence between a fragment's covariate distribution and that of the standard cross-validation baseline. We s{how} an equivalence with popular importance-weighting methods. {The method}'s numerical solution poses a computational challenge owing to the overparametrized nature of a neural network, and we derive a Fisher Information approximation. When accumulated over fragments, this provides a global estimate of the amount of shift remediation thus far needed, and we incorporate that as a prior via the minimization objective. In the paper, we run extensive classification experiments on multiple data classes, over $40$ datasets, and with data batched over multiple sequence lengths. We extend the study to the $k$-fold cross-validation setting through a similar set of experiments. An ablation study exposes the method to varying amounts of shift and demonstrates slower degradation with $FIcsR$ in place. The results are promising under all these conditions; with improved accuracy against batch and fold state-of-the-art by more than $5\%$ and $10\%$, respectively.
- Published
- 2024
42. MA-DV2F: A Multi-Agent Navigation Framework using Dynamic Velocity Vector Field
- Author
-
Ma, Yining, Khan, Qadeer, and Cremers, Daniel
- Subjects
Computer Science - Multiagent Systems - Abstract
In this paper we propose MA-DV2F: Multi-Agent Dynamic Velocity Vector Field. It is a framework for simultaneously controlling a group of vehicles in challenging environments. DV2F is generated for each vehicle independently and provides a map of reference orientation and speed that a vehicle must attain at any point on the navigation grid such that it safely reaches its target. The field is dynamically updated depending on the speed and proximity of the ego-vehicle to other agents. This dynamic adaptation of the velocity vector field allows prevention of imminent collisions. Experimental results show that MA-DV2F outperforms concurrent methods in terms of safety, computational efficiency and accuracy in reaching the target when scaling to a large number of vehicles. Project page for this work can be found here: https://yininghase.github.io/MA-DV2F/
- Published
- 2024
43. The GHY boundary term from the string worldsheet to linear order
- Author
-
Ahmadain, Amr, Akhtar, Shoaib, and Khan, Rifath
- Subjects
High Energy Physics - Theory - Abstract
Using the method of images we derive the boundary term of the Einstein-$\Gamma^2$ action in half-space from the spherical worldsheet to first order in $\alpha'$ and to linear order in the metric perturbation around flat half-space. The $\Gamma^2$ action, written down by Einstein more than 100 years ago, includes a boundary term that consists of the Gibbons-Hawking-York action along with two additional terms that are functions of the metric, normal vector, and tangential derivatives. With this boundary term, the total (bulk + boundary) sphere effective action has a well-posed variational principle for Dirichlet boundary conditions., Comment: 20 pages
- Published
- 2024
44. A Hybrid Approach for COVID-19 Detection: Combining Wasserstein GAN with Transfer Learning
- Author
-
Rounaq, Sumera, Shah, Shahid Munir, Aljawarneh, Mahmoud, Khan, Sarah, and Muhammad, Ghulam
- Subjects
Electrical Engineering and Systems Science - Image and Video Processing ,Computer Science - Computer Vision and Pattern Recognition - Abstract
COVID-19 is extremely contagious and its rapid growth has drawn attention towards its early diagnosis. Early diagnosis of COVID-19 enables healthcare professionals and government authorities to break the chain of transition and flatten the epidemic curve. With the number of cases accelerating across the developed world, COVID-19 induced Viral Pneumonia cases is a big challenge. Overlapping of COVID-19 cases with Viral Pneumonia and other lung infections with limited dataset and long training hours is a serious problem to cater. Limited amount of data often results in over-fitting models and due to this reason, model does not predict generalized results. To fill this gap, we proposed GAN-based approach to synthesize images which later fed into the deep learning models to classify images of COVID-19, Normal, and Viral Pneumonia. Specifically, customized Wasserstein GAN is proposed to generate 19% more Chest X-ray images as compare to the real images. This expanded dataset is then used to train four proposed deep learning models: VGG-16, ResNet-50, GoogLeNet and MNAST. The result showed that expanded dataset utilized deep learning models to deliver high classification accuracies. In particular, VGG-16 achieved highest accuracy of 99.17% among all four proposed schemes. Rest of the models like ResNet-50, GoogLeNet and MNAST delivered 93.9%, 94.49% and 97.75% testing accuracies respectively. Later, the efficiency of these models is compared with the state of art models on the basis of accuracy. Further, our proposed models can be applied to address the issue of scant datasets for any problem of image analysis.
- Published
- 2024
45. PDRs4All XI. Empirical prescriptions for the interpretation of JWST imaging observations of star-forming regions
- Author
-
Chown, Ryan, Okada, Yoko, Peeters, Els, Sidhu, Ameek, Khan, Baria, Schefter, Bethany, Trahin, Boris, Canin, Amelie, Van De Putte, Dries, Alarcon, Felipe, Schroetter, Ilane, Kannavou, Olga, Habart, Emilie, Berne, Olivier, Boersma, Christiaan, Cami, Jan, Dartois, Emmanuel, Goicoechea, Javier, Gordon, Karl, and Onaka, Takashi
- Subjects
Astrophysics - Astrophysics of Galaxies - Abstract
(Abridged) JWST continues to deliver incredibly detailed infrared (IR) images of star forming regions in the Milky Way and beyond. IR emission from star-forming regions is very spectrally rich due to emission from gas-phase atoms, ions, and polycyclic aromatic hydrocarbons (PAHs). Physically interpreting IR images of these regions relies on assumptions about the underlying spectral energy distribution in the imaging bandpasses. We aim to provide empirical prescriptions linking line, PAH, and continuum intensities from JWST images, to facilitate the interpretation of JWST images in a wide variety of contexts. We use JWST PDRs4All Near-Infrared Camera (NIRCam) and Mid-Infrared Instrument (MIRI) imaging and Near-Infrared Spectrograph (NIRSpec) integral field unit (IFU) and MIRI Medium Resolution Spectrograph (MRS) spectroscopic observations of the Orion Bar photodissociation region (PDR), to directly compare and cross-calibrate imaging and IFU data at ~100 AU resolution over a region where the radiation field and ISM environment evolves from the hot ionized gas to the cold molecular gas. We measure the relative contributions of line, PAH, and continuum emission to the NIRCam and MIRI filters as functions of local physical conditions. We provide empirical prescriptions based on NIRCam and MIRI images to derive intensities of emission lines and PAH features. Within the range of the environments probed in this study, these prescriptions accurately predict Pa-alpha, Br-alpha, PAH 3.3 um and 11.2 um intensities, while those for FeII 1.644 um, H_2 1--0 S(1) 2.12 um and 1--0 S(9) 4.96 um, and PAH 7.7 um show more complicated environmental dependencies. Linear combinations of JWST NIRCam and MIRI images provide effective tracers of ionized gas, H_2, and PAH emission in PDRs. We expect these recipes to be useful for both the Galactic and extragalactic communities., Comment: 31 pages, 24 figures, submitted to A&A
- Published
- 2024
46. Sdn Intrusion Detection Using Machine Learning Method
- Author
-
Mahmud, Muhammad Zawad, Alve, Shahran Rahman, Islam, Samiha, and Khan, Mohammad Monirujjaman
- Subjects
Computer Science - Cryptography and Security ,Computer Science - Machine Learning ,Computer Science - Networking and Internet Architecture - Abstract
Software-defined network (SDN) is a new approach that allows network control to become directly programmable, and the underlying infrastructure can be abstracted from applications and network services. Control plane). When it comes to security, the centralization that this demands is ripe for a variety of cyber threats that are not typically seen in other network architectures. The authors in this research developed a novel machine-learning method to capture infections in networks. We applied the classifier to the UNSW-NB 15 intrusion detection benchmark and trained a model with this data. Random Forest and Decision Tree are classifiers used to assess with Gradient Boosting and AdaBoost. Out of these best-performing models was Gradient Boosting with an accuracy, recall, and F1 score of 99.87%,100%, and 99.85%, respectively, which makes it reliable in the detection of intrusions for SDN networks. The second best-performing classifier was also a Random Forest with 99.38% of accuracy, followed by Ada Boost and Decision Tree. The research shows that the reason that Gradient Boosting is so effective in this task is that it combines weak learners and creates a strong ensemble model that can predict if traffic belongs to a normal or malicious one with high accuracy. This paper indicates that the GBDT-IDS model is able to improve network security significantly and has better features in terms of both real-time detection accuracy and low false positive rates. In future work, we will integrate this model into live SDN space to observe its application and scalability. This research serves as an initial base on which one can make further strides forward to enhance security in SDN using ML techniques and have more secure, resilient networks., Comment: 15 Pages, 14 Figures
- Published
- 2024
47. Efficient Spintronic THz Emitters Without External Magnetic Field
- Author
-
Khan, Amir, Beermann, Nicolas Sylvester, Sharma, Shalini, Schneider, Tiago de Oliveira, Zhang, Wentao, Turchinovich, Dmitry, and Meinert, Markus
- Subjects
Physics - Applied Physics ,Condensed Matter - Mesoscale and Nanoscale Physics ,Condensed Matter - Materials Science - Abstract
We investigate the performance of state-of-the-art spintronic THz emitters (W or Ta)/CoFeB/Pt with non-magnetic underlayer deposited using oblique angle deposition. The THz emission amplitude in the presence or absence of an external magnetic field remains the same and remarkably stable over time. This stability is attributed to the enhanced uniaxial magnetic anisotropy in the ferromagnetic layer, achieved by oblique angle deposition of the underlying non-magnetic layer. Our findings could be used for the development of practical field-free emitters of linearly polarized THz radiation, potentially enabling novel applications in future THz technologies., Comment: 6 pages including references and 5 figures
- Published
- 2024
48. BhasaAnuvaad: A Speech Translation Dataset for 13 Indian Languages
- Author
-
Jain, Sparsh, Sankar, Ashwin, Choudhary, Devilal, Suman, Dhairya, Narasimhan, Nikhil, Khan, Mohammed Safi Ur Rahman, Kunchukuttan, Anoop, Khapra, Mitesh M, and Dabre, Raj
- Subjects
Computer Science - Computation and Language - Abstract
Automatic Speech Translation (AST) datasets for Indian languages remain critically scarce, with public resources covering fewer than 10 of the 22 official languages. This scarcity has resulted in AST systems for Indian languages lagging far behind those available for high-resource languages like English. In this paper, we first evaluate the performance of widely-used AST systems on Indian languages, identifying notable performance gaps and challenges. Our findings show that while these systems perform adequately on read speech, they struggle significantly with spontaneous speech, including disfluencies like pauses and hesitations. Additionally, there is a striking absence of systems capable of accurately translating colloquial and informal language, a key aspect of everyday communication. To this end, we introduce BhasaAnuvaad, the largest publicly available dataset for AST involving 13 out of 22 scheduled Indian languages and English spanning over 44,400 hours and 17M text segments. BhasaAnuvaad contains data for English speech to Indic text, as well as Indic speech to English text. This dataset comprises three key categories: (1) Curated datasets from existing resources, (2) Large-scale web mining, and (3) Synthetic data generation. By offering this diverse and expansive dataset, we aim to bridge the resource gap and promote advancements in AST for Indian languages., Comment: Work in Progress
- Published
- 2024
49. Cybercrime Prediction via Geographically Weighted Learning
- Author
-
Khan, Muhammad Al-Zafar, Al-Karaki, Jamal, and Mahafzah, Emad
- Subjects
Computer Science - Machine Learning - Abstract
Inspired by the success of Geographically Weighted Regression and its accounting for spatial variations, we propose GeogGNN -- A graph neural network model that accounts for geographical latitude and longitudinal points. Using a synthetically generated dataset, we apply the algorithm for a 4-class classification problem in cybersecurity with seemingly realistic geographic coordinates centered in the Gulf Cooperation Council region. We demonstrate that it has higher accuracy than standard neural networks and convolutional neural networks that treat the coordinates as features. Encouraged by the speed-up in model accuracy by the GeogGNN model, we provide a general mathematical result that demonstrates that a geometrically weighted neural network will, in principle, always display higher accuracy in the classification of spatially dependent data by making use of spatial continuity and local averaging features., Comment: 17 pages, 8 figures, Submitted to the International Jordanian Cybersecurity Conference 2024 (IJCC24)
- Published
- 2024
50. Variational Low-Rank Adaptation Using IVON
- Author
-
Cong, Bai, Daheim, Nico, Shen, Yuesong, Cremers, Daniel, Yokota, Rio, Khan, Mohammad Emtiyaz, and Möllenhoff, Thomas
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Computation and Language ,Statistics - Machine Learning - Abstract
We show that variational learning can significantly improve the accuracy and calibration of Low-Rank Adaptation (LoRA) without a substantial increase in the cost. We replace AdamW by the Improved Variational Online Newton (IVON) algorithm to finetune large language models. For Llama-2 with 7 billion parameters, IVON improves the accuracy over AdamW by 2.8% and expected calibration error by 4.6%. The accuracy is also better than the other Bayesian alternatives, yet the cost is lower and the implementation is easier. Our work provides additional evidence for the effectiveness of IVON for large language models. The code is available at https://github.com/team-approx-bayes/ivon-lora., Comment: Published at 38th Workshop on Fine-Tuning in Machine Learning (NeurIPS 2024). Code available at https://github.com/team-approx-bayes/ivon-lora. In version 2 we fixed a typo in the equation of prior in section 2
- Published
- 2024
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.