287 results on '"Alam, Firoj"'
Search Results
2. Improving the Quality of Higher Education using Information and Communication Technologies-Changing Paradigm Shifting
- Author
-
Alam, Firoj
- Published
- 2022
- Full Text
- View/download PDF
3. MemeIntel: Explainable Detection of Propagandistic and Hateful Memes
- Author
-
Kmainasi, Mohamed Bayan, Hasnat, Abul, Hasan, Md Arid, Shahroor, Ali Ezzat, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,68T50 ,I.2.7 - Abstract
The proliferation of multimodal content on social media presents significant challenges in understanding and moderating complex, context-dependent issues such as misinformation, hate speech, and propaganda. While efforts have been made to develop resources and propose new methods for automatic detection, limited attention has been given to label detection and the generation of explanation-based rationales for predicted labels. To address this challenge, we introduce MemeIntel, an explanation-enhanced dataset for propaganda memes in Arabic and hateful memes in English, making it the first large-scale resource for these tasks. To solve these tasks, we propose a multi-stage optimization approach and train Vision-Language Models (VLMs). Our results demonstrate that this approach significantly improves performance over the base model for both \textbf{label detection} and explanation generation, outperforming the current state-of-the-art with an absolute improvement of ~3% on ArMeme and ~7% on Hateful Memes. For reproducibility and future research, we aim to make the MemeIntel dataset and experimental resources publicly available., Comment: disinformation, misinformation, factuality, harmfulness, fake news, propaganda, hateful meme, multimodality, text, images
- Published
- 2025
4. Reasoning About Persuasion: Can LLMs Enable Explainable Propaganda Detection?
- Author
-
Hasanain, Maram, Hasan, Md Arid, Kmainasi, Mohamed Bayan, Sartori, Elisa, Shahroor, Ali Ezzat, Martino, Giovanni Da San, and Alam, Firoj
- Subjects
Computer Science - Computation and Language - Abstract
There has been significant research on propagandistic content detection across different modalities and languages. However, most studies have primarily focused on detection, with little attention given to explanations justifying the predicted label. This is largely due to the lack of resources that provide explanations alongside annotated labels. To address this issue, we propose a multilingual (i.e., Arabic and English) explanation-enhanced dataset, the first of its kind. Additionally, we introduce an explanation-enhanced LLM for both label detection and rationale-based explanation generation. Our findings indicate that the model performs comparably while also generating explanations. We will make the dataset and experimental resources publicly available for the research community.
- Published
- 2025
5. TituLLMs: A Family of Bangla LLMs with Comprehensive Benchmarking
- Author
-
Nahin, Shahriar Kabir, Nandi, Rabindra Nath, Sarker, Sagor, Muhtaseem, Quazi Sarwar, Kowsher, Md, Shill, Apu Chandraw, Ibrahim, Md, Menon, Mehadi Hasan, Muntasir, Tareq Al, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,68T50 ,F.2.2 ,I.2.7 - Abstract
In this paper, we present TituLLMs, the first large pretrained Bangla LLMs, available in 1b and 3b parameter sizes. Due to computational constraints during both training and inference, we focused on smaller models. To train TituLLMs, we collected a pretraining dataset of approximately ~37 billion tokens. We extended the Llama-3.2 tokenizer to incorporate language- and culture-specific knowledge, which also enables faster training and inference. There was a lack of benchmarking datasets to benchmark LLMs for Bangla. To address this gap, we developed five benchmarking datasets. We benchmarked various LLMs, including TituLLMs, and demonstrated that TituLLMs outperforms its initial multilingual versions. However, this is not always the case, highlighting the complexities of language adaptation. Our work lays the groundwork for adapting existing multilingual open models to other low-resource languages. To facilitate broader adoption and further research, we have made the TituLLMs models and benchmarking datasets publicly available (https://huggingface.co/collections/hishab/titulm-llama-family-6718d31fc1b83529276f490a)., Comment: LLMs, Benchmarking, Large Language Models, Bangla
- Published
- 2025
6. BnTTS: Few-Shot Speaker Adaptation in Low-Resource Setting
- Author
-
Basher, Mohammad Jahid Ibna, Kowsher, Md, Islam, Md Saiful, Nandi, Rabindra Nath, Prottasha, Nusrat Jahan, Menon, Mehadi Hasan, Muntasir, Tareq Al, Chowdhury, Shammur Absar, Alam, Firoj, Yousefi, Niloofar, and Garibay, Ozlem Ozmen
- Subjects
Computer Science - Computation and Language - Abstract
This paper introduces BnTTS (Bangla Text-To-Speech), the first framework for Bangla speaker adaptation-based TTS, designed to bridge the gap in Bangla speech synthesis using minimal training data. Building upon the XTTS architecture, our approach integrates Bangla into a multilingual TTS pipeline, with modifications to account for the phonetic and linguistic characteristics of the language. We pre-train BnTTS on 3.85k hours of Bangla speech dataset with corresponding text labels and evaluate performance in both zero-shot and few-shot settings on our proposed test dataset. Empirical evaluations in few-shot settings show that BnTTS significantly improves the naturalness, intelligibility, and speaker fidelity of synthesized Bangla speech. Compared to state-of-the-art Bangla TTS systems, BnTTS exhibits superior performance in Subjective Mean Opinion Score (SMOS), Naturalness, and Clarity metrics., Comment: Accepted paper in NAACL 2025
- Published
- 2025
7. Fanar: An Arabic-Centric Multimodal Generative AI Platform
- Author
-
Fanar Team, Abbas, Ummar, Ahmad, Mohammad Shahmeer, Alam, Firoj, Altinisik, Enes, Asgari, Ehsannedin, Boshmaf, Yazan, Boughorbel, Sabri, Chawla, Sanjay, Chowdhury, Shammur, Dalvi, Fahim, Darwish, Kareem, Durrani, Nadir, Elfeky, Mohamed, Elmagarmid, Ahmed, Eltabakh, Mohamed, Fatehkia, Masoomali, Fragkopoulos, Anastasios, Hasanain, Maram, Hawasly, Majd, Husaini, Mus'ab, Jung, Soon-Gyo, Lucas, Ji Kim, Magdy, Walid, Messaoud, Safa, Mohamed, Abubakr, Mohiuddin, Tasnim, Mousi, Basel, Mubarak, Hamdy, Musleh, Ahmad, Naeem, Zan, Ouzzani, Mourad, Popovic, Dorde, Sadeghi, Amin, Sencar, Husrev Taha, Shinoy, Mohammed, Sinan, Omar, Zhang, Yifan, Ali, Ahmed, Kheir, Yassine El, Ma, Xiaosong, and Ruan, Chaoyi
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,I.2.0 ,D.2.0 - Abstract
We present Fanar, a platform for Arabic-centric multimodal generative AI systems, that supports language, speech and image generation tasks. At the heart of Fanar are Fanar Star and Fanar Prime, two highly capable Arabic Large Language Models (LLMs) that are best in the class on well established benchmarks for similar sized models. Fanar Star is a 7B (billion) parameter model that was trained from scratch on nearly 1 trillion clean and deduplicated Arabic, English and Code tokens. Fanar Prime is a 9B parameter model continually trained on the Gemma-2 9B base model on the same 1 trillion token set. Both models are concurrently deployed and designed to address different types of prompts transparently routed through a custom-built orchestrator. The Fanar platform provides many other capabilities including a customized Islamic Retrieval Augmented Generation (RAG) system for handling religious prompts, a Recency RAG for summarizing information about current or recent events that have occurred after the pre-training data cut-off date. The platform provides additional cognitive capabilities including in-house bilingual speech recognition that supports multiple Arabic dialects, voice and image generation that is fine-tuned to better reflect regional characteristics. Finally, Fanar provides an attribution service that can be used to verify the authenticity of fact based generated content. The design, development, and implementation of Fanar was entirely undertaken at Hamad Bin Khalifa University's Qatar Computing Research Institute (QCRI) and was sponsored by Qatar's Ministry of Communications and Information Technology to enable sovereign AI technology development.
- Published
- 2025
8. GenAI Content Detection Task 3: Cross-Domain Machine-Generated Text Detection Challenge
- Author
-
Dugan, Liam, Zhu, Andrew, Alam, Firoj, Nakov, Preslav, Apidianaki, Marianna, and Callison-Burch, Chris
- Subjects
Computer Science - Computation and Language ,Computer Science - Machine Learning ,I.2.7 - Abstract
Recently there have been many shared tasks targeting the detection of generated text from Large Language Models (LLMs). However, these shared tasks tend to focus either on cases where text is limited to one particular domain or cases where text can be from many domains, some of which may not be seen during test time. In this shared task, using the newly released RAID benchmark, we aim to answer whether or not models can detect generated text from a large, yet fixed, number of domains and LLMs, all of which are seen during training. Over the course of three months, our task was attempted by 9 teams with 23 detector submissions. We find that multiple participants were able to obtain accuracies of over 99% on machine-generated text from RAID while maintaining a 5% False Positive Rate -- suggesting that detectors are able to robustly detect text from many domains and models simultaneously. We discuss potential interpretations of this result and provide directions for future research., Comment: COLING 2025
- Published
- 2025
9. Naringin: Sources, Chemistry, Toxicity, Pharmacokinetics, Pharmacological Evidences, Molecular Docking and Cell line Study
- Author
-
Alam, Firoj, Badruddeen, Kharya, Anil Kumar, Juber, Akhtar, and Khan, Mohammad Irfan
- Published
- 2020
- Full Text
- View/download PDF
10. GenAI Content Detection Task 2: AI vs. Human -- Academic Essay Authenticity Challenge
- Author
-
Chowdhury, Shammur Absar, Almerekhi, Hind, Kutlu, Mucahid, Keles, Kaan Efe, Ahmad, Fatema, Mohiuddin, Tasnim, Mikros, George, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,68T50 ,F.2.2 ,I.2.7 - Abstract
This paper presents a comprehensive overview of the first edition of the Academic Essay Authenticity Challenge, organized as part of the GenAI Content Detection shared tasks collocated with COLING 2025. This challenge focuses on detecting machine-generated vs. human-authored essays for academic purposes. The task is defined as follows: "Given an essay, identify whether it is generated by a machine or authored by a human.'' The challenge involves two languages: English and Arabic. During the evaluation phase, 25 teams submitted systems for English and 21 teams for Arabic, reflecting substantial interest in the task. Finally, seven teams submitted system description papers. The majority of submissions utilized fine-tuned transformer-based models, with one team employing Large Language Models (LLMs) such as Llama 2 and Llama 3. This paper outlines the task formulation, details the dataset construction process, and explains the evaluation framework. Additionally, we present a summary of the approaches adopted by participating teams. Nearly all submitted systems outperformed the n-gram-based baseline, with the top-performing systems achieving F1 scores exceeding 0.98 for both languages, indicating significant progress in the detection of machine-generated text., Comment: AI Generated Content, Academic Essay, LLMs, Arabic, English
- Published
- 2024
11. LlamaLens: Specialized Multilingual LLM for Analyzing News and Social Media Content
- Author
-
Kmainasi, Mohamed Bayan, Shahroor, Ali Ezzat, Hasanain, Maram, Laskar, Sahinur Rahman, Hassan, Naeemul, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,68T50 ,F.2.2 ,I.2.7 - Abstract
Large Language Models (LLMs) have demonstrated remarkable success as general-purpose task solvers across various fields. However, their capabilities remain limited when addressing domain-specific problems, particularly in downstream NLP tasks. Research has shown that models fine-tuned on instruction-based downstream NLP datasets outperform those that are not fine-tuned. While most efforts in this area have primarily focused on resource-rich languages like English and broad domains, little attention has been given to multilingual settings and specific domains. To address this gap, this study focuses on developing a specialized LLM, LlamaLens, for analyzing news and social media content in a multilingual context. To the best of our knowledge, this is the first attempt to tackle both domain specificity and multilinguality, with a particular focus on news and social media. Our experimental setup includes 18 tasks, represented by 52 datasets covering Arabic, English, and Hindi. We demonstrate that LlamaLens outperforms the current state-of-the-art (SOTA) on 23 testing sets, and achieves comparable performance on 8 sets. We make the models and resources publicly available for the research community (https://huggingface.co/collections/QCRI/llamalens-672f7e0604a0498c6a2f0fe9)., Comment: LLMs, Multilingual, Language Diversity, Large Language Models, Social Media, News Media, Specialized LLMs, Fact-checking, Media Analysis, Arabic, Hindi, English
- Published
- 2024
12. AraDiCE: Benchmarks for Dialectal and Cultural Capabilities in LLMs
- Author
-
Mousi, Basel, Durrani, Nadir, Ahmad, Fatema, Hasan, Md. Arid, Hasanain, Maram, Kabbani, Tameem, Dalvi, Fahim, Chowdhury, Shammur Absar, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,68T50 ,F.2.2 ,I.2.7 - Abstract
Arabic, with its rich diversity of dialects, remains significantly underrepresented in Large Language Models, particularly in dialectal variations. We address this gap by introducing seven synthetic datasets in dialects alongside Modern Standard Arabic (MSA), created using Machine Translation (MT) combined with human post-editing. We present AraDiCE, a benchmark for Arabic Dialect and Cultural Evaluation. We evaluate LLMs on dialect comprehension and generation, focusing specifically on low-resource Arabic dialects. Additionally, we introduce the first-ever fine-grained benchmark designed to evaluate cultural awareness across the Gulf, Egypt, and Levant regions, providing a novel dimension to LLM evaluation. Our findings demonstrate that while Arabic-specific models like Jais and AceGPT outperform multilingual models on dialectal tasks, significant challenges persist in dialect identification, generation, and translation. This work contributes $\approx$45K post-edited samples, a cultural benchmark, and highlights the importance of tailored training to improve LLM performance in capturing the nuances of diverse Arabic dialects and cultural contexts. We have released the dialectal translation models and benchmarks developed in this study (https://huggingface.co/datasets/QCRI/AraDiCE)., Comment: Benchmarking, Culturally Informed, Large Language Models, Arabic NLP, LLMs, Arabic Dialect, Dialectal Benchmarking
- Published
- 2024
13. Propaganda to Hate: A Multimodal Analysis of Arabic Memes with Multi-Agent LLMs
- Author
-
Alam, Firoj, Biswas, Md. Rafiul, Shah, Uzair, Zaghouani, Wajdi, and Mikros, Georgios
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,68T50 ,F.2.2 ,I.2.7 - Abstract
In the past decade, social media platforms have been used for information dissemination and consumption. While a major portion of the content is posted to promote citizen journalism and public awareness, some content is posted to mislead users. Among different content types such as text, images, and videos, memes (text overlaid on images) are particularly prevalent and can serve as powerful vehicles for propaganda, hate, and humor. In the current literature, there have been efforts to individually detect such content in memes. However, the study of their intersection is very limited. In this study, we explore the intersection between propaganda and hate in memes using a multi-agent LLM-based approach. We extend the propagandistic meme dataset with coarse and fine-grained hate labels. Our finding suggests that there is an association between propaganda and hate in memes. We provide detailed experimental results that can serve as a baseline for future studies. We will make the experimental resources publicly available to the community (https://github.com/firojalam/propaganda-and-hateful-memes)., Comment: propaganda, hate-speech, disinformation, misinformation, fake news, LLMs, GPT-4, multimodality, multimodal LLMs
- Published
- 2024
14. Native vs Non-Native Language Prompting: A Comparative Analysis
- Author
-
Kmainasi, Mohamed Bayan, Khan, Rakif, Shahroor, Ali Ezzat, Bendou, Boushra, Hasanain, Maram, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,68T50 ,F.2.2 ,I.2.7 - Abstract
Large language models (LLMs) have shown remarkable abilities in different fields, including standard Natural Language Processing (NLP) tasks. To elicit knowledge from LLMs, prompts play a key role, consisting of natural language instructions. Most open and closed source LLMs are trained on available labeled and unlabeled resources--digital content such as text, images, audio, and videos. Hence, these models have better knowledge for high-resourced languages but struggle with low-resourced languages. Since prompts play a crucial role in understanding their capabilities, the language used for prompts remains an important research question. Although there has been significant research in this area, it is still limited, and less has been explored for medium to low-resourced languages. In this study, we investigate different prompting strategies (native vs. non-native) on 11 different NLP tasks associated with 12 different Arabic datasets (9.7K data points). In total, we conducted 197 experiments involving 3 LLMs, 12 datasets, and 3 prompting strategies. Our findings suggest that, on average, the non-native prompt performs the best, followed by mixed and native prompts., Comment: Foundation Models, Large Language Models, Arabic NLP, LLMs, Native, Contextual Understanding, Arabic LLM
- Published
- 2024
15. NativQA: Multilingual Culturally-Aligned Natural Query for LLMs
- Author
-
Hasan, Md. Arid, Hasanain, Maram, Ahmad, Fatema, Laskar, Sahinur Rahman, Upadhyay, Sunaya, Sukhadia, Vrunda N, Kutlu, Mucahid, Chowdhury, Shammur Absar, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,68T50 ,F.2.2 ,I.2.7 - Abstract
Natural Question Answering (QA) datasets play a crucial role in evaluating the capabilities of large language models (LLMs), ensuring their effectiveness in real-world applications. Despite the numerous QA datasets that have been developed, there is a notable lack of region-specific datasets generated by native users in their own languages. This gap hinders the effective benchmarking of LLMs for regional and cultural specificities. Furthermore, it also limits the development of fine-tuned models. In this study, we propose a scalable, language-independent framework, NativQA, to seamlessly construct culturally and regionally aligned QA datasets in native languages, for LLM evaluation and tuning. We demonstrate the efficacy of the proposed framework by designing a multilingual natural QA dataset, \mnqa, consisting of ~64k manually annotated QA pairs in seven languages, ranging from high to extremely low resource, based on queries from native speakers from 9 regions covering 18 topics. We benchmark open- and closed-source LLMs with the MultiNativQA dataset. We also showcase the framework efficacy in constructing fine-tuning data especially for low-resource and dialectally-rich languages. We made both the framework NativQA and MultiNativQA dataset publicly available for the community (https://nativqa.gitlab.io)., Comment: LLMs, Native, Multilingual, Language Diversity, Contextual Understanding, Minority Languages, Culturally Informed, Foundation Models, Large Language Models
- Published
- 2024
16. ArAIEval Shared Task: Propagandistic Techniques Detection in Unimodal and Multimodal Arabic Content
- Author
-
Hasanain, Maram, Hasan, Md. Arid, Ahmed, Fatema, Suwaileh, Reem, Biswas, Md. Rafiul, Zaghouani, Wajdi, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Computer Vision and Pattern Recognition ,68T50 ,F.2.2 ,I.2.7 - Abstract
We present an overview of the second edition of the ArAIEval shared task, organized as part of the ArabicNLP 2024 conference co-located with ACL 2024. In this edition, ArAIEval offers two tasks: (i) detection of propagandistic textual spans with persuasion techniques identification in tweets and news articles, and (ii) distinguishing between propagandistic and non-propagandistic memes. A total of 14 teams participated in the final evaluation phase, with 6 and 9 teams participating in Tasks 1 and 2, respectively. Finally, 11 teams submitted system description papers. Across both tasks, we observed that fine-tuning transformer models such as AraBERT was at the core of the majority of the participating systems. We provide a description of the task setup, including a description of the dataset construction and the evaluation setup. We further provide a brief overview of the participating systems. All datasets and evaluation scripts are released to the research community (https://araieval.gitlab.io/). We hope this will enable further research on these important tasks in Arabic., Comment: propaganda, span detection, disinformation, misinformation, fake news, LLMs, GPT-4, multimodality, multimodal LLMs
- Published
- 2024
17. Propaganda to Hate: A Multimodal Analysis of Arabic Memes with Multi-agent LLMs
- Author
-
Alam, Firoj, Biswas, Md. Rafiul, Shah, Uzair, Zaghouani, Wajdi, Mikros, Georgios, Goos, Gerhard, Series Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Barhamgi, Mahmoud, editor, Wang, Hua, editor, and Wang, Xin, editor
- Published
- 2025
- Full Text
- View/download PDF
18. Native vs Non-native Language Prompting: A Comparative Analysis
- Author
-
Kmainasi, Mohamed Bayan, Khan, Rakif, Shahroor, Ali Ezzat, Bendou, Boushra, Hasanain, Maram, Alam, Firoj, Goos, Gerhard, Series Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Barhamgi, Mahmoud, editor, Wang, Hua, editor, and Wang, Xin, editor
- Published
- 2025
- Full Text
- View/download PDF
19. ThatiAR: Subjectivity Detection in Arabic News Sentences
- Author
-
Suwaileh, Reem, Hasanain, Maram, Hubail, Fatema, Zaghouani, Wajdi, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,68T50 ,F.2.2 ,I.2.7 - Abstract
Detecting subjectivity in news sentences is crucial for identifying media bias, enhancing credibility, and combating misinformation by flagging opinion-based content. It provides insights into public sentiment, empowers readers to make informed decisions, and encourages critical thinking. While research has developed methods and systems for this purpose, most efforts have focused on English and other high-resourced languages. In this study, we present the first large dataset for subjectivity detection in Arabic, consisting of ~3.6K manually annotated sentences, and GPT-4o based explanation. In addition, we included instructions (both in English and Arabic) to facilitate LLM based fine-tuning. We provide an in-depth analysis of the dataset, annotation process, and extensive benchmark results, including PLMs and LLMs. Our analysis of the annotation process highlights that annotators were strongly influenced by their political, cultural, and religious backgrounds, especially at the beginning of the annotation process. The experimental results suggest that LLMs with in-context learning provide better performance. We aim to release the dataset and resources for the community., Comment: Subjectivity, Sentiment, Disinformation, Misinformation, Fake news, LLMs, Transformers, Instruction Dataset
- Published
- 2024
20. ArMeme: Propagandistic Content in Arabic Memes
- Author
-
Alam, Firoj, Hasnat, Abul, Ahmed, Fatema, Hasan, Md Arid, and Hasanain, Maram
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Computer Vision and Pattern Recognition ,68T50 ,I.2.7 - Abstract
With the rise of digital communication, memes have become a significant medium for cultural and political expression that is often used to mislead audiences. Identification of such misleading and persuasive multimodal content has become more important among various stakeholders, including social media platforms, policymakers, and the broader society as they often cause harm to individuals, organizations, and/or society. While there has been effort to develop AI-based automatic systems for resource-rich languages (e.g., English), it is relatively little to none for medium to low resource languages. In this study, we focused on developing an Arabic memes dataset with manual annotations of propagandistic content. We annotated ~6K Arabic memes collected from various social media platforms, which is a first resource for Arabic multimodal research. We provide a comprehensive analysis aiming to develop computational tools for their detection. We will make them publicly available for the community., Comment: disinformation, misinformation, factuality, harmfulness, fake news, propaganda, multimodality, text, images
- Published
- 2024
21. Can GPT-4 Identify Propaganda? Annotation and Detection of Propaganda Spans in News Articles
- Author
-
Hasanain, Maram, Ahmed, Fatema, and Alam, Firoj
- Subjects
Computer Science - Computation and Language - Abstract
The use of propaganda has spiked on mainstream and social media, aiming to manipulate or mislead users. While efforts to automatically detect propaganda techniques in textual, visual, or multimodal content have increased, most of them primarily focus on English content. The majority of the recent initiatives targeting medium to low-resource languages produced relatively small annotated datasets, with a skewed distribution, posing challenges for the development of sophisticated propaganda detection models. To address this challenge, we carefully develop the largest propaganda dataset to date, ArPro, comprised of 8K paragraphs from newspaper articles, labeled at the text span level following a taxonomy of 23 propagandistic techniques. Furthermore, our work offers the first attempt to understand the performance of large language models (LLMs), using GPT-4, for fine-grained propaganda detection from text. Results showed that GPT-4's performance degrades as the task moves from simply classifying a paragraph as propagandistic or not, to the fine-grained task of detecting propaganda techniques and their manifestation in text. Compared to models fine-tuned on the dataset for propaganda detection at different classification granularities, GPT-4 is still far behind. Finally, we evaluate GPT-4 on a dataset consisting of six other languages for span detection, and results suggest that the model struggles with the task across languages. Our dataset and resources will be released to the community., Comment: Accepted as a full paper at LREC-COLING 2024
- Published
- 2024
22. Large Language Models for Propaganda Span Annotation
- Author
-
Hasanain, Maram, Ahmad, Fatema, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,68T50 ,F.2.2 ,I.2.7 - Abstract
The use of propagandistic techniques in online content has increased in recent years aiming to manipulate online audiences. Fine-grained propaganda detection and extraction of textual spans where propaganda techniques are used, are essential for more informed content consumption. Automatic systems targeting the task over lower resourced languages are limited, usually obstructed by lack of large scale training datasets. Our study investigates whether Large Language Models (LLMs), such as GPT-4, can effectively extract propagandistic spans. We further study the potential of employing the model to collect more cost-effective annotations. Finally, we examine the effectiveness of labels provided by GPT-4 in training smaller language models for the task. The experiments are performed over a large-scale in-house manually annotated dataset. The results suggest that providing more annotation context to GPT-4 within prompts improves its performance compared to human annotators. Moreover, when serving as an expert annotator (consolidator), the model provides labels that have higher agreement with expert annotators, and lead to specialized models that achieve state-of-the-art over an unseen Arabic testing set. Finally, our work is the first to show the potential of utilizing LLMs to develop annotated datasets for propagandistic spans detection task prompting it with annotations from human annotators with limited expertise. All scripts and annotations will be shared with the community., Comment: propaganda, span detection, disinformation, misinformation, fake news, LLMs, GPT-4
- Published
- 2023
23. Pseudo-Labeling for Domain-Agnostic Bangla Automatic Speech Recognition
- Author
-
Nandi, Rabindra Nath, Menon, Mehadi Hasan, Muntasir, Tareq Al, Sarker, Sagor, Muhtaseem, Quazi Sarwar, Islam, Md. Tariqul, Chowdhury, Shammur Absar, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,68T50 ,F.2.2 ,I.2.7 - Abstract
One of the major challenges for developing automatic speech recognition (ASR) for low-resource languages is the limited access to labeled data with domain-specific variations. In this study, we propose a pseudo-labeling approach to develop a large-scale domain-agnostic ASR dataset. With the proposed methodology, we developed a 20k+ hours labeled Bangla speech dataset covering diverse topics, speaking styles, dialects, noisy environments, and conversational scenarios. We then exploited the developed corpus to design a conformer-based ASR system. We benchmarked the trained ASR with publicly available datasets and compared it with other available models. To investigate the efficacy, we designed and developed a human-annotated domain-agnostic test set composed of news, telephony, and conversational data among others. Our results demonstrate the efficacy of the model trained on psuedo-label data for the designed test-set along with publicly-available Bangla datasets. The experimental resources will be publicly available.(https://github.com/hishab-nlp/Pseudo-Labeling-for-Domain-Agnostic-Bangla-ASR), Comment: Accepted at BLP-2023 (at EMNLP 2023), ASR, low-resource, out-of-distribution, domain-agnostic
- Published
- 2023
24. Nexus at ArAIEval Shared Task: Fine-Tuning Arabic Language Models for Propaganda and Disinformation Detection
- Author
-
Xiao, Yunze and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Social and Information Networks ,68T50 ,F.2.2 ,I.2.7 - Abstract
The spread of disinformation and propagandistic content poses a threat to societal harmony, undermining informed decision-making and trust in reliable sources. Online platforms often serve as breeding grounds for such content, and malicious actors exploit the vulnerabilities of audiences to shape public opinion. Although there have been research efforts aimed at the automatic identification of disinformation and propaganda in social media content, there remain challenges in terms of performance. The ArAIEval shared task aims to further research on these particular issues within the context of the Arabic language. In this paper, we discuss our participation in these shared tasks. We competed in subtasks 1A and 2A, where our submitted system secured positions 9th and 10th, respectively. Our experiments consist of fine-tuning transformer models and using zero- and few-shot learning with GPT-4., Comment: propaganda, disinformation, misinformation, fake news
- Published
- 2023
25. ArAIEval Shared Task: Persuasion Techniques and Disinformation Detection in Arabic Text
- Author
-
Hasanain, Maram, Alam, Firoj, Mubarak, Hamdy, Abdaljalil, Samir, Zaghouani, Wajdi, Nakov, Preslav, Martino, Giovanni Da San, and Freihat, Abed Alhakim
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,68T50 ,F.2.2 ,I.2.7 - Abstract
We present an overview of the ArAIEval shared task, organized as part of the first ArabicNLP 2023 conference co-located with EMNLP 2023. ArAIEval offers two tasks over Arabic text: (i) persuasion technique detection, focusing on identifying persuasion techniques in tweets and news articles, and (ii) disinformation detection in binary and multiclass setups over tweets. A total of 20 teams participated in the final evaluation phase, with 14 and 16 teams participating in Tasks 1 and 2, respectively. Across both tasks, we observed that fine-tuning transformer models such as AraBERT was at the core of the majority of the participating systems. We provide a description of the task setup, including a description of the dataset construction and the evaluation setup. We further give a brief overview of the participating systems. All datasets and evaluation scripts from the shared task are released to the research community. (https://araieval.gitlab.io/) We hope this will enable further research on these important tasks in Arabic., Comment: Accepted at ArabicNLP-23 (EMNLP-23), propaganda, disinformation, misinformation, fake news
- Published
- 2023
26. BLP-2023 Task 2: Sentiment Analysis
- Author
-
Hasan, Md. Arid, Alam, Firoj, Anjum, Anika, Das, Shudipta, and Anjum, Afiyat
- Subjects
Computer Science - Computation and Language ,Computer Science - Machine Learning ,I.2.7 - Abstract
We present an overview of the BLP Sentiment Shared Task, organized as part of the inaugural BLP 2023 workshop, co-located with EMNLP 2023. The task is defined as the detection of sentiment in a given piece of social media text. This task attracted interest from 71 participants, among whom 29 and 30 teams submitted systems during the development and evaluation phases, respectively. In total, participants submitted 597 runs. However, a total of 15 teams submitted system description papers. The range of approaches in the submitted systems spans from classical machine learning models, fine-tuning pre-trained models, to leveraging Large Language Model (LLMs) in zero- and few-shot settings. In this paper, we provide a detailed account of the task setup, including dataset development and evaluation setup. Additionally, we provide a brief overview of the systems submitted by the participants. All datasets and evaluation scripts from the shared task have been made publicly available for the research community, to foster further research in this domain., Comment: Accepted in BLP Workshop at EMNLP-23
- Published
- 2023
- Full Text
- View/download PDF
27. Zero- and Few-Shot Prompting with LLMs: A Comparative Study with Fine-tuned Models for Bangla Sentiment Analysis
- Author
-
Hasan, Md. Arid, Das, Shudipta, Anjum, Afiyat, Alam, Firoj, Anjum, Anika, Sarker, Avijit, and Noori, Sheak Rashed Haider
- Subjects
Computer Science - Computation and Language ,Computer Science - Machine Learning ,68T50 ,I.2.7 - Abstract
The rapid expansion of the digital world has propelled sentiment analysis into a critical tool across diverse sectors such as marketing, politics, customer service, and healthcare. While there have been significant advancements in sentiment analysis for widely spoken languages, low-resource languages, such as Bangla, remain largely under-researched due to resource constraints. Furthermore, the recent unprecedented performance of Large Language Models (LLMs) in various applications highlights the need to evaluate them in the context of low-resource languages. In this study, we present a sizeable manually annotated dataset encompassing 33,606 Bangla news tweets and Facebook comments. We also investigate zero- and few-shot in-context learning with several language models, including Flan-T5, GPT-4, and Bloomz, offering a comparative analysis against fine-tuned models. Our findings suggest that monolingual transformer-based models consistently outperform other models, even in zero and few-shot scenarios. To foster continued exploration, we intend to make this dataset and our research tools publicly available to the broader research community., Comment: Accepted at LREC-COLING 2024. Zero-Shot Prompting, Few-Shot Prompting, LLMs, Comparative Study, Fine-tuned Models, Bangla, Sentiment Analysis
- Published
- 2023
28. LLMeBench: A Flexible Framework for Accelerating LLMs Benchmarking
- Author
-
Dalvi, Fahim, Hasanain, Maram, Boughorbel, Sabri, Mousi, Basel, Abdaljalil, Samir, Nazar, Nizi, Abdelali, Ahmed, Chowdhury, Shammur Absar, Mubarak, Hamdy, Ali, Ahmed, Hawasly, Majd, Durrani, Nadir, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,68T50 ,F.2.2 ,I.2.7 - Abstract
The recent development and success of Large Language Models (LLMs) necessitate an evaluation of their performance across diverse NLP tasks in different languages. Although several frameworks have been developed and made publicly available, their customization capabilities for specific tasks and datasets are often complex for different users. In this study, we introduce the LLMeBench framework, which can be seamlessly customized to evaluate LLMs for any NLP task, regardless of language. The framework features generic dataset loaders, several model providers, and pre-implements most standard evaluation metrics. It supports in-context learning with zero- and few-shot settings. A specific dataset and task can be evaluated for a given LLM in less than 20 lines of code while allowing full flexibility to extend the framework for custom datasets, models, or tasks. The framework has been tested on 31 unique NLP tasks using 53 publicly available datasets within 90 experimental setups, involving approximately 296K data points. We open-sourced LLMeBench for the community (https://github.com/qcri/LLMeBench/) and a video demonstrating the framework is available online. (https://youtu.be/9cC2m_abk3A), Comment: Accepted as a demo paper at EACL 2024
- Published
- 2023
29. LAraBench: Benchmarking Arabic AI with Large Language Models
- Author
-
Abdelali, Ahmed, Mubarak, Hamdy, Chowdhury, Shammur Absar, Hasanain, Maram, Mousi, Basel, Boughorbel, Sabri, Kheir, Yassine El, Izham, Daniel, Dalvi, Fahim, Hawasly, Majd, Nazar, Nizi, Elshahawy, Yousseif, Ali, Ahmed, Durrani, Nadir, Milic-Frayling, Natasa, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,68T50 ,F.2.2 ,I.2.7 - Abstract
Recent advancements in Large Language Models (LLMs) have significantly influenced the landscape of language and speech research. Despite this progress, these models lack specific benchmarking against state-of-the-art (SOTA) models tailored to particular languages and tasks. LAraBench addresses this gap for Arabic Natural Language Processing (NLP) and Speech Processing tasks, including sequence tagging and content classification across different domains. We utilized models such as GPT-3.5-turbo, GPT-4, BLOOMZ, Jais-13b-chat, Whisper, and USM, employing zero and few-shot learning techniques to tackle 33 distinct tasks across 61 publicly available datasets. This involved 98 experimental setups, encompassing ~296K data points, ~46 hours of speech, and 30 sentences for Text-to-Speech (TTS). This effort resulted in 330+ sets of experiments. Our analysis focused on measuring the performance gap between SOTA models and LLMs. The overarching trend observed was that SOTA models generally outperformed LLMs in zero-shot learning, with a few exceptions. Notably, larger computational models with few-shot learning techniques managed to reduce these performance gaps. Our findings provide valuable insights into the applicability of LLMs for Arabic NLP and speech processing tasks., Comment: Foundation Models, Large Language Models, Arabic NLP, Arabic Speech, Arabic AI, GPT3.5 Evaluation, USM Evaluation, Whisper Evaluation, GPT-4, BLOOMZ, Jais13b
- Published
- 2023
30. Detecting and Reasoning of Deleted Tweets before they are Posted
- Author
-
Mubarak, Hamdy, Abdaljalil, Samir, Nassar, Azza, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Computers and Society ,68T50 ,F.2.2 ,I.2.7 - Abstract
Social media platforms empower us in several ways, from information dissemination to consumption. While these platforms are useful in promoting citizen journalism, public awareness etc., they have misuse potentials. Malicious users use them to disseminate hate-speech, offensive content, rumor etc. to gain social and political agendas or to harm individuals, entities and organizations. Often times, general users unconsciously share information without verifying it, or unintentionally post harmful messages. Some of such content often get deleted either by the platform due to the violation of terms and policies, or users themselves for different reasons, e.g., regrets. There is a wide range of studies in characterizing, understanding and predicting deleted content. However, studies which aims to identify the fine-grained reasons (e.g., posts are offensive, hate speech or no identifiable reason) behind deleted content, are limited. In this study we address this gap, by identifying deleted tweets, particularly within the Arabic context, and labeling them with a corresponding fine-grained disinformation category. We then develop models that can predict the potentiality of tweets getting deleted, as well as the potential reasons behind deletion. Such models can help in moderating social media posts before even posting., Comment: disinformation, misinformation, fake news
- Published
- 2023
31. QCRI at SemEval-2023 Task 3: News Genre, Framing and Persuasion Techniques Detection using Multilingual Models
- Author
-
Hasanain, Maram, El-Shangiti, Ahmed Oumar, Nandi, Rabindra Nath, Nakov, Preslav, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Computers and Society ,68T50 ,F.2.2 ,I.2.7 - Abstract
Misinformation spreading in mainstream and social media has been misleading users in different ways. Manual detection and verification efforts by journalists and fact-checkers can no longer cope with the great scale and quick spread of misleading information. This motivated research and industry efforts to develop systems for analyzing and verifying news spreading online. The SemEval-2023 Task 3 is an attempt to address several subtasks under this overarching problem, targeting writing techniques used in news articles to affect readers' opinions. The task addressed three subtasks with six languages, in addition to three ``surprise'' test languages, resulting in 27 different test setups. This paper describes our participating system to this task. Our team is one of the 6 teams that successfully submitted runs for all setups. The official results show that our system is ranked among the top 3 systems for 10 out of the 27 setups., Comment: Accepted at SemEval-23 (ACL-23, propaganda, disinformation, misinformation, fake news
- Published
- 2023
32. Overview of the WANLP 2022 Shared Task on Propaganda Detection in Arabic
- Author
-
Alam, Firoj, Mubarak, Hamdy, Zaghouani, Wajdi, Martino, Giovanni Da San, and Nakov, Preslav
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning ,68T50 ,F.2.2 ,I.2.7 - Abstract
Propaganda is the expression of an opinion or an action by an individual or a group deliberately designed to influence the opinions or the actions of other individuals or groups with reference to predetermined ends, which is achieved by means of well-defined rhetorical and psychological devices. Propaganda techniques are commonly used in social media to manipulate or to mislead users. Thus, there has been a lot of recent research on automatic detection of propaganda techniques in text as well as in memes. However, so far the focus has been primarily on English. With the aim to bridge this language gap, we ran a shared task on detecting propaganda techniques in Arabic tweets as part of the WANLP 2022 workshop, which included two subtasks. Subtask~1 asks to identify the set of propaganda techniques used in a tweet, which is a multilabel classification problem, while Subtask~2 asks to detect the propaganda techniques used in a tweet together with the exact span(s) of text in which each propaganda technique appears. The task attracted 63 team registrations, and eventually 14 and 3 teams made submissions for subtask 1 and 2, respectively. Finally, 11 teams submitted system description papers., Comment: Accepted at WANLP-22 (EMNLP-22), propaganda, disinformation, misinformation, fake news, memes, multimodality. arXiv admin note: text overlap with arXiv:2109.08013, arXiv:2105.09284
- Published
- 2022
33. ConceptX: A Framework for Latent Concept Analysis
- Author
-
Alam, Firoj, Dalvi, Fahim, Durrani, Nadir, Sajjad, Hassan, Khan, Abdul Rafae, and Xu, Jia
- Subjects
Computer Science - Computation and Language - Abstract
The opacity of deep neural networks remains a challenge in deploying solutions where explanation is as important as precision. We present ConceptX, a human-in-the-loop framework for interpreting and annotating latent representational space in pre-trained Language Models (pLMs). We use an unsupervised method to discover concepts learned in these models and enable a graphical interface for humans to generate explanations for the concepts. To facilitate the process, we provide auto-annotations of the concepts (based on traditional linguistic ontologies). Such annotations enable development of a linguistic resource that directly represents latent concepts learned within deep NLP models. These include not just traditional linguistic concepts, but also task-specific or sensitive concepts (words grouped based on gender or religious connotation) that helps the annotators to mark bias in the model. The framework consists of two parts (i) concept discovery and (ii) annotation platform., Comment: AAAI 23
- Published
- 2022
34. On the Transformation of Latent Space in Fine-Tuned NLP Models
- Author
-
Durrani, Nadir, Sajjad, Hassan, Dalvi, Fahim, and Alam, Firoj
- Subjects
Computer Science - Computation and Language - Abstract
We study the evolution of latent space in fine-tuned NLP models. Different from the commonly used probing-framework, we opt for an unsupervised method to analyze representations. More specifically, we discover latent concepts in the representational space using hierarchical clustering. We then use an alignment function to gauge the similarity between the latent space of a pre-trained model and its fine-tuned version. We use traditional linguistic concepts to facilitate our understanding and also study how the model space transforms towards task-specific information. We perform a thorough analysis, comparing pre-trained and fine-tuned models across three models and three downstream tasks. The notable findings of our work are: i) the latent space of the higher layers evolve towards task-specific concepts, ii) whereas the lower layers retain generic concepts acquired in the pre-trained model, iii) we discovered that some concepts in the higher layers acquire polarity towards the output class, and iv) that these concepts can be used for generating adversarial triggers., Comment: EMNLP 2022
- Published
- 2022
35. Overview of the CLEF-2024 CheckThat! Lab: Check-Worthiness, Subjectivity, Persuasion, Roles, Authorities, and Adversarial Robustness
- Author
-
Barrón-Cedeño, Alberto, Alam, Firoj, Struß, Julia Maria, Nakov, Preslav, Chakraborty, Tanmoy, Elsayed, Tamer, Przybyła, Piotr, Caselli, Tommaso, Da San Martino, Giovanni, Haouari, Fatima, Hasanain, Maram, Li, Chengkai, Piskorski, Jakub, Ruggeri, Federico, Song, Xingyi, Suwaileh, Reem, Goos, Gerhard, Series Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Goeuriot, Lorraine, editor, Mulhem, Philippe, editor, Quénot, Georges, editor, Schwab, Didier, editor, Di Nunzio, Giorgio Maria, editor, Soulier, Laure, editor, Galuščáková, Petra, editor, García Seco de Herrera, Alba, editor, Faggioli, Guglielmo, editor, and Ferro, Nicola, editor
- Published
- 2024
- Full Text
- View/download PDF
36. The CLEF-2024 CheckThat! Lab: Check-Worthiness, Subjectivity, Persuasion, Roles, Authorities, and Adversarial Robustness
- Author
-
Barrón-Cedeño, Alberto, Alam, Firoj, Chakraborty, Tanmoy, Elsayed, Tamer, Nakov, Preslav, Przybyła, Piotr, Struß, Julia Maria, Haouari, Fatima, Hasanain, Maram, Ruggeri, Federico, Song, Xingyi, Suwaileh, Reem, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Goharian, Nazli, editor, Tonellotto, Nicola, editor, He, Yulan, editor, Lipani, Aldo, editor, McDonald, Graham, editor, Macdonald, Craig, editor, and Ounis, Iadh, editor
- Published
- 2024
- Full Text
- View/download PDF
37. Z-Index at CheckThat! Lab 2022: Check-Worthiness Identification on Tweet Text
- Author
-
Tarannum, Prerona, Alam, Firoj, Hasan, Md. Arid, and Noori, Sheak Rashed Haider
- Subjects
Computer Science - Computation and Language ,Computer Science - Machine Learning ,I.2.7 - Abstract
The wide use of social media and digital technologies facilitates sharing various news and information about events and activities. Despite sharing positive information misleading and false information is also spreading on social media. There have been efforts in identifying such misleading information both manually by human experts and automatic tools. Manual effort does not scale well due to the high volume of information, containing factual claims, are appearing online. Therefore, automatically identifying check-worthy claims can be very useful for human experts. In this study, we describe our participation in Subtask-1A: Check-worthiness of tweets (English, Dutch and Spanish) of CheckThat! lab at CLEF 2022. We performed standard preprocessing steps and applied different models to identify whether a given text is worthy of fact checking or not. We use the oversampling technique to balance the dataset and applied SVM and Random Forest (RF) with TF-IDF representations. We also used BERT multilingual (BERT-m) and XLM-RoBERTa-base pre-trained models for the experiments. We used BERT-m for the official submissions and our systems ranked as 3rd, 5th, and 12th in Spanish, Dutch, and English, respectively. In further experiments, our evaluation shows that transformer models (BERT-m and XLM-RoBERTa-base) outperform the SVM and RF in Dutch and English languages where a different scenario is observed for Spanish., Comment: Accepted in CLEF 2022
- Published
- 2022
38. Analyzing Encoded Concepts in Transformer Language Models
- Author
-
Sajjad, Hassan, Durrani, Nadir, Dalvi, Fahim, Alam, Firoj, Khan, Abdul Rafae, and Xu, Jia
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence - Abstract
We propose a novel framework ConceptX, to analyze how latent concepts are encoded in representations learned within pre-trained language models. It uses clustering to discover the encoded concepts and explains them by aligning with a large set of human-defined concepts. Our analysis on seven transformer language models reveal interesting insights: i) the latent space within the learned representations overlap with different linguistic concepts to a varying degree, ii) the lower layers in the model are dominated by lexical concepts (e.g., affixation), whereas the core-linguistic concepts (e.g., morphological or syntactic relations) are better represented in the middle and higher layers, iii) some encoded concepts are multi-faceted and cannot be adequately explained using the existing human-defined concepts., Comment: 20 pages, 10 figures
- Published
- 2022
39. Discovering Latent Concepts Learned in BERT
- Author
-
Dalvi, Fahim, Khan, Abdul Rafae, Alam, Firoj, Durrani, Nadir, Xu, Jia, and Sajjad, Hassan
- Subjects
Computer Science - Computation and Language - Abstract
A large number of studies that analyze deep neural network models and their ability to encode various linguistic and non-linguistic concepts provide an interpretation of the inner mechanics of these models. The scope of the analyses is limited to pre-defined concepts that reinforce the traditional linguistic knowledge and do not reflect on how novel concepts are learned by the model. We address this limitation by discovering and analyzing latent concepts learned in neural network models in an unsupervised fashion and provide interpretations from the model's perspective. In this work, we study: i) what latent concepts exist in the pre-trained BERT model, ii) how the discovered latent concepts align or diverge from classical linguistic hierarchy and iii) how the latent concepts evolve across layers. Our findings show: i) a model learns novel concepts (e.g. animal categories and demographic groups), which do not strictly adhere to any pre-defined categorization (e.g. POS, semantic tags), ii) several latent concepts are based on multiple properties which may include semantics, syntax, and morphology, iii) the lower layers in the model dominate in learning shallow lexical concepts while the higher layers learn semantic relations and iv) the discovered latent concepts highlight potential biases learned in the model. We also release a novel BERT ConceptNet dataset (BCN) consisting of 174 concept labels and 1M annotated instances., Comment: ICLR 2022
- Published
- 2022
40. Detecting and Understanding Harmful Memes: A Survey
- Author
-
Sharma, Shivam, Alam, Firoj, Akhtar, Md. Shad, Dimitrov, Dimitar, Martino, Giovanni Da San, Firooz, Hamed, Halevy, Alon, Silvestri, Fabrizio, Nakov, Preslav, and Chakraborty, Tanmoy
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Computer Vision and Pattern Recognition - Abstract
The automatic identification of harmful content online is of major concern for social media platforms, policymakers, and society. Researchers have studied textual, visual, and audio content, but typically in isolation. Yet, harmful content often combines multiple modalities, as in the case of memes, which are of particular interest due to their viral nature. With this in mind, here we offer a comprehensive survey with a focus on harmful memes. Based on a systematic analysis of recent literature, we first propose a new typology of harmful memes, and then we highlight and summarize the relevant state of the art. One interesting finding is that many types of harmful memes are not really studied, e.g., such featuring self-harm and extremism, partly due to the lack of suitable datasets. We further find that existing datasets mostly capture multi-class scenarios, which are not inclusive of the affective spectrum that memes can represent. Another observation is that memes can propagate globally through repackaging in different languages and that they can also be multilingual, blending different cultures. We conclude by highlighting several challenges related to multimodal semiotics, technological constraints, and non-trivial social engagement, and we present several open-ended aspects such as delineating online harm and empirically examining related frameworks and assistive interventions, which we believe will motivate and drive future research., Comment: Accepted at IJCAI-ECAI 2022 (Survey Track) - Editorial Feedback Revised, 9 pages (7 main + 2 reference pages)
- Published
- 2022
41. TeamX@DravidianLangTech-ACL2022: A Comparative Analysis for Troll-Based Meme Classification
- Author
-
Nandi, Rabindra Nath, Alam, Firoj, and Nakov, Preslav
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Multimedia ,Computer Science - Social and Information Networks ,68T50 ,I.2.7 - Abstract
The spread of fake news, propaganda, misinformation, disinformation, and harmful content online raised concerns among social media platforms, government agencies, policymakers, and society as a whole. This is because such harmful or abusive content leads to several consequences to people such as physical, emotional, relational, and financial. Among different harmful content \textit{trolling-based} online content is one of them, where the idea is to post a message that is provocative, offensive, or menacing with an intent to mislead the audience. The content can be textual, visual, a combination of both, or a meme. In this study, we provide a comparative analysis of troll-based memes classification using the textual, visual, and multimodal content. We report several interesting findings in terms of code-mixed text, multimodal setting, and combining an additional dataset, which shows improvements over the majority baseline., Comment: Accepted at DravidianLangTech-ACL2022 (Colocated with ACL-2022). disinformation, misinformation, factuality, harmfulness, fake news, propaganda, multimodality, text, images, videos, network structure, temporality
- Published
- 2022
42. Detecting the Role of an Entity in Harmful Memes: Techniques and Their Limitations
- Author
-
Nandi, Rabindra Nath, Alam, Firoj, and Nakov, Preslav
- Subjects
Computer Science - Computation and Language ,Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Multimedia ,Computer Science - Social and Information Networks ,68T50 ,I.2.7 - Abstract
Harmful or abusive online content has been increasing over time, raising concerns for social media platforms, government agencies, and policymakers. Such harmful or abusive content can have major negative impact on society, e.g., cyberbullying can lead to suicides, rumors about COVID-19 can cause vaccine hesitance, promotion of fake cures for COVID-19 can cause health harms and deaths. The content that is posted and shared online can be textual, visual, or a combination of both, e.g., in a meme. Here, we describe our experiments in detecting the roles of the entities (hero, villain, victim) in harmful memes, which is part of the CONSTRAINT-2022 shared task, as well as our system for the task. We further provide a comparative analysis of different experimental settings (i.e., unimodal, multimodal, attention, and augmentation). For reproducibility, we make our experimental code publicly available. \url{https://github.com/robi56/harmful_memes_block_fusion}, Comment: Accepted at CONSTRAINT 2022 (Colocated with ACL-2022), disinformation, misinformation, factuality, harmfulness, fake news, propaganda, multimodality, text, images, videos, network structure, temporality
- Published
- 2022
43. QCRI's COVID-19 Disinformation Detector: A System to Fight the COVID-19 Infodemic in Social Media
- Author
-
Nakov, Preslav, Alam, Firoj, Zhang, Yifan, Prakash, Animesh, and Dalvi, Fahim
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Computers and Society ,Computer Science - Information Retrieval ,Computer Science - Social and Information Networks ,68T50 ,I.2 ,I.2.7 - Abstract
Fighting the ongoing COVID-19 infodemic has been declared as one of the most important focus areas by the World Health Organization since the onset of the COVID-19 pandemic. While the information that is consumed and disseminated consists of promoting fake cures, rumors, and conspiracy theories to spreading xenophobia and panic, at the same time there is information (e.g., containing advice, promoting cure) that can help different stakeholders such as policy-makers. Social media platforms enable the infodemic and there has been an effort to curate the content on such platforms, analyze and debunk them. While a majority of the research efforts consider one or two aspects (e.g., detecting factuality) of such information, in this study we focus on a multifaceted approach, including an API,\url{https://app.swaggerhub.com/apis/yifan2019/Tanbih/0.8.0/} and a demo system,\url{https://covid19.tanbih.org}, which we made freely and publicly available. We believe that this will facilitate researchers and different stakeholders. A screencast of the API services and demo is available.\url{https://youtu.be/zhbcSvxEKMk}, Comment: disinformation, misinformation, factuality, fact-checking, fact-checkers, check-worthiness, Social Media Platforms, COVID-19, social media
- Published
- 2022
44. ArabGend: Gender Analysis and Inference on Arabic Twitter
- Author
-
Mubarak, Hamdy, Chowdhury, Shammur Absar, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Computers and Society ,Computer Science - Social and Information Networks ,68T50 ,I.2.7 - Abstract
Gender analysis of Twitter can reveal important socio-cultural differences between male and female users. There has been a significant effort to analyze and automatically infer gender in the past for most widely spoken languages' content, however, to our knowledge very limited work has been done for Arabic. In this paper, we perform an extensive analysis of differences between male and female users on the Arabic Twitter-sphere. We study differences in user engagement, topics of interest, and the gender gap in professions. Along with gender analysis, we also propose a method to infer gender by utilizing usernames, profile pictures, tweets, and networks of friends. In order to do so, we manually annotated gender and locations for ~166K Twitter accounts associated with ~92K user location, which we plan to make publicly available at http://anonymous.com. Our proposed gender inference method achieve an F1 score of 82.1%, which is 47.3% higher than majority baseline. In addition, we also developed a demo and made it publicly available., Comment: Gender Analysis Dataset, Demography, Arabic Twitter Accounts, Arabic Social Media Content
- Published
- 2022
45. ArCovidVac: Analyzing Arabic Tweets About COVID-19 Vaccination
- Author
-
Mubarak, Hamdy, Hassan, Sabit, Chowdhury, Shammur Absar, and Alam, Firoj
- Subjects
Computer Science - Computation and Language ,Computer Science - Social and Information Networks - Abstract
The emergence of the COVID-19 pandemic and the first global infodemic have changed our lives in many different ways. We relied on social media to get the latest information about the COVID-19 pandemic and at the same time to disseminate information. The content in social media consisted not only health related advises, plans, and informative news from policy makers, but also contains conspiracies and rumors. It became important to identify such information as soon as they are posted to make actionable decisions (e.g., debunking rumors, or taking certain measures for traveling). To address this challenge, we develop and publicly release the first largest manually annotated Arabic tweet dataset, ArCovidVac, for the COVID-19 vaccination campaign, covering many countries in the Arab region. The dataset is enriched with different layers of annotation, including, (i) Informativeness (more vs. less importance of the tweets); (ii) fine-grained tweet content types (e.g., advice, rumors, restriction, authenticate news/information); and (iii) stance towards vaccination (pro-vaccination, neutral, anti-vaccination). Further, we performed in-depth analysis of the data, exploring the popularity of different vaccines, trending hashtags, topics and presence of offensiveness in the tweets. We studied the data for individual types of tweets and temporal changes in stance towards vaccine. We benchmarked the ArCovidVac dataset using transformer architectures for informativeness, content types, and stance detection., Comment: 8 pages, 9 figures
- Published
- 2022
46. Adversarial Attacks and Defenses for Social Network Text Processing Applications: Techniques, Challenges and Future Research Directions
- Author
-
Alsmadi, Izzat, Ahmad, Kashif, Nazzal, Mahmoud, Alam, Firoj, Al-Fuqaha, Ala, Khreishah, Abdallah, and Algosaibi, Abdulelah
- Subjects
Computer Science - Computation and Language - Abstract
The growing use of social media has led to the development of several Machine Learning (ML) and Natural Language Processing(NLP) tools to process the unprecedented amount of social media content to make actionable decisions. However, these MLand NLP algorithms have been widely shown to be vulnerable to adversarial attacks. These vulnerabilities allow adversaries to launch a diversified set of adversarial attacks on these algorithms in different applications of social media text processing. In this paper, we provide a comprehensive review of the main approaches for adversarial attacks and defenses in the context of social media applications with a particular focus on key challenges and future research directions. In detail, we cover literature on six key applications, namely (i) rumors detection, (ii) satires detection, (iii) clickbait & spams identification, (iv) hate speech detection, (v)misinformation detection, and (vi) sentiment analysis. We then highlight the concurrent and anticipated future research questions and provide recommendations and directions for future work., Comment: 21 pages, 6 figures, 10 tables
- Published
- 2021
47. Overview of the CLEF--2021 CheckThat! Lab on Detecting Check-Worthy Claims, Previously Fact-Checked Claims, and Fake News
- Author
-
Nakov, Preslav, Martino, Giovanni Da San, Elsayed, Tamer, Barrón-Cedeño, Alberto, Míguez, Rubén, Shaar, Shaden, Alam, Firoj, Haouari, Fatima, Hasanain, Maram, Mansour, Watheq, Hamdan, Bayan, Ali, Zien Sheikh, Babulkov, Nikolay, Nikolov, Alex, Shahi, Gautam Kishore, Struß, Julia Maria, Mandl, Thomas, Kutlu, Mucahid, and Kartal, Yavuz Selim
- Subjects
Computer Science - Computation and Language ,Computer Science - Information Retrieval ,Computer Science - Machine Learning ,Computer Science - Social and Information Networks ,68T50 ,F.2.2 ,I.2.7 - Abstract
We describe the fourth edition of the CheckThat! Lab, part of the 2021 Conference and Labs of the Evaluation Forum (CLEF). The lab evaluates technology supporting tasks related to factuality, and covers Arabic, Bulgarian, English, Spanish, and Turkish. Task 1 asks to predict which posts in a Twitter stream are worth fact-checking, focusing on COVID-19 and politics (in all five languages). Task 2 asks to determine whether a claim in a tweet can be verified using a set of previously fact-checked claims (in Arabic and English). Task 3 asks to predict the veracity of a news article and its topical domain (in English). The evaluation is based on mean average precision or precision at rank k for the ranking tasks, and macro-F1 for the classification tasks. This was the most popular CLEF-2021 lab in terms of team registrations: 132 teams. Nearly one-third of them participated: 15, 5, and 25 teams submitted official runs for tasks 1, 2, and 3, respectively., Comment: Check-Worthiness Estimation, Fact-Checking, Veracity, Evidence-based Verification, Detecting Previously Fact-Checked Claims, Social Media Verification, Computational Journalism, COVID-19
- Published
- 2021
48. Findings of the NLP4IF-2021 Shared Tasks on Fighting the COVID-19 Infodemic and Censorship Detection
- Author
-
Shaar, Shaden, Alam, Firoj, Martino, Giovanni Da San, Nikolov, Alex, Zaghouani, Wajdi, Nakov, Preslav, and Feldman, Anna
- Subjects
Computer Science - Computation and Language ,Computer Science - Information Retrieval ,Computer Science - Machine Learning ,Computer Science - Social and Information Networks ,68T50 ,F.2.2 ,I.2.7 - Abstract
We present the results and the main findings of the NLP4IF-2021 shared tasks. Task 1 focused on fighting the COVID-19 infodemic in social media, and it was offered in Arabic, Bulgarian, and English. Given a tweet, it asked to predict whether that tweet contains a verifiable claim, and if so, whether it is likely to be false, is of general interest, is likely to be harmful, and is worthy of manual fact-checking; also, whether it is harmful to society, and whether it requires the attention of policy makers. Task~2 focused on censorship detection, and was offered in Chinese. A total of ten teams submitted systems for task 1, and one team participated in task 2; nine teams also submitted a system description paper. Here, we present the tasks, analyze the results, and discuss the system submissions and the methods they used. Most submissions achieved sizable improvements over several baselines, and the best systems used pre-trained Transformers and ensembles. The data, the scorers and the leaderboards for the tasks are available at http://gitlab.com/NLP4IF/nlp4if-2021., Comment: COVID-19, infodemic, harmfulness, check-worthiness, censorship, social media, tweets, Arabic, Bulgarian, English, Chinese
- Published
- 2021
49. A Second Pandemic? Analysis of Fake News About COVID-19 Vaccines in Qatar
- Author
-
Nakov, Preslav, Alam, Firoj, Shaar, Shaden, Martino, Giovanni Da San, and Zhang, Yifan
- Subjects
Computer Science - Computation and Language ,Computer Science - Social and Information Networks ,68T50 ,F.2.2 ,I.2.7 - Abstract
While COVID-19 vaccines are finally becoming widely available, a second pandemic that revolves around the circulation of anti-vaxxer fake news may hinder efforts to recover from the first one. With this in mind, we performed an extensive analysis of Arabic and English tweets about COVID-19 vaccines, with focus on messages originating from Qatar. We found that Arabic tweets contain a lot of false information and rumors, while English tweets are mostly factual. However, English tweets are much more propagandistic than Arabic ones. In terms of propaganda techniques, about half of the Arabic tweets express doubt, and 1/5 use loaded language, while English tweets are abundant in loaded language, exaggeration, fear, name-calling, doubt, and flag-waving. Finally, in terms of framing, Arabic tweets adopt a health and safety perspective, while in English economic concerns dominate., Comment: COVID-19, disinformation, misinformation, factuality, fact-checking, fact-checkers, check-worthiness, framing, harmfulness, social media platforms, social media
- Published
- 2021
50. Assisting the Human Fact-Checkers: Detecting All Previously Fact-Checked Claims in a Document
- Author
-
Shaar, Shaden, Georgiev, Nikola, Alam, Firoj, Martino, Giovanni Da San, Mohamed, Aisha, and Nakov, Preslav
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Computers and Society ,Computer Science - Information Retrieval ,Computer Science - Machine Learning ,68T50 ,F.2.2 ,I.2.7 - Abstract
Given the recent proliferation of false claims online, there has been a lot of manual fact-checking effort. As this is very time-consuming, human fact-checkers can benefit from tools that can support them and make them more efficient. Here, we focus on building a system that could provide such support. Given an input document, it aims to detect all sentences that contain a claim that can be verified by some previously fact-checked claims (from a given database). The output is a re-ranked list of the document sentences, so that those that can be verified are ranked as high as possible, together with corresponding evidence. Unlike previous work, which has looked into claim retrieval, here we take a document-level perspective. We create a new manually annotated dataset for this task, and we propose suitable evaluation measures. We further experiment with a learning-to-rank approach, achieving sizable performance gains over several strong baselines. Our analysis demonstrates the importance of modeling text similarity and stance, while also taking into account the veracity of the retrieved previously fact-checked claims. We believe that this research would be of interest to fact-checkers, journalists, media, and regulatory authorities., Comment: detecting previously fact-checked claims, fact-checking, disinformation, fake news, social media, political debates
- Published
- 2021
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.