2,618 results on '"Natural language processing (Computer science)"'
Search Results
2. Machine learning for medical coding in healthcare surveys.
- Subjects
Health surveys -- United States. ,Machine learning. ,Diagnosis related groups -- United States. ,Natural language processing (Computer science) - Published
- 2021
3. The word-based pyramid
- Author
-
Thompson, Andrew A.
- Subjects
Computational linguistics. ,Natural language processing (Computer science) ,Machine translating. - Published
- 2012
4. Significance test in speaker recognition data analysis with data dependency
- Subjects
Bootstrap (Statistics) ,Natural language processing (Computer science) - Published
- 2012
5. Longt5Rank: A novel integrated hybrid approach for text summarisation
- Author
-
Lau, Agatha Jin Jin and Tan, Chi Wee
- Published
- 2024
6. Framing the future: The 'Foundation' series, foundation models and framing AI
- Author
-
Williams, Clare
- Published
- 2022
7. A rule-based machine learning model for career selection through MBTI personality
- Author
-
Fatima, Noureen, Gul, Sana, Ahmed, Javed, Khand, Zahid Hussain, and Mujtaba, Ghulam
- Published
- 2022
8. Machine learning and data science blueprints for finance : from building trading strategies to robo-advisors using Python.
- Author
-
Tatsat, Hariom, Lookabaugh, Brad, and Puri, Sahil
- Subjects
Finance -- Data processing ,Finance -- Mathematical models ,Machine learning ,Natural language processing (Computer science) ,Python (Computer program language) - Abstract
Summary: Machine learning and data science will significantly transform the finance industry in the next few years. With this practical guide, professionals at hedge funds, investment and retail banks, and fintech firms will learn how to build ML algorithms crucial to this industry. You'll examine ML concepts and over 20 case studies in supervised, unsupervised, and reinforcement learning, along with natural language processing (NLP).
- Published
- 2020
9. Resume classification system using natural language processing and machine learning techniques
- Author
-
Ali, Irfan, Mughal, Nimra, Khand, Zahid Hussain, Ahmed, Javed, and Mujtaba, Ghulam
- Published
- 2022
10. Natural language processing for detecting brand hate speech
- Author
-
Mednini, Latifa, Noubigh, Zouhaira, and Turki, Mouna Damak
- Published
- 2024
11. Artificial intelligence analysis: What the rise of AI means for human intelligence analysts
- Author
-
Turner, Brett
- Published
- 2024
12. Computational analysis and understanding of natural languages : principles, methods and applications.
- Author
-
Gudivada, Venkat N. and Rao, C.R.
- Subjects
Computational linguistics ,Natural language processing (Computer science) ,Semantics -- Data processing - Abstract
Summary: Computational Analysis and Understanding of Natural Languages: Principles, Methods and Applications, Volume 38, the latest release in this monograph that provides a cohesive and integrated exposition of these advances and associated applications, includes new chapters on Linguistics: Core Concepts and Principles, Grammars, Open-Source Libraries, Application Frameworks, Workflow Systems, Mathematical Essentials, Probability, Inference and Prediction Methods, Random Processes, Bayesian Methods, Machine Learning,
- Published
- 2018
13. Deep learning for natural language processing : creating neural networks with Python.
- Author
-
Goyal, Palash, Jain, Karan, and Pandey, Sumit
- Subjects
Natural language processing (Computer science) ,Neural networks (Computer science) ,Python (Computer program language) - Abstract
Summary: Discover the concepts of deep learning used for natural language processing (NLP), with full-fledged examples of neural network models such as recurrent neural networks, long short-term memory networks, and sequence-2-sequence models. You'll start by covering the mathematical prerequisites and the fundamentals of deep learning and NLP with practical examples. The first three chapters of the book cover the basics of NLP, starting with word-vector representation before moving onto advanced algorithms. The final chapters focus entirely on implementation, and deal with sophisticated architectures such as RNN, LSTM, and Seq2seq, using Python tools: TensorFlow, and Keras. Deep Learning for Natural Language Processing follows a progressive approach and combines all the knowledge you have gained to build a question-answer chatbot system. This book is a good starting point for people who want to get started in deep learning for NLP. All the code presented in the book will be available in the form of IPython notebooks and scripts, which allow you to try out the examples and extend them in interesting ways. You will: Gain the fundamentals of deep learning and its mathematical prerequisites Discover deep learning frameworks in Python Develop a chatbot Implement a research paper on sentiment classification.
- Published
- 2018
14. Issues in the multilingual information processing of spoken political and journalistic texts.
- Author
-
Alexandris, Christina
- Subjects
Journalism -- Translating ,Machine translating ,Natural language processing (Computer science) - Abstract
Summary: From television screens to mobile phones, spoken political and journalistic texts in the media are accessible to recipients of almost any kind, including the international public. These texts constitute a remarkable source of empirical data for human behavior and for linguistic phenomena, but pose significant challenges in terms of their evaluation, processing and translation due to a set of distinctive characteristics. This volume presents and describes a number of features of spoken political and journalistic texts, and proposes strategies for their correct and efficient analysis and processing both by human evaluators and by Natural Language Processing applications. The book also discusses the accessibility of "complex" information content and transfer for an international audience.
- Published
- 2016
15. Speaker Dependent Voice Recognition with Word-Tense Association and Part-of-Speech Tagging
- Author
-
Ernst, Theodore Philip King, Ernst, Theodore Philip King, Ernst, Theodore Philip King, and Ernst, Theodore Philip King
- Subjects
- Automatic speech recognition., Natural language processing (Computer science), Reconnaissance automatique de la parole., Traitement automatique des langues naturelles., Automatic speech recognition, Natural language processing (Computer science)
- Abstract
This thesis deals with speaker recognition and natural language processing. The most common speaker recognition systems are Text-Dependent and identify the speaker after a key word/phrase is uttered. This thesis presents Text-Independent Speaker recognition systems that incorporate the collaborative effort and research of noise-filtering, Speech Segmentation, Feature extraction, speaker verification and finally, Partial Language Modelling. The filtering process was accomplished using 4th order Butterworth Band-pass filters to dampen ambient noise outside normal speech frequencies of 300Hzto3000Hz. Speech segmentation utilizes Hamming windows to segment the speech, after which speech detection occurs by calculating the Short time Energy and Zero-crossing rates over a particular time period and identifying voices from unvoiced using a threshold. Audio data collected from different people is run consecutively through a Speaker Training and Recognition Algorithm which uses neutral networks to create a training group and target group for the recognition process. The output of the segmentation module is then processed by the neutral network to recognize the speaker. Though not implemented here due to database and computational requirements, the last module suggests a new model for the Part of Speech tagging process that involves a combination of Artificial Neural Networks (ANN) and Hidden Markov Models (HMM) in a series configuration to achieve higher accuracy. This differs from existing research by diverging from the usual single model approach of the creation of hybrid ANN and HMM models.
- Published
- 2024
16. Natural language annotation for machine learning.
- Author
-
Pustejovsky, James and Stubbs, Amber
- Subjects
Corpora (Linguistics) -- Data processing ,Machine learning ,Natural language processing (Computer science) - Abstract
Summary: Create your own natural language trainingcorpus for machine learning. Whether you're working with English, Chinese, orany other natural language, this hands-on book guides you through a provenannotation development cycle-the process of adding metadata to your trainingcorpus to help ML algorithms work more efficiently. You don't need anyprogramming or linguistics experience to get started. Using detailed examples at every step, you'll learn how the MATTER AnnotationDevelopment Process helps you Model, Annotate, Train, Test, Evaluate, and Reviseyour training corpus. You also get a complete walkthrough of a real-worldannotation project.
- Published
- 2013
17. Mobile speech and advanced natural language solutions.
- Author
-
Markowitz, Judith A. and Neustein, Amy
- Subjects
Automatic speech recognition ,Mobile computing ,Natural language processing (Computer science) - Abstract
Summary: Machine Talk: The Next Generation of Natural Language Processing and Speech Technology' presents the discussion of the most recent advances in intelligent human-computer interaction, including fascinating new study findings on talk-in-interaction, which is the province of conversation analysis, a subfield in sociology/sociolinguistics, a new and emerging area in natural language understanding. Editors Amy Neustein and Judith A. Markowitz have recruited a talented group of contributors to introduce the next generation natural language technologies for practical speech processing applications that serve the consumer's need for well-functioning natural language-driven personal assistants and other mobile devices, while also addressing business' need for better functioning IVR-driven call centers that yield a more satisfying experience for the caller. -- Source other than Library of Congress.
- Published
- 2013
18. Categorical tools for natural language processing
- Author
-
de Felice, Giovanni and Coecke, Bob
- Subjects
Categories (Mathematics) ,Linguistics ,Natural language processing (Computer science) ,Quantum theory - Abstract
This thesis develops the translation between category theory and computational linguistics as a foundation for natural language processing. The three chapters deal with syntax, semantics and pragmatics. First, string diagrams provide a unified model of syntactic structures in formal grammars. Second, functors compute semantics by turning diagrams into logical, tensor, neural or quantum computation. Third, the resulting functorial models can be composed to form games where equilibria are the solutions of language processing tasks. This framework is implemented as part of DisCoPy, the Python library for computing with string diagrams. We give a hierarchy of categorical and linguistic notions and an overview of their applications in compositional natural language processing.
- Published
- 2022
19. Category theory for quantum natural language processing
- Author
-
Toumi, Alexis Naïm Hubert, Coecke, Bob, and Marsden, Daniel
- Subjects
Natural language processing (Computer science) ,Quantum computing ,Categories (Mathematics) - Abstract
This thesis introduces quantum natural language processing (QNLP) models based on a simple yet powerful analogy between computational linguistics and quantum mechanics: grammar as entanglement. The grammatical structure of text and sentences connects the meaning of words in the same way that entanglement structure connects the states of quantum systems. Category theory allows to make this language-to-qubit analogy formal: it is a monoidal functor from grammar to vector spaces. We turn this abstract analogy into a concrete algorithm that translates the grammatical structure onto the architecture of parameterised quantum circuits. We then use a hybrid classical-quantum algorithm to train the model so that evaluating the circuits computes the meaning of sentences in data-driven tasks. The implementation of QNLP models motivated the development of DisCoPy (Distributional Compositional Python), the toolkit for applied category theory of which the first chapter gives a comprehensive overview. String diagrams are the core data structure of DisCoPy, they allow to reason about computation at a high level of abstraction. We show how they can encode both grammatical structures and quantum circuits, but also logical formulae, neural networks or arbitrary Python code. Monoidal functors allow to translate these abstract diagrams into concrete computation, interfacing with optimised task-specific libraries. The second chapter uses DisCopy to implement QNLP models as parameterised functors from grammar to quantum circuits. It gives a first proof-of-concept for the more general concept of functorial learning: generalising machine learning from functions to functors by learning from diagram-like data. In order to learn optimal functor parameters via gradient descent, we introduce the notion of diagrammatic differentiation: a graphical calculus for computing the gradients of parameterised diagrams.
- Published
- 2022
20. Coarse-to-fine natural language processing.
- Author
-
Petrov, Slav Orlinov
- Subjects
COMPUTERS -- General ,Sciences sociales ,Natural language processing (Computer science) - Abstract
Summary: The impact of computer systems that can understand natural language will be tremendous. To develop this capability we need to be able to automatically and efficiently analyze large amounts of text. Manually devised rules are not sufficient to provide coverage to handle the complex structure of natural language, necessitating systems that can automatically learn from examples. To handle the flexibility of natural language, it has become standard practice to use statistical models, which assign probabilities for example to the different meanings of a word or the plausibility of grammatical constructions. This book develops a general coarse-to-fine framework for learning and inference in large statistical models for natural language processing. Coarse-to-fine approaches exploit a sequence of models which introduce complexity gradually. At the top of the sequence is a trivial model in which learning and inference are both cheap. Each subsequent model refines the previous one, until a final, full-complexity model is reached. Applications of this framework to syntactic parsing, speech recognition and machine translation are presented, demonstrating the effectiveness of the approach in terms of accuracy and speed. This book is intended for students and researchers interested in statistical approaches to Natural Language Processing.¡ Slav's work¡Coarse-to-Fine Natural Language Processing represents a major advance in the area of syntactic parsing, and a great advertisement for the superiority of the machine-learning approach. Eugene Charniak (Brown University).
- Published
- 2012
21. Grammars for language and genes : theoretical and empirical investigations.
- Author
-
Chiang, David and Joshi, Aravind K.
- Subjects
Computational linguistics ,Grammar, Comparative and general ,Natural language processing (Computer science) ,Sequence alignment (Bioinformatics) - Published
- 2012
22. Handbook of natural language processing and machine translation : DARPA global autonomous language exploitation.
- Author
-
Christianson, Caitlin, McCary, John, and Olive, Joseph P.
- Subjects
Machine translating ,Natural language processing (Computer science) - Abstract
Summary: Data Acquisition and Linguistic Resources -- Machine Translation from Text -- Machine Translation From Speech -- Distillation -- Machine Translation Evaluation and Optimization -- Operational Engines.
- Published
- 2011
23. Graph-based natural language processing and information retrieval.
- Author
-
Mihalcea, Rada and Radev, Dragomir
- Subjects
Graph based information retrieval ,Graphical user interfaces (Computer systems) ,Natural language processing (Computer science) - Abstract
Summary: "Graph theory and the fields of natural language processing and information retrieval are well-studied disciplines. Traditionally, these areas have been perceived as distinct, with different algorithms, different applications, and different potential end-users. However, recent research has shown that these disciplines are intimately connected, with a large variety of natural language processing and information retrieval applications finding efficient solutions within graph-theoretical frameworks. This book extensively covers the use of graph-based algorithms for natural language processing and information retrieval. It brings together topics as diverse as lexical semantics, text summarization, text mining, ontology construction, text classification, and information retrieval, which are connected by the common underlying theme of the use of graph-theoretical methods for text and information processing tasks. Readers will come away with a firm understanding of the major methods and applications in natural language processing and information retrieval that rely on graph-based representations and algorithms"-- Provided by publisher.
- Published
- 2011
24. The spectrums of automated privacy policy compliance
- Author
-
Burdon, Mark, Hardy, Sam, Klaver, Rianne, Yang, Nick, Gaunt, Sam, and Casey, Samuel Irvine
- Published
- 2023
25. Phishing message detection based on keyword matching
- Author
-
Tham, Keng-Theen, Ng, Kok-Why, and Haw, Su-Cheng
- Published
- 2023
26. Essays in macroeconomics and machine learning
- Author
-
Ashwin, Julian and Ellison, Martin
- Subjects
Economics ,Machine learning ,Macroeconomics ,Natural language processing (Computer science) - Abstract
The unifying theme of this thesis is the use of techniques from machine learning and data science to address questions in macroeconomics. It makes both theoretical contributions by applying neural networks as a learning algorithm in models that are indeterminate under rational expectations, and empirical contributions by developing and applying natural language processing methods to datasets including news media articles and central bank communication. The first Chapter, 'Resolving Indeterminacy with Neural Network Learning: Sinks become Sources', aims to make a theoretical contribution to the literature on indeterminacy in rational expectations models. Indeterminacy (i.e. non-uniqueness of equilibrium under rational expectations) is a pervasive and often neglected challenge for macroeconomists. Previous literature on learning in macroeconomics has shown that the equilibria in linear indeterminate models are almost always not learnable. This Chapter examines the equilibrium sophisticated learning agents converge to in models where there is indeterminacy, but it is bounded. This neural network learning converges to a stable equilibrium, in which indeterminate regions are sources and determinate regions are sinks. Furthermore, this equilibrium is consistent with Rational Expectations. There are multiple steady states and agents have correct beliefs about transitions between them, which means that transitory shocks can have permanent effects. This is demonstrated with an application to a well-known example of indeterminacy: a New Keynesian model with a Zero Lower Bound in interest rates. The resolution to the challenge of indeterminacy presented here is plausible, as it's based on learnability, and also appealing because the identified equilibrium is globally stable, can have multiple steady states with well-defined transitions between them, and passes tests for rationality. It also demonstrates the value of using machine learning, as the neural network acts as a very flexible function approximator that is fast to train, allowing the use of learnability as an equilibrium selection device in non-linear models. Alternative learning algorithms like Recursive Least Squares yield qualitatively similar results, but are not sufficiently flexible to pass tests of rationality. Chapter 2, 'Bayesian Topic Regression for Causal Inference' has a more method- ological focus, developing Bayesian Topic Regression, a model for causal inference with text data. This methodology is then applied in Chapter 3 to help identify a potentially causal effect of media coverage on stock price volatility. The Bayesian Topic Regression model jointly estimates topics in text documents and a regression using these topics and associated numerical data to predict a response variable. As well as showing that per- forming text feature extraction and prediction in separate stages can lead to incorrect inference, we benchmark our model on two real-world customer review datasets and show markedly improved out-of-sample prediction in comparison to competing approaches. Chapters 3 and 4 use text data to address to empirical questions relating to how agents in the economy acquire information and on what they focus their attention. In Chapter 3 'Financial news media and volatility: is there more to newspapers that news?' identifies a co-movement media coverage in the Financial Times newspaper and a firm's intra-day stock price volatility is identified. I argue that part of this co-movement is causal, relying on an identification strategy based on the publication time of the newspa- per, controlling for persistence and anticipation effects as well as the content of articles using the Bayesian Topic Regression introduced in Chapter 2. These results are consis- tent with a salience-based view of the media's role in financial markets: media coverage does not (only) provide information, but influences where investors choose to direct their focus. This identified effect also has interesting spillovers to firms in sectors that are linked by the production network. In Chapter 4 'The Shifting focus of Central Bankers'. I use an unsupervised topic model to quantify the focus of central bank communication and economic news media. I offer an explanation for the variation of this focus over time, and identify a robust co-movement between central bank and media focus. A model of multidimensional uncertainty and limited attention is proposed to explain the shifting focus of central bank communication. Evidence from the Survey of Professional Forecasters is used to support this explanation, showing that focus shifts to cover variables about which there is greater uncertainty. An event study approach is used to show a potentially causal influence of Federal Reserve communication on the focus of US news media and on the communication of other central banks.
- Published
- 2021
27. Generating textual captions for ultrasound visuals in an automated fashion
- Author
-
Alsharid, Mohammad and Noble, J. Alison
- Subjects
Natural language processing (Computer science) ,Computer vision ,Imaging ,Fetus--Ultrasonic imaging - Abstract
Generating captions for ultrasound images and videos is an area that is yet to be fully studied and explored. The aim of the work in this thesis is to learn joint image-text representations to describe ultrasound images with rich vocabulary consisting of nouns, verbs, and adjectives. Preparing medical image captioning benchmarks is challenging for two reasons: (a) describing medical images with specific terminology requires expert knowledge of medical professionals; and (b) the sensitive nature of medical images prevents wide-scale annotation, for instance, using crowd-sourcing services (e.g. Amazon Mechanical Turk) and similar methods. Therefore, automatic image captioning has not been widely studied on ultrasound images before, the challenge being enhanced by the lack of readily available large datasets of ultrasound images with captions. First, the thesis explores different combinations of recurrent neural networks, concatenation techniques, word embedding vectors in different model architecture configurations. We identify in this process the configuration most suitable for the fetal ultrasound image captioning task and the dataset at hand. We show that a configuration incorporating an LSTM-RNN and word2vec embeddings and using a merge-by-concatenation operation performed best. The thesis then explores three solutions to the challenge of working with real world datasets. We introduce a curriculum learning based strategy that incorporates the complexities of the image and text information to prepare the data for training. We show that by training captioning models with the order of data samples determined by the curriculum, we can achieve higher scores on the evaluation metrics with the same amount of data. We also look into augmenting the data through the creation of pseudocaptions to pair up with caption-less images. Finally, we explore leveraging other available data from a different modality, specifically eye gaze points, to supplement available image-text data. We find that using eye gaze data can help in training models that score relatively higher on the evaluation metrics; however since the improvements are small and the pre-training steps involved are considerable, this leads us to the recommendation that improving base models should take precedence over relying on data from other modalities to improve the performance of captioning models. To the best of our knowledge, the work in this thesis is the first attempt to perform automatic image captioning on fetal ultrasound images (video frames), using sonographer spoken words to describe their scanning experience. The thesis can help serve as a blue print for future endeavours in fetal ultrasound captioning by providing guidelines to follow and pitfalls to avoid and as an aid for those attempting medical image captioning, more generally.
- Published
- 2021
28. Representation learning of linguistic structures with neural networks
- Author
-
Kawakami, Kazuya and Blunsom, Philip
- Subjects
Natural language processing (Computer science) - Abstract
In the first part of the thesis, I explore the explicit modelling of word creation and reuse in the context of open-vocabulary language modelling. I propose a neural network augmented with a hierarchical structure and a memory component that explicitly models the generation of new words and supports the frequent reuse of new words. The research question is whether the explicit modelling assumption is useful for improving the performance of language modelling compared to the implicit model without using any linguistic structures. The model is evaluated in terms of language modelling performance (i.e. held-out perplexity) on typologically diverse languages and compared with a character-level neural language model which does not explicitly represent any linguistic structure. The results show that the proposed explicit model improve the performance on language modelling in all tested languages and analysis demonstrates that the model is able to use the memory architecture appropriately. In the second part, I extend the open-vocabulary language model to discover word-like structures without any supervision of word boundaries. In contrast to previous work on word segmentation and language modelling that focuses only on either structure discovery or language modelling, the hypothesis is that it is possible to learn good predictive distributions of language at the same time as discovering good explicit structures. Thus, the proposed model combines the benefit of explicit and implicit modelling by parameterizing an explicit probabilistic model using neural networks. The proposal includes a differential learning algorithm that efficiently marginalizes all possible segmentation decisions and a regularization method that is crucial for successful structure induction and language modelling. The model is evaluated in terms of both language modelling performance (i.e. held-out perplexity) and the quality of induced word structures (i.e. precision metrics compared to the human reference). The results show that the proposed model improves language modelling performance over neural language models and discovers word-like units better than Bayesian word segmentation models. Moreover, conditioning on visual context improves performance on both. In the last part, I present a method to discover acoustic structures implicitly from raw audio signals and show that the model can learn useful representations from largescale, real-world data. The aim is to learn representations that are robust to domain shifts (e.g. read English to spoken English) and generalize well to many languages. Since the structures are not induced explicitly, the representations are evaluated based on the impact on downstream speech recognition tasks which predicts phonetic structure in utterances. The results show that the representations learned from diverse and noisy data provide significant improvements on speech recognition tasks in terms of performance, data efficiency and robustness. Moreover, the representations generalize well to many languages including tonal and low-resource languages.
- Published
- 2021
29. Justification Mining : developing a novel machine learning method for identifying representative sentences and summarising sentiment in financial text
- Author
-
Patel, Kevin, Coecke, Bob, and Simpson, Edwin
- Subjects
Transfer learning (Machine learning) ,Quantitative analysts ,Computational finance ,Deep learning (Machine learning) ,Machine learning ,Quantitative research ,Trading rooms (Finance) ,Ensemble learning (Machine learning) ,Support vector machines ,Sentiment analysis ,Financial engineering ,Algorithmic trading ,Natural language processing (Computer science) ,Finance ,Supervised learning (Machine learning) - Abstract
As the fields of machine learning, natural language processing, and big data become increasingly important throughout various industries, including finance, it becomes crucial to evaluate how they are being utilised and understand the motivations behind the recommendations of algorithms, in order to make sound decisions based on this information. In finance, particularly, being able to create concise and easy to comprehend tools for understanding machine learning models is extremely beneficial. These tools could serve multiple purposes, including allowing managers to better explain to and convince investors of machine learning based trading strategies. Decision-makers themselves might also be able to act according to the information provided by these tools, which could assist in, for example, making more ethically informed decisions. Thus, this type of research not only has implications in the areas of machine learning, Natural Language Processing (NLP), and finance but also in fields like AI explainability, model interpretability and Responsible Research and Innovation. The central problem this thesis addresses is whether machine learning methods can be used to mine representative sentences from financial text that are able to capture the majority of sentiment in a full document with only a few sentences. These mined sentences are referred to as 'justifications', and the process has been called 'justification mining'. The full documents used here are data taken from 10-K filings. Before examining justification mining, however, transfer learning methods suitable for training data annotated at a different level than testing data (sentence- versus document-level) are assessed. The purpose of this is to address a problem that often occurs in NLP research, where no annotated training data perfectly suited to the research is readily available. Therefore, methods must be created to make use of what is available, in this case sentence-annotated training data being employed to train classifiers for prediction on document-level testing data. These transfer learning methods are then employed in the next step, which focuses on justification mining. The process of justification mining itself is first developed using transformer models to encode embeddings and, then, using clustering algorithms and cosine similarity to extract or mine justifications. This process is evaluated by comparing sentiment from mined justifications to sentiment from full documents, in order to assess the ability of justification mining to capture or summarise sentiment. It is also evaluated by correlating predicted sentiment from mined justifications and full documents to future stock returns, to gauge whether justification mining offers any benefit in identifying signals in the data for financial purposes. Little work has been done previously that evaluates the results of financial sentiment analyses in this way. In the NLP domain, the best way of extracting aggregate justifications for sentiment is still an open question. Moreover, few research papers attempt to apply transfer learning from lower-level data to entire documents. Nor are 10-K filings widely studied in this context, as they are difficult to parse. The methods created in this thesis might offer a novel means of providing information that can assess the motivations behind sentiment analyses. In the transfer learning process, feature engineering and preprocessing steps were modified to obtain accuracies up to 0.903. For justification mining, considering statistically significant (p≤0.05) correlations of |r| > 0.2, sentiment from mined justifications more often correlated with future stock returns than did full document sentiment. Although the correlations (r) were somewhat weak overall, they could potentially be combined with traditional alpha signals to enhance these signals (see Section 11 for further discussion). Moreover, high degrees of similarity were found between aggregated sentiment from mined justifications and full documents, with similarity scores up to 0.9999, supporting the efficacy of justification mining in capturing full document sentiment. In these evaluations, transformer models performed better in numericising text for input into ML models than traditional approaches like Bag of Words. In fact, every statistically significant sentiment to stock return correlation bar one, as well as every mined justification to full document similarity and correlation score above 0.7, used transformer models for numericisation. These results imply that justification mining might be successful in eliminating sentiment noise in financial data, as well as in capturing the majority of sentiment from a full document. Moreover, transformer models might provide an advantage over traditional approaches like Bag of Words for numericisation of text. Mined justifications, themselves, provide an easily interpretable and presentable means of explaining the output of sentiment analysis and have numerous uses, including identifying the driving factors behind sentiment in a financial document, which could be helpful for making more ethically informed decisions or building investor trust in the methodology of sentiment analysis algorithms.
- Published
- 2021
30. Machine learning for multimodal interaction. [electronic resource] : third international workshop, MLMI 2006, Bethesda, MD, USA, May 1-4, 2006 : revised selected papers.
- Author
-
Bengio, Samy, Fiscus, Jonathan G., Renals, Steve, and National Institute of Standards and Technology (U.S.)
- Subjects
Automatic speech recognition ,Human-computer interaction ,Machine learning ,Natural language processing (Computer science) ,Speech processing systems - Abstract
Summary: This book constitutes the thoroughly refereed post-proceedings of the Third International Workshop on Machine Learning for Multimodal Interaction, MLMI 2006, held in Bethesda, MD, USA, in May 2006. The papers are organized in topical sections on multimodal processing, image and video processing, HCI and applications, discourse and dialogue, speech and audio processing, and NIST meeting recognition evaluation.
- Published
- 2006
31. Computational metaphor processing
- Author
-
Mao, Rui, Lin, Chenghua, Guerin, Frank, Reiter, Ehud, and Sripada, Gowri
- Subjects
006.35 ,Computational linguistics ,Metaphor ,Figures of speech ,Natural language processing (Computer science) - Published
- 2020
32. Novel symbolic and sub-symbolic approaches for text based and multimodal sentiment analysis
- Author
-
Dashtipour, Kia and Li, Jingpeng
- Subjects
006.3 ,Sentiment analysis ,ML ,Text processing (Computer science) ,Natural language processing (Computer science) ,Computer vision ,Multimedia systems - Abstract
In the era of digital media, e-commerce and social networks, websites allow users to share opinions and feedback about products and services. Customers can make informed decisions by reading the experiences of other users. In addition, customer feedback can be used by the organizations to further improve the offered services. However, the quintillion bytes of data generated per day in different languages such as Persian consisting of user feedback cannot be manually read and analyzed by an individual or an organization, for gauging public opinion. Sentiment analysis is an automated process of computationally understanding and classifying subjective information in multi disciplinary fields such as products, movies, news, public opinion etc. In this thesis, we focus on developing novel methods for Persian text based sentiment analysis. We exploit the developed text-based methods to improve multimodal polarity detection. Specifically, we develop a novel hybrid framework that integrates dependency-based rules and deep neural networks for detecting polarity in Persian natural language sentences. In addition, we develop a Persian multimodal sentiment analysis framework that integrates audio, visual and textual cues to computationally understand and harvest sentiments from videos posted on social media platforms such as YouTube and Facebook. Specifically, a first of its kind, multimodal Persian sentiment analysis dataset is developed, which is then used evaluate the proposed multimodal framework that exploits the hybrid dependency-based sentiment analysis framework and deep neural network based multimodal fusion. Extensive experimental results have proven the effectiveness of the proposed approaches as compared to state-of-the-art approaches including deep neural networks.
- Published
- 2019
33. Build a Large Language Model (From Scratch)
- Author
-
Sebastian Raschka and Sebastian Raschka
- Subjects
- Artificial intelligence, Natural language processing (Computer science)
- Abstract
Learn how to create, train, and tweak large language models (LLMs) by building one from the ground up!In Build a Large Language Model (from Scratch) bestselling author Sebastian Raschka guides you step by step through creating your own LLM. Each stage is explained with clear text, diagrams, and examples. You'll go from the initial design and creation, to pretraining on a general corpus, and on to fine-tuning for specific tasks. Build a Large Language Model (from Scratch) teaches you how to: • Plan and code all the parts of an LLM • Prepare a dataset suitable for LLM training • Fine-tune LLMs for text classification and with your own data • Use human feedback to ensure your LLM follows instructions • Load pretrained weights into an LLM Build a Large Language Model (from Scratch) takes you inside the AI black box to tinker with the internal systems that power generative AI. As you work through each key stage of LLM creation, you'll develop an in-depth understanding of how LLMs work, their limitations, and their customization methods. Your LLM can be developed on an ordinary laptop, and used as your own personal assistant. About the technology Physicist Richard P. Feynman reportedly said, “I don't understand anything I can't build.” Based on this same powerful principle, bestselling author Sebastian Raschka guides you step by step as you build a GPT-style LLM that you can run on your laptop. This is an engaging book that covers each stage of the process, from planning and coding to training and fine-tuning. About the book Build a Large Language Model (From Scratch) is a practical and eminently-satisfying hands-on journey into the foundations of generative AI. Without relying on any existing LLM libraries, you'll code a base model, evolve it into a text classifier, and ultimately create a chatbot that can follow your conversational instructions. And you'll really understand it because you built it yourself! What's inside • Plan and code an LLM comparable to GPT-2 • Load pretrained weights • Construct a complete training pipeline • Fine-tune your LLM for text classification • Develop LLMs that follow human instructions About the reader Readers need intermediate Python skills and some knowledge of machine learning. The LLM you create will run on any modern laptop and can optionally utilize GPUs. About the author Sebastian Raschka is a Staff Research Engineer at Lightning AI, where he works on LLM research and develops open-source software. The technical editor on this book was David Caswell. Table of Contents 1 Understanding large language models 2 Working with text data 3 Coding attention mechanisms 4 Implementing a GPT model from scratch to generate text 5 Pretraining on unlabeled data 6 Fine-tuning for classification 7 Fine-tuning to follow instructions A Introduction to PyTorch B References and further reading C Exercise solutions D Adding bells and whistles to the training loop E Parameter-efficient fine-tuning with LoRA
- Published
- 2025
34. The AI Revolution in Customer Service and Support : A Practical Guide to Impactful Deployment of AI to Best Serve Your Customers
- Author
-
Ross Smith, Mayte Cubino, Emily McKeon, Ross Smith, Mayte Cubino, and Emily McKeon
- Subjects
- Customer services--Data processing, Artificial intelligence, Natural language processing (Computer science)
- Abstract
In the rapidly evolving AI landscape, customer service and support professionals find themselves in a prime position to take advantage of this innovative technology to drive customer success. The AI Revolution in Customer Service and Support is a practical guide for professionals who want to harness the power of generative AI within their organizations to create more powerful customer and employee experiences. This book is designed to equip you with the knowledge and confidence to embrace the AI revolution and integrate the technology, such as large language models (LLMs), machine learning, predictive analytics, and gamified learning, into the customer experience. Start your journey toward leveraging this technology effectively to optimize organizational productivity. A portion of the book's proceeds will be donated to the nonprofit Future World Alliance, dedicated to K-12 AI ethics education. IN THIS BOOK YOU'LL LEARN About AI, machine learning, and data science How to develop an AI vision for your organization How and where to incorporate AI technology in your customer experience fl ow About new roles and responsibilities for your organization How to improve customer experience while optimizing productivity How to implement responsible AI practices How to strengthen your culture across all generations in the workplace How to address concerns and build strategies for reskilling and upskilling your people How to incorporate games, play, and other techniques to engage your agents with AI Explore thought experiments for the future of support in your organization “Insightful & comprehensive—if you run a service & support operation, put this book on your essential reading list right now!” —PHIL WOLFENDEN, Cisco, VP, Customer Experience “This book is both timely and relevant as we enter an unprecedented period in our industry and the broader world driven by Generative AI. The magnitude and speed of change we're experiencing is astounding and this book does an outstanding job balancing technical knowledge with the people and ethical considerations we must also keep front of mind.” —BRYAN BELMONT, Microsoft, Corporate VP, Customer Service & Support “The authors of this book are undoubtedly on the front lines of operationalizing Gen AI implementations in customer support environments… and they know undoubtedly that at its core, support is about people and genuine human connections. This book walks you through their journey to keep people at the center of this technical tsunami.” —PHAEDRA BOINODIRIS, Author, AI for the Rest of Us
- Published
- 2025
35. Generative AI in Writing Education : Policy and Pedagogical Implications
- Author
-
Dylan Medina and Dylan Medina
- Subjects
- Natural language processing (Computer science), Artificial intelligence--Educational applications, English language--Composition and exercises--Study and teaching, Rhetoric--Data processing, English language--Rhetoric--Study and teaching
- Abstract
This book provides a theoretical framework to allow educators, researchers, and policymakers to better understand computer‑generated writing and the policy and pedagogical implications of generative AI.Generative AI, such as ChatGPT and Gemini, has substantially disrupted educational spaces, forcing educators, policymakers, and other stakeholders to reconsider writing and how it should be used in education. Responding to this disruption, this book provides technically sound guidance on how various stakeholders should engage with generative AI. After providing a foundational and technical discussion of the technology, this book directly addresses the educational context. Informed by theories of learning and knowledge transfer and utilizing rhetorical theories of writing, this book assesses the impact of AI on student learning, student performance, and academic honesty and integrity. In doing so, the book outlines how generative AI can be both a help and a hindrance for students, enabling readers to craft informed and meaningful policies and successfully integrate AI in the composition classroom.This book will be of interest to scholars in the fields of Rhetoric and Composition, Technical Writing, Communication Studies, Linguistics, and TESOL, as well as to Education and Machine Learning policymakers, program directors, and researchers.
- Published
- 2025
36. Next Generation AI Language Models in Research : Promising Perspectives and Valid Concerns
- Author
-
Kashif Naseer Qureshi, Gwanggil Jeon, Kashif Naseer Qureshi, and Gwanggil Jeon
- Subjects
- Artificial intelligence--Scientific applications, Research--Data processing, Natural language processing (Computer science)
- Abstract
In this comprehensive and cutting-edge volume, Qureshi and Jeon bring together experts from around the world to explore the potential of artificial intelligence models in research and discuss the potential benefits and the concerns and challenges that the rapid development of this field has raised.The international chapter contributor group provides a wealth of technical information on different aspects of AI, including key aspects of AI, deep learning and machine learning models for AI, natural language processing and computer vision, reinforcement learning, ethics and responsibilities, security, practical implementation, and future directions. The contents are balanced in terms of theory, methodologies, and technical aspects, and contributors provide case studies to clearly illustrate the concepts and technical discussions throughout. Readers will gain valuable insights into how AI can revolutionize their work in fields including data analytics and pattern identification, healthcare research, social science research, and more, and improve their technical skills, problem-solving abilities, and evidence-based decision-making. Additionally, they will be cognizant of the limitations and challenges, the ethical implications, and security concerns related to language models, which will enable them to make more informed choices regarding their implementation.This book is an invaluable resource for undergraduate and graduate students who want to understand AI models, recent trends in the area, and technical and ethical aspects of AI. Companies involved in AI development or implementing AI in various fields will also benefit from the book's discussions on both the technical and ethical aspects of this rapidly growing field.
- Published
- 2025
37. Understanding Natural Language Understanding
- Author
-
Erik Cambria and Erik Cambria
- Subjects
- Natural language processing (Computer science)
- Abstract
About half a century ago, AI pioneers like Marvin Minsky embarked on the ambitious project of emulating how the human mind encodes and decodes meaning. While today we have a better understanding of the brain thanks to neuroscience, we are still far from unlocking the secrets of the mind, especially when it comes to language, the prime example of human intelligence. “Understanding natural language understanding”, i.e., understanding how the mind encodes and decodes meaning through language, is a significant milestone in our journey towards creating machines that genuinely comprehend human language. Large language models (LLMs) such as GPT-4 have astounded us with their ability to generate coherent, contextually relevant text, seemingly bridging the gap between human and machine communication. Yet, despite their impressive capabilities, these models operate on statistical patterns rather than true comprehension. This textbook delves into the nuanced differences between these two paradigms and explores the future of AI as we strive to achieve true natural language understanding (NLU). LLMs excel at identifying and replicating patterns within vast datasets, producing responses that appear intelligent and meaningful. They can generate text that mimics human writing styles, provide summaries of complex documents, and even engage in extended dialogues with users. However, their limitations become evident when they encounter tasks that require deeper understanding, reasoning, and contextual knowledge. An NLU system that deconstructs meaning leveraging linguistics and semiotics (on top of statistical analysis) represents a more profound level of language comprehension. It involves understanding context in a manner similar to human cognition, discerning subtle meanings, implications, and nuances that current LLMs might miss or misinterpret. NLU grasps the semantics behind words and sentences, comprehending synonyms, metaphors, idioms, and abstract concepts with precision. This textbook explores the current state of LLMs, their capabilities and limitations, and contrasts them with the aspirational goals of NLU. The author delves into the technical foundations required for achieving true NLU, including advanced knowledge representation, hybrid AI systems, and neurosymbolic integration, while also examining the ethical implications and societal impacts of developing AI systems that genuinely understand human language. Containing exercises, a final assignment and a comprehensive quiz, the textbook is meant as a reference for courses on information retrieval, AI, NLP, data analytics, data mining and more.
- Published
- 2025
38. The Pragmatics of Governmental Discourse : Resilience, Sustainability and Wellbeing
- Author
-
Ayan-Yue Gupta and Ayan-Yue Gupta
- Subjects
- Rhetoric--Political aspects--Great Britain--History--21st century, Discourse analysis--Social aspects, Discourse analysis--Political aspects, Discourse analysis--Data processing, Pragmatics, Natural language processing (Computer science)
- Abstract
This book presents a novel methodological framework for analysing governmental discourse. It involves combining pragmatist perspectives on language with computational sociolinguistics and large language models (LLMs).The first half discusses traditional critical approaches to investigating discursive practices, principally those employing Critical Discourse Analysis (CDA) and those based on methods developed by Michel Foucault. These are critiqued in terms of pragmatist views on meaning, which are rarely taken up in this area. It is argued that to understand the grounding of social structures and power relations in discourse, we must begin with a systematic account of how meaning is contextually fixed. It is proposed that a pragmatist reading of Foucault's arguments about governmentality offers a productive framework for discourse analysis. To illustrate the advantages of this framework, the book presents a case study of the British government's adoption of resilience, sustainability, and wellbeing discourses in the period 2000-2020. A dataset of 179 million tokens sampled from approximately 170,000 government documents is used to illustrate how this framework can be combined with natural language processing (NLP) to make robust inferences.This study will be of interest to both sociologists interested in language and in the methodological potential of recent developments in NLP. Importantly, the book demonstrates how LLMs can be harnessed to bring new perspectives to long-standing sociological questions.
- Published
- 2025
39. Large Language Models : Concepts, Techniques and Applications
- Author
-
John Atkinson-Abutridy and John Atkinson-Abutridy
- Subjects
- Machine translating, Natural language processing (Computer science), Artificial intelligence, Chatbots
- Abstract
This book serves as an introduction to the science and applications of Large Language Models (LLMs). You'll discover the common thread that drives some of the most revolutionary recent applications of artificial intelligence (AI): from conversational systems like ChatGPT or BARD, to machine translation, summary generation, question answering, and much more.At the heart of these innovative applications is a powerful and rapidly evolving discipline, natural language processing (NLP). For more than 60 years, research in this science has been focused on enabling machines to efficiently understand and generate human language. The secrets behind these technological advances lie in LLMs, whose power lies in their ability to capture complex patterns and learn contextual representations of language. How do these LLMs work? What are the available models and how are they evaluated? This book will help you answer these and many other questions. With a technical but accessible introduction: You will explore the fascinating world of LLMs, from its foundations to its most powerful applications You will learn how to build your own simple applications with some of the LLMs Designed to guide you step by step, with six chapters combining theory and practice, along with exercises in Python on the Colab platform, you will master the secrets of LLMs and their application in NLP.From deep neural networks and attention mechanisms, to the most relevant LLMs such as BERT, GPT-4, LLaMA, Palm-2 and Falcon, this book guides you through the most important achievements in NLP. Not only will you learn the benchmarks used to evaluate the capabilities of these models, but you will also gain the skill to create your own NLP applications. It will be of great value to professionals, researchers and students within AI, data science and beyond.
- Published
- 2025
40. AI chatbots not yet ready for clinical use
- Author
-
Joshua Au Yeung, Zeljko Kraljevic, Akish Luintel, Alfred Balston, Esther Idowu, Richard J. Dobson, and James T. Teo
- Subjects
large language models ,chatbot ,natural language processing (computer science) ,digital health ,AI safety ,transformer ,Medicine ,Public aspects of medicine ,RA1-1270 ,Electronic computers. Computer science ,QA75.5-76.95 - Abstract
As large language models (LLMs) expand and become more advanced, so do the natural language processing capabilities of conversational AI, or “chatbots”. OpenAI's recent release, ChatGPT, uses a transformer-based model to enable human-like text generation and question-answering on general domain knowledge, while a healthcare-specific Large Language Model (LLM) such as GatorTron has focused on the real-world healthcare domain knowledge. As LLMs advance to achieve near human-level performances on medical question and answering benchmarks, it is probable that Conversational AI will soon be developed for use in healthcare. In this article we discuss the potential and compare the performance of two different approaches to generative pretrained transformers—ChatGPT, the most widely used general conversational LLM, and Foresight, a GPT (generative pretrained transformer) based model focused on modelling patients and disorders. The comparison is conducted on the task of forecasting relevant diagnoses based on clinical vignettes. We also discuss important considerations and limitations of transformer-based chatbots for clinical use.
- Published
- 2023
- Full Text
- View/download PDF
41. Building Applications with Large Language Models : Techniques, Implementation, and Applications
- Author
-
Bhawna Singh and Bhawna Singh
- Subjects
- Artificial intelligence, Machine learning, Python (Computer program language), Natural language processing (Computer science)
- Abstract
This book delves into a broad spectrum of topics, covering the foundational aspects of Large Language Models (LLMs) such as PaLM, LLaMA, BERT, and GPT, among others. The book takes you through the complexities involved in creating and deploying applications based on LLMs, providing you with an in-depth understanding of the model architecture. You will explore techniques such as fine-tuning, prompt engineering, and retrieval augmented generation (RAG). The book also addresses different ways to evaluate LLM outputs and discusses the benefits and limitations of large models. The book focuses on the tools, techniques, and methods essential for developing Large Language Models. It includes hands-on examples and tips to guide you in building applications using the latest technology in Natural Language Processing (NLP). It presents a roadmap to assist you in navigating challenges related to constructing and deploying LLM-based applications. By the end of the book, you will understand LLMs and build applications with use cases that align with emerging business needs and address various problems in the realm of language processing. What You Will Learn Be able to answer the question: What are Large Language Models? Understand techniques such as prompt engineering, fine-tuning, RAG, and vector databases Know the best practices for effective implementation Know the metrics and frameworks essential for evaluating the performance of Large Language Models Who This Book Is For An essential resource for AI-ML developers and enthusiasts eager to acquire practical, hands-on experience in this domain; also applies to individuals seeking a technical understanding of Large Language Models (LLMs) and those aiming to build applications using LLMs
- Published
- 2024
42. Inevitable Knowledge
- Author
-
Janos J. Sarbo and Janos J. Sarbo
- Subjects
- Artificial intelligence, Expert systems (Computer science), Neural networks (Computer science), Natural language processing (Computer science), Multiagent systems
- Abstract
The Holy Grail of AI is artificial generative intelligence, a computer that can think human-like. However, human thinking is qualitatively more complex than computer calculations. So, the ultimate goal of AI cannot be achieved. Not quite. This book shows that a model of human-like, meaningful processing can be introduced based on a theory of cognition (how human processing can be abstracted in a series of events), semiotics (what signs are and what kind of distinctions can be communicated by signs), and computer science (how all this can be realized as a procedure). The emerging model offers a solution to the problem of artificial intelligence, not by itself, but in collaboration with the human agent by augmenting its intelligence. But there is more to it than that. Because of the fundamental nature of signs, the semiotic concept of meaning can be transformative for AI research. The book comprehensively covers several applications, including language processing, analyzing integrative negation processes, and solving mathematical problems. It delves into the intricate characteristics of the meaningful processing problem and the fascinating journey that led to its solution. The book provides insight into the historical background of the problem and the solution, enriching the reader's understanding and engagement. The text is self-contained. All necessary technical terms are explained.
- Published
- 2024
43. Intelligent Systems and Data Science : Second International Conference, ISDS 2024, Nha Trang, Vietnam, November 9–10, 2024, Proceedings, Part II
- Author
-
Nguyen Thai-Nghe, Thanh-Nghi Do, Salem Benferhat, Nguyen Thai-Nghe, Thanh-Nghi Do, and Salem Benferhat
- Subjects
- Application software, Artificial intelligence, Artificial intelligence—Data processing, Machine learning, Natural language processing (Computer science), Big data
- Abstract
This two-volume set constitutes the refereed proceedings of the Second International Conference, ISDS 2024, held in Nha Trang, Vietnam, during November 9–10, 2024. The 38 full papers and 10 short papers were carefully reviewed and selected from 129 submissions. They were categorized under the topical sections as follows: AI in E-Commerce, Agriculture, and Aquaculture; AI in Health Care Analytics; Big Data, IoT, and Cloud Computing; and Natural Language Processing.
- Published
- 2024
44. Natural Language Processing and Chinese Computing : 13th National CCF Conference, NLPCC 2024, Hangzhou, China, November 1–3, 2024, Proceedings, Part V
- Author
-
Derek F. Wong, Zhongyu Wei, Muyun Yang, Derek F. Wong, Zhongyu Wei, and Muyun Yang
- Subjects
- Artificial intelligence, Natural language processing (Computer science)
- Abstract
The five-volume set LNCS 15359 - 15363 constitutes the refereed proceedings of the 13th National CCF Conference on Natural Language Processing and Chinese Computing, NLPCC 2024, held in Hangzhou, China, during November 2024. The 161 full papers and 33 evaluation workshop papers included in these proceedings were carefully reviewed and selected from 451 submissions. They deal with the following areas: Fundamentals of NLP; Information Extraction and Knowledge Graph; Information Retrieval, Dialogue Systems, and Question Answering; Large Language Models and Agents; Machine Learning for NLP; Machine Translation and Multilinguality; Multi-modality and Explainability; NLP Applications and Text Mining; Sentiment Analysis, Argumentation Mining, and Social Media; Summarization and Generation.
- Published
- 2024
45. Natural Language Processing and Chinese Computing : 13th National CCF Conference, NLPCC 2024, Hangzhou, China, November 1–3, 2024, Proceedings, Part IV
- Author
-
Derek F. Wong, Zhongyu Wei, Muyun Yang, Derek F. Wong, Zhongyu Wei, and Muyun Yang
- Subjects
- Artificial intelligence, Natural language processing (Computer science)
- Abstract
The five-volume set LNCS 15359 - 15363 constitutes the refereed proceedings of the 13th National CCF Conference on Natural Language Processing and Chinese Computing, NLPCC 2024, held in Hangzhou, China, during November 2024. The 161 full papers and 33 evaluation workshop papers included in these proceedings were carefully reviewed and selected from 451 submissions. They deal with the following areas: Fundamentals of NLP; Information Extraction and Knowledge Graph; Information Retrieval, Dialogue Systems, and Question Answering; Large Language Models and Agents; Machine Learning for NLP; Machine Translation and Multilinguality; Multi-modality and Explainability; NLP Applications and Text Mining; Sentiment Analysis, Argumentation Mining, and Social Media; Summarization and Generation.
- Published
- 2024
46. Hands-On Large Language Models
- Author
-
Jay Alammar, Maarten Grootendorst, Jay Alammar, and Maarten Grootendorst
- Subjects
- Natural language processing (Computer science), Artificial intelligence--Computer programs, Natural language generation (Computer science)
- Abstract
AI has acquired startling new language capabilities in just the past few years. Driven by the rapid advances in deep learning, language AI systems are able to write and understand text better than ever before. This trend enables the rise of new features, products, and entire industries. With this book, Python developers will learn the practical tools and concepts they need to use these capabilities today.You'll learn how to use the power of pre-trained large language models for use cases like copywriting and summarization; create semantic search systems that go beyond keyword matching; build systems that classify and cluster text to enable scalable understanding of large amounts of text documents; and use existing libraries and pre-trained models for text classification, search, and clusterings.This book also shows you how to:Build advanced LLM pipelines to cluster text documents and explore the topics they belong toBuild semantic search engines that go beyond keyword search with methods like dense retrieval and rerankersLearn various use cases where these models can provide valueUnderstand the architecture of underlying Transformer models like BERT and GPTGet a deeper understanding of how LLMs are trainedUnderstanding how different methods of fine-tuning optimize LLMs for specific applications (generative model fine-tuning, contrastive fine-tuning, in-context learning, etc.)
- Published
- 2024
47. Rules and Reasoning : 8th International Joint Conference, RuleML+RR 2024, Bucharest, Romania, September 16–18, 2024, Proceedings
- Author
-
Sabrina Kirrane, Mantas Šimkus, Ahmet Soylu, Dumitru Roman, Sabrina Kirrane, Mantas Šimkus, Ahmet Soylu, and Dumitru Roman
- Subjects
- Computer science, Information technology—Management, Database management, Logic programming, Natural language processing (Computer science), Expert systems (Computer science)
- Abstract
This book constitutes the proceedings of the 8th International Joint Conference on Rules and Reasoning, RuleML+RR 2024, held in Bucharest, Romania, during September 16-18, 2024. The 12 full papers and 4 short papers included in this book were carefully reviewed and selected from 35 submissions. The RuleML+RR symposia were devoted to disseminating research, applications, languages, and standards for rule technologies, with attention to both theoretical and practical developments, to challenging new ideas and to industrial applications.
- Published
- 2024
48. Formal Methods Teaching : 6th Formal Methods Teaching Workshop, FMTea 2024, Milan, Italy, September 10, 2024, Proceedings
- Author
-
Emil Sekerinski, Leila Ribeiro, Emil Sekerinski, and Leila Ribeiro
- Subjects
- Mathematical logic, Logic programming, Natural language processing (Computer science), Social sciences—Data processing, Software engineering, Microprogramming
- Abstract
This book constitutes the proceedings of the 6th International Workshop on Formal Methods Teaching, FMTea 2024, which was held in Milan, Italy, on September 10, 2024. The 7 full papers included in these proceedings were carefully reviewed and selected from 9 submissions. The book also contains one invited talk in full paper length. The papers focus on learning formal methods for the purpose of teaching and self-learning.
- Published
- 2024
49. UX for Enterprise ChatGPT Solutions : A Practical Guide to Designing Enterprise-grade LLMs
- Author
-
Richard H. Miller and Richard H. Miller
- Subjects
- Natural language processing (Computer science), Artificial intelligence, Business--Data processing
- Abstract
Create engaging AI experiences by mastering ChatGPT for business and leveraging user interface design practices, research methods, prompt engineering, the feeding lifecycle, and moreKey FeaturesLearn in-demand design thinking and user research techniques applicable to all conversational AI platformsMeasure the quality and evaluate ChatGPT from a customer's perspective for optimal user experienceSet up and use your secure private data, documents, and materials to enhance your ChatGPT modelsPurchase of the print or Kindle book includes a free PDF eBookBook DescriptionMany enterprises grapple with new technology, often hopping on the bandwagon only to abandon it when challenges emerge. This book is your guide to seamlessly integrating ChatGPT into enterprise solutions with a UX-centered approach. UX for Enterprise ChatGPT Solutions empowers you to master effective use case design and adapt UX guidelines through an engaging learning experience. Discover how to prepare your content for success by tailoring interactions to match your audience's voice, style, and tone using prompt-engineering and fine-tuning. For UX professionals, this book is the key to anchoring your expertise in this evolving field. Writers, researchers, product managers, and linguists will learn to make insightful design decisions. You'll explore use cases like ChatGPT-powered chat and recommendation engines, while uncovering the AI magic behind the scenes. The book introduces a and feeding model, enabling you to leverage feedback and monitoring to iterate and refine any Large Language Model solution. Packed with hundreds of tips and tricks, this guide will help you build a continuous improvement cycle suited for AI solutions. By the end, you'll know how to craft powerful, accurate, responsive, and brand-consistent generative AI experiences, revolutionizing your organization's use of ChatGPT.What you will learnAlign with user needs by applying design thinking to tailor ChatGPT to meet customer expectationsHarness user research to enhance chatbots and recommendation enginesTrack quality metrics and learn methods to evaluate and monitor ChatGPT's quality and usabilityEstablish and maintain a uniform style and tone with prompt engineering and fine-tuningApply proven heuristics by monitoring and assessing the UX for conversational experiences with trusted methodsRefine continuously by implementing an ongoing process for chatbot and feedingWho this book is forThis book is for user experience designers, product managers, and product owners of business and enterprise ChatGPT solutions who are interested in learning how to design and implement ChatGPT-4 solutions for enterprise needs. You should have a basic-to-intermediate level of understanding in UI/UX design concepts and fundamental knowledge of ChatGPT-4 and its capabilities.
- Published
- 2024
50. Build a Website with ChatGPT : No Coding Experience Necessary
- Author
-
Paul McFedries and Paul McFedries
- Subjects
- Web site development, Natural language processing (Computer science), Artificial intelligence
- Abstract
Create a portfolio of cool and creative websites—all without having to write your own code.Build a Website with ChatGPT teaches you zero-coding web development utilizing powerful generative AI tools like ChatGPT. If you can open a web browser, you're ready to start building—absolutely no coding experience required. Inside Build a Website with ChatGPT you'll learn the important skills of AI-assisted web programming, such as: • Crafting effective prompts to generate HTML, CSS, and JavaScript • Converting text into images with DALL-E integration • Building navigation bars, image galleries, and contact forms • Deploying fully functional sites to the web for free • Customizing the generated code for unique sites Inside Build a Website with ChatGPT you'll learn the high-level coding concepts that let you check and perfect AI output, prompting skills that deliver the exact code you need, and how to properly deploy your site to the web—for free! Annotated code samples and advice on code customization give you the perfect balance of understanding and convenience. Plus, you'll get access to a tried-and-tested repository of prompts and working code. About the technology You can build amazing websites even if you don't know HTML, CSS, and JavaScript. Just describe what you want in plain English, and let ChatGPT take care of the gnarly details! This book guides you step-by-step as you create user-friendly forms, interesting graphics, and interactive web pages using nothing but AI and your imagination. About the book Build a Website with ChatGPT shows you how to make websites in an AI-first world—no experience required! You'll start with the basics of generating pages with ChatGPT, and by the end of the second chapter your first site will be up and running. Author Paul McFedries then shows you how to add interesting text and graphics, forms for user input, and even custom CSS to give your pages some pizzazz. As you go, you'll expand your new AI skills to create photo galleries, portfolios, catalog pages and more. What's inside • Writing effective prompts to create code, text, and graphics • Adding navigation bars, image galleries, and contact forms • Deploying your sites to the web for free • Adding your unique touches to AI-generated pages About the reader No experience with web development or programming required. If you can create a Word document, you can build a website! About the author Paul McFedries has written over 100 books on web development and other technology topics including Web Design Playground (Manning Publications). The technical editor on this book was Anirudh V. Prabhu. Table of Contents 1 Introducing website creation with ChatGPT 2 Creating and deploying your first web page 3 Working with fonts, colors, and headings 4 Adding structure to a page 5 Publishing page posts 6 Adding links and navigation 7 Creating site content 8 Generating site forms 9 Adding lists to your pages 10 Setting up a photo gallery 11 Creating a portfolio page 12 Building an article page 13 Coding an interactive course catalog A Getting ready to build web pages with ChatGPT B Deploying your site C Learning a few ChatGPT best practices
- Published
- 2024
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.