54 results on '"Dinh Phung"'
Search Results
2. Improved speech emotion recognition based on music-related audio features
- Author
-
Linh Vu, Raphael C.-W. Phan, Lim Wern Han, and Dinh Phung
- Published
- 2022
3. MED-TEX: Transfer and Explain Knowledge with Less Data from Pretrained Medical Imaging Models
- Author
-
Thanh Nguyen-Duc, He Zhao, Jianfei Cai, and Dinh Phung
- Published
- 2022
4. STEM: An approach to Multi-source Domain Adaptation with Guarantees
- Author
-
Van-Anh Nguyen, Tuan Nguyen, Trung Le, Quan Hung Tran, and Dinh Phung
- Published
- 2021
5. Information-theoretic Source Code Vulnerability Highlighting
- Author
-
Trung Le, Van N.B. Nguyen, Dinh Phung, John Grundy, Paul Montague, and Olivier Y. de Vel
- Subjects
Source code ,business.industry ,Computer science ,media_common.quotation_subject ,Deep learning ,Feature extraction ,Data science ,Variety (cybernetics) ,Identification (information) ,Software ,Artificial intelligence ,business ,Function (engineering) ,media_common ,Vulnerability (computing) - Abstract
Software vulnerabilities are a crucial and serious concern in the software industry and computer security. A variety of methods have been proposed to detect vulnerabilities in real-world software. Recent methods based on deep learning approaches for automatic feature extraction have improved software vulnerability identification compared with machine learning approaches based on hand-crafted feature extraction. However, these methods can usually only detect software vulnerabilities at a function or program level, which is much less informative because, out of hundreds (thousands) of code statements in a program or function, only a few core statements contribute to a software vulnerability. This requires us to find a way to detect software vulnerabilities at a fine-grained level. In this paper, we propose a novel method based on the concept of mutual information that can help us to detect and isolate software vulnerabilities at a fine-grained level (i.e., several statements that are highly relevant to a software vulnerability that include the core vulnerable statements) in both unsupervised and semi-supervised contexts. We conduct comprehensive experiments on real-world software projects to demonstrate that our proposed method can detect vulnerabilities at a fine-grained level by identifying several statements that mostly contribute to the vulnerability detection decision.
- Published
- 2021
6. Stein Variational Gradient Descent with Variance Reduction
- Author
-
Dinh Phung, Viet Huynh, Nhan Dam, and Trung Le
- Subjects
Random field ,Artificial neural network ,Markov chain ,business.industry ,Computer science ,Stochastic process ,Probabilistic logic ,Inference ,Estimator ,010501 environmental sciences ,Probabilistic inference ,Machine learning ,computer.software_genre ,Bayesian inference ,01 natural sciences ,010104 statistics & probability ,Binary classification ,Variance reduction ,Artificial intelligence ,0101 mathematics ,Gradient descent ,business ,computer ,0105 earth and related environmental sciences - Abstract
Probabilistic inference is a common and important task in statistical machine learning. The recently proposed Stein variational gradient descent (SVGD) is a generic Bayesian inference method that has been shown to be successfully applied in a wide range of contexts, especially in dealing with large datasets, where existing probabilistic inference methods have been known to be ineffective. In a large-scale data setting, SVGD employs the mini-batch strategy but its mini-batch estimator has large variance, hence compromising its estimation quality in practice. To this end, we propose in this paper a generic SVGD-based inference method that can significantly reduce the variance of mini-batch estimator when working with large datasets. Our experiments on 14 datasets show that the proposed method enjoys substantial and consistent improvements compared with baseline methods in binary classification task and its pseudo-online learning setting, and regression task. Furthermore, our framework is generic and applicable to a wide range of probabilistic inference problems such as in Bayesian neural networks and Markov random fields.
- Published
- 2020
7. Code Pointer Network for Binary Function Scope Identification
- Author
-
Dinh Phung, Van N.B. Nguyen, Tue Le, Trung Le, Olivier Y. de Vel, Khanh Nguyen, and Paul Montague
- Subjects
Binary analysis ,Binary function ,business.industry ,Computer science ,Deep learning ,02 engineering and technology ,010501 environmental sciences ,computer.software_genre ,01 natural sciences ,Pointer (computer programming) ,0202 electrical engineering, electronic engineering, information engineering ,Microsoft Windows ,Malware ,020201 artificial intelligence & image processing ,Data mining ,Artificial intelligence ,business ,computer ,0105 earth and related environmental sciences - Abstract
Function identification is a preliminary step in binary analysis for many extensive applications from malware detection, common vulnerability detection and binary instrumentation to name a few. In this paper, we propose the Code Pointer Network that leverages the underlying idea of a pointer network to efficiently and effectively tackle function scope identification – the hardest and most crucial task in function identification. We establish extensive experiments to compare our proposed method with the deep learning based baseline. Experimental results demonstrate that our proposed method significantly outperforms the state-of-the-art baseline in terms of both predictive performance and running time.
- Published
- 2020
8. Deep Domain Adaptation for Vulnerable Code Function Identification
- Author
-
Lizhen Qu, Olivier deVel, Trung Le, Paul Montague, Van N.B. Nguyen, Khanh Nguyen, Tue Le, and Dinh Phung
- Subjects
Exploit ,business.industry ,Computer science ,Vulnerability ,Context (language use) ,Machine learning ,computer.software_genre ,Identification (information) ,Software ,Margin (machine learning) ,Software security assurance ,Artificial intelligence ,business ,Cluster analysis ,Feature learning ,computer ,Vulnerability (computing) - Abstract
Due to the ubiquity of computer software, software vulnerability detection (SVD) has become crucial in the software industry and in the field of computer security. Two significant issues in SVD arise when using machine learning, namely: i) how to learn automatic features that can help improve the predictive performance of vulnerability detection and ii) how to overcome the scarcity of labeled vulnerabilities in projects that require the laborious labeling of code by software security experts. In this paper, we address these two crucial concerns by proposing a novel architecture which leverages deep domain adaptation with automatic feature learning for software vulnerability identification. Based on this architecture, we keep the principles and reapply the state-of-the-art deep domain adaptation methods to indicate that deep domain adaptation for SVD is plausible and promising. Moreover, we further propose a novel method named Semi-supervised Code Domain Adaptation Network (SCDAN) that can efficiently utilize and exploit information carried in unlabeled target data by considering them as the unlabeled portion in a semi-supervised learning context. The proposed SCDAN method enforces the clustering assumption, which is a key principle in semi-supervised learning. The experimental results using six real-world software project datasets show that our SCDAN method and the baselines using our architecture have better predictive performance by a wide margin compared with the Deep Code Network (VulDeePecker) method without domain adaptation. Also, the proposed SCDAN significantly outperforms the DIRT-T which to the best of our knowledge is currently the-state-of-the-art method in deep domain adaptation and other baselines.
- Published
- 2019
9. Bayesian Multi-Hyperplane Machine for Pattern Recognition
- Author
-
Tu Nguyen Dinh, Dinh Phung, Khanh Nguyen, and Trung Le
- Subjects
021103 operations research ,Optimization problem ,Discretization ,Computer science ,business.industry ,Bayesian probability ,0211 other engineering and technologies ,Probabilistic logic ,Markov chain Monte Carlo ,02 engineering and technology ,01 natural sciences ,010104 statistics & probability ,symbols.namesake ,Stochastic gradient descent ,Hyperplane ,Hyperparameter optimization ,symbols ,Artificial intelligence ,0101 mathematics ,business ,Algorithm ,Parametric statistics - Abstract
Current existing multi-hyperplane machine approach deals with high-dimensional and complex datasets by approximating the input data region using a parametric mixture of hyperplanes. Consequently, this approach requires an excessively time-consuming parameter search to find the set of optimal hyper-parameters. Another serious drawback of this approach is that it is often suboptimal since the optimal choice for the hyper-parameter is likely to lie outside the searching space due to the space discretization step required in grid search. To address these challenges, we propose in this paper BAyesian Multi-hyperplane Machine (BAMM). Our approach departs from a Bayesian perspective, and aims to construct an alternative probabilistic view in such a way that its maximum-a-posteriori (MAP) estimation reduces exactly to the original optimization problem of a multi-hyperplane machine. This view allows us to endow prior distributions over hyper-parameters and augment auxiliary variables to efficiently infer model parameters and hyper-parameters via Markov chain Monte Carlo (MCMC) method. We then employ a Stochastic Gradient Descent (SGD) framework to scale our model up with ever-growing large datasets. Extensive experiments demonstrate the capability of our proposed method in learning the optimal model without using any parameter tuning, and in achieving comparable accuracies compared with the state-of-art baselines; in the meantime our model can seamlessly handle with large-scale datasets.
- Published
- 2018
10. GoGP: Fast Online Regression with Gaussian Processes
- Author
-
Khanh Nguyen, Trung Le, Tu Dinh Nguyen, Dinh Phung, and Vu Nguyen
- Subjects
Speedup ,Computational complexity theory ,Computer science ,02 engineering and technology ,010501 environmental sciences ,Perceptron ,01 natural sciences ,Support vector machine ,symbols.namesake ,Rate of convergence ,Kriging ,0202 electrical engineering, electronic engineering, information engineering ,symbols ,020201 artificial intelligence & image processing ,Gaussian process ,Algorithm ,0105 earth and related environmental sciences - Abstract
One of the most current challenging problems in Gaussian process regression (GPR) is to handle large-scale datasets and to accommodate an online learning setting where data arrive irregularly on the fly. In this paper, we introduce a novel online Gaussian process model that could scale with massive datasets. Our approach is formulated based on alternative representation of the Gaussian process under geometric and optimization views, hence termed geometric-based online GP (GoGP). We developed theory to guarantee that with a good convergence rate our proposed algorithm always produces a (sparse) solution which is close to the true optima to any arbitrary level of approximation accuracy specified a priori. Furthermore, our method is proven to scale seamlessly not only with large-scale datasets, but also to adapt accurately with streaming data. We extensively evaluated our proposed model against state-of-the-art baselines using several large-scale datasets for online regression task. The experimental results show that our GoGP delivered comparable, or slightly better, predictive performance while achieving a magnitude of computational speedup compared with its rivals under online setting. More importantly, its convergence behavior is guaranteed through our theoretical analysis, which is rapid and stable while achieving lower errors.
- Published
- 2017
11. Forward-Backward Smoothing for Hidden Markov Models of Point Pattern Data
- Author
-
Viet Huynh, Ba-Ngu Vo, Nhan Dam, and Dinh Phung
- Subjects
Estimation theory ,Computer science ,Computation ,0202 electrical engineering, electronic engineering, information engineering ,Inference ,020206 networking & telecommunications ,Point (geometry) ,02 engineering and technology ,Hidden Markov model ,Algorithm ,Smoothing ,Point process ,Data modeling - Abstract
This paper considers a discrete-time sequential latent model for point pattern data, specifically a hidden Markov model (HMM) where each observation is an instantiation of a random finite set (RFS). This so-called RFS-HMM is worthy of investigation since point pattern data are ubiquitous in artificial intelligence and data science. We address the three basic problems typically encountered in such a sequential latent model, namely likelihood computation, hidden state inference, and parameter estimation. Moreover, we develop algorithms for solving these problems including forward-backward smoothing for likelihood computation and hidden state inference, and expectation-maximisation for parameter estimation. Simulation studies are used to demonstrate key properties of RFS-HMM, whilst real data in the domain of human dynamics are used to demonstrate its applicability.
- Published
- 2017
12. Animal Recognition and Identification with Deep Convolutional Neural Networks for Automated Wildlife Monitoring
- Author
-
Tu Dinh Nguyen, Paul Flemons, Kylie Andrews, Euan G. Ritchie, Thin Nguyen, Dinh Phung, Sarah J. Maclagan, and Hung Nguyen
- Subjects
0106 biological sciences ,business.industry ,Computer science ,Deep learning ,Wildlife ,02 engineering and technology ,Machine learning ,computer.software_genre ,010603 evolutionary biology ,01 natural sciences ,Automation ,Convolutional neural network ,Covert ,Obstacle ,0202 electrical engineering, electronic engineering, information engineering ,Citizen science ,020201 artificial intelligence & image processing ,Identification (biology) ,Artificial intelligence ,business ,computer - Abstract
Efficient and reliable monitoring of wild animals in their natural habitats is essential to inform conservation and management decisions. Automatic covert cameras or "camera traps" are being an increasingly popular tool for wildlife monitoring due to their effectiveness and reliability in collecting data of wildlife unobtrusively, continuously and in large volume. However, processing such a large volume of images and videos captured from camera traps manually is extremely expensive, time-consuming and also monotonous. This presents a major obstacle to scientists and ecologists to monitor wildlife in an open environment. Leveraging on recent advances in deep learning techniques in computer vision, we propose in this paper a framework to build automated animal recognition in the wild, aiming at an automated wildlife monitoring system. In particular, we use a single-labeled dataset from Wildlife Spotter project, done by citizen scientists, and the state-of-the-art deep convolutional neural network architectures, to train a computational system capable of filtering animal images and identifying species automatically. Our experimental results achieved an accuracy at 96.6% for the task of detecting images containing animal, and 90.4% for identifying the three most common species among the set of images of wild animals taken in South-central Victoria, Australia, demonstrating the feasibility of building fully automated wildlife observation. This, in turn, can therefore speed up research findings, construct more efficient citizen sciencebased monitoring systems and subsequent management decisions, having the potential to make significant impacts to the world of ecology and trap camera images analysis.
- Published
- 2017
13. Distributed data augmented support vector machine on Spark
- Author
-
Trung Le, Vu Nguyen, Tu Dinh Nguyen, and Dinh Phung
- Subjects
Structured support vector machine ,Distributed database ,Computer science ,business.industry ,Big data ,02 engineering and technology ,computer.software_genre ,Machine learning ,01 natural sciences ,Data modeling ,Support vector machine ,010104 statistics & probability ,ComputingMethodologies_PATTERNRECOGNITION ,Data point ,020204 information systems ,Scalability ,Spark (mathematics) ,0202 electrical engineering, electronic engineering, information engineering ,Data mining ,Artificial intelligence ,0101 mathematics ,business ,computer - Abstract
Support vector machines (SVMs) are widely-used for classification in machine learning and data mining tasks. However, they traditionally have been applied to small to medium datasets. Recent need to scale up with data size has attracted research attention to develop new methods and implementation for SVM to perform tasks at scale. Distributed SVMs are relatively new and studied recently, but the distributed implementation for SVM with data augmentation has not been developed. This paper introduces a distributed data augmentation implementation for SVM on Apache Spark, a recent advanced and popular platform for distributed computing that has been employed widely in research as well as in industry. We term our implementation sparkling vector machine (SkVM) which supports both classification and regression tasks by scanning through the data exactly once. In addition, we further develop a framework to handle the data with new classes arriving under an online classification setting where new data points can have labels that have not previously seen - a problem we term label-drift classification. We demonstrate the scalability of our proposed method on large-scale datasets with more than one hundred million data points. The experimental results show that the predictive performances of our method are comparable or better than those of baselines whilst the execution time is much faster at an order of magnitude.
- Published
- 2016
14. One-Pass Logistic Regression for Label-Drift and Large-Scale Classification on Distributed Systems
- Author
-
Trung Le, Vu Nguyen, Dinh Phung, Svetha Venkatesh, and Tu Dinh Nguyen
- Subjects
Hyperparameter ,Computer science ,business.industry ,Big data ,02 engineering and technology ,computer.software_genre ,Machine learning ,Logistic regression ,01 natural sciences ,Data modeling ,010104 statistics & probability ,ComputingMethodologies_PATTERNRECOGNITION ,Data point ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Data mining ,Artificial intelligence ,0101 mathematics ,Invariant (mathematics) ,business ,computer - Abstract
Logistic regression (LR) for classification is the workhorse in industry, where a set of predefined classes is required. The model, however, fails to work in the case where the class labels are not known in advance, a problem we term label-drift classification. Label-drift classification problem naturally occurs in many applications, especially in the context of streaming settings where the incoming data may contain samples categorized with new classes that have not been previously seen. Additionally, in the wave of big data, traditional LR methods may fail due to their expense of running time. In this paper, we introduce a novel variant of LR, namely one-pass logistic regression (OLR) to offer a principled treatment for label-drift and large-scale classifications. To handle largescale classification for big data, we further extend our OLR to a distributed setting for parallelization, termed sparkling OLR (Spark-OLR). We demonstrate the scalability of our proposed methods on large-scale datasets with more than one hundred million data points. The experimental results show that the predictive performances of our methods are comparable orbetter than those of state-of-the-art baselines whilst the executiontime is much faster at an order of magnitude. In addition, the OLR and Spark-OLR are invariant to data shuffling and have no hyperparameter to tune that significantly benefits data practitioners and overcomes the curse of big data cross-validationto select optimal hyperparameters.
- Published
- 2016
15. MCNC: Multi-Channel Nonparametric Clustering from heterogeneous data
- Author
-
Thanh-Binh Nguyen, Dinh Phung, Vu Nguyen, and Svetha Venkatesh
- Subjects
Flexibility (engineering) ,Context model ,Computer science ,business.industry ,Nonparametric statistics ,Context (language use) ,computer.software_genre ,Missing data ,Machine learning ,Data modeling ,Key (cryptography) ,Artificial intelligence ,Data mining ,business ,Cluster analysis ,computer - Abstract
Bayesian nonparametric (BNP) models have recently become popular due to their flexibility in identifying the unknown number of clusters. However, they have difficulties handling heterogeneous data from multiple sources. Existing BNP methods either treat each of these sources independently - hence do not get benefits from the correlating information between them, or require to explicitly specify data sources as primary and context channels. In this paper, we present a BNP framework, termed MCNC, which has the ability to (1) discover co-patterns from multiple sources; (2) explore multi-channel data simultaneously and treat them equally; (3) automatically identify a suitable number of patterns from data; and (4) handle missing data. The key idea is to utilize a richer base measure of a BNP model being a product-space. We demonstrate our framework on synthetic and real-world datasets to discover the identity-location-time (a.k.a who-where-when) patterns. The experimental results highlight the effectiveness of our MCNC framework in both cases of complete and missing data.
- Published
- 2016
16. Stable clinical prediction using graph support vector machines
- Author
-
Dinh Phung, Svetha Venkatesh, Iman Kamkar, Sunil Gupta, and Cheng Li
- Subjects
0301 basic medicine ,Clustering high-dimensional data ,Jaccard index ,business.industry ,Model selection ,Feature selection ,Pattern recognition ,01 natural sciences ,Support vector machine ,010104 statistics & probability ,03 medical and health sciences ,030104 developmental biology ,Graph (abstract data type) ,Pairwise comparison ,Artificial intelligence ,0101 mathematics ,Convex function ,business ,Mathematics - Abstract
The stability matters in clinical prediction models because it makes the model to be interpretable and generalizable. It is paramount for high dimensional data, which employ sparse models with feature selection ability. We propose a new method to stabilize sparse support vector machines using intrinsic graph structure of the electronic medical records. The graph structure is exploited using the Jaccard similarity among features. Our method employs a convex function to penalize the pairwise l ∞ -norm of connected feature coefficients in the graph. We apply the alternating direction method of multipliers to solve the proposed formulation. Our experiments are conducted on a synthetic and three real-world hospital datasets. We show that our proposed method is more stable than the state-of-the-art feature selection and classification techniques in terms of three stability measures namely, Jaccard similarity measure, Spearman's rank correlation coefficient and Kuncheva index. We further show that our method has resulted in better classification performance compared to the baselines.
- Published
- 2016
17. Transfer learning for rare cancer problems via Discriminative Sparse Gaussian Graphical model
- Author
-
Sunil Gupta, Svetha Venkatesh, Dinh Phung, and Budhaditya Saha
- Subjects
Generalization ,business.industry ,Computer science ,Gaussian ,Pattern recognition ,02 engineering and technology ,Machine learning ,computer.software_genre ,01 natural sciences ,Data modeling ,Task (project management) ,010104 statistics & probability ,symbols.namesake ,Data point ,Discriminative model ,0202 electrical engineering, electronic engineering, information engineering ,symbols ,020201 artificial intelligence & image processing ,Graphical model ,Artificial intelligence ,0101 mathematics ,Transfer of learning ,business ,computer - Abstract
Mortality prediction of rare cancer types with a small number of high-dimensional samples is a challenging task. We propose a transfer learning model where both classes in rare cancers (target task) are modeled in a joint framework by transferring knowledge from the source task. The knowledge transfer is at the data level where only “related” data points are chosen to train the target task. Moreover, both positive and negative class in training enhances the discrimination power of the proposed framework. Overall, this approach boosts the generalization performance of target task with a small number of data points. The formulation of the proposed framework is convex and expressed as a primal problem. We convert this to a dual problem and efficiently solve by alternating direction multipliers method. Our experiments with both synthetic and three real-world datasets show that our framework outperforms state-of-the-art single-task, multi-task, and transfer learning baselines.
- Published
- 2016
18. Effect of social capital on emotion, language style and latent topics in online depression community
- Author
-
Bo Dao, Dinh Phung, Svetha Venkatesh, and Thin Nguyen
- Subjects
Social computing ,medicine ,Social media ,Social competence ,Social engagement ,Psychology ,Mental illness ,medicine.disease ,Mental health ,Social psychology ,Style (sociolinguistics) ,Social capital - Abstract
Social capital is linked to mental illness. It has been proposed that higher social capital is associated with better mental well-being in both individuals and groups in offline setting. However, in online settings, the association between online social capital and mental health conditions has not yet been explored. Social media offer us a rich opportunity to determine the link between social capital and aspects of mental wellbeing. In this paper, we examine social capital based on levels of social connectivity of bloggers can be connected to aspects of depression in individuals and online depression community. We explore apparent properties of textual contents, including expressed emotions, language styles and latent topics, of a large corpus of blog posts, to analyze the aspect of social capital in the community. Using data collected from online Livejoumal depression community, we apply both statistical tests and machine learning approaches to examine how predictive factors vary between low and high social capital groups. Significant differences are found between low and high social capital groups when characterized by a set of latent topics, language features derived from blog posts, suggesting discriminative features, proved to be useful in the classification task. This shows that linguistic styles are better predictors than latent topics as features. The findings indicate the potential of using social media as a sensor for monitoring mental well-being in online settings.
- Published
- 2016
19. Learning Multifaceted Latent Activities from Heterogeneous Mobile Data
- Author
-
Dinh Phung, Mohan Kumar, Vu Nguyen, Svetha Venkatesh, Thuong Nguyen, and Thanh-Binh Nguyen
- Subjects
Hierarchical Dirichlet process ,Structure (mathematical logic) ,Topic model ,business.industry ,Computer science ,Context (language use) ,02 engineering and technology ,Machine learning ,computer.software_genre ,01 natural sciences ,Synthetic data ,Data modeling ,Activity recognition ,010104 statistics & probability ,ComputingMethodologies_PATTERNRECOGNITION ,0202 electrical engineering, electronic engineering, information engineering ,Data analysis ,020201 artificial intelligence & image processing ,Artificial intelligence ,Data mining ,0101 mathematics ,business ,computer - Abstract
Inferring abstract contexts and activities from heterogeneous data is vital to context-aware ubiquitous applications but still remains one of the most challenging problems. Recent advances in Bayesian nonparametric machine learning, in particular the theory of topic models based on Hierarchical Dirichlet Process (HDP), has provided an elegant solution towards these challenges. However, limited existing methods have addressed the problem of inferring latent multifaceted activities and contexts from heterogeneous data sources such as those collected from mobile devices. In this paper, we extend the original HDP to model heterogeneous data using a richer structure of the base measure being a product-space. The proposed model, called product-space HDP (PS-HDP), naturally handles the heterogeneous data from multiple sources and identify the unknown number of latent structures in a principle way. Although this framework is generic, our current work primarily focuses on inferring (latent) threefold activities of who-when-where simultaneously, which corresponds to inducing activities from data collected for identity, location and time. We demonstrate our model on synthetic data as well as on a real-world dataset – the StudentLife dataset. We report results and provide analysis on the discovered activities and patterns to demonstrate the merit of the model. We also quantitatively evaluate the performance of PS-HDP model using standard metrics including F1-score, NMI, RI, purity, and compare them with well-known existing baseline methods.
- Published
- 2016
20. Analysing the History of Autism Spectrum Disorder Using Topic Models
- Author
-
Dinh Phung, Adham Beykikhoshk, Ognjen Arandjelovic, Svetha Venkatesh, and University of St Andrews. School of Computer Science
- Subjects
QA75 ,0301 basic medicine ,Topic model ,Source code ,Computer science ,QA75 Electronic computers. Computer science ,QH301 Biology ,media_common.quotation_subject ,NDAS ,02 engineering and technology ,Data modeling ,Bayesian nonparametrics ,QH301 ,03 medical and health sciences ,020204 information systems ,Multiple time dimensions ,0202 electrical engineering, electronic engineering, information engineering ,Social media ,Autism spectrum disorder ,Data mining ,media_common ,Structure (mathematical logic) ,Probabilistic logic ,Popularity ,Data science ,030104 developmental biology ,RC0321 ,RC0321 Neuroscience. Biological psychiatry. Neuropsychiatry - Abstract
We describe a novel framework for the discovery of underlying topics of a longitudinal collection of scholarly data, and the tracking of their lifetime and popularity over time. Unlike the social media or news data where the underlying topics evolve over time, the topic nuances in science result in new scientific directions to emerge. Therefore, we model the longitudinal literature data with a new approach that uses topics which remain identifiable over the course of time. Current studies either disregard the time dimension or treat it as an exchangeable covariate when they fix the topics over time or do not share the topics over epochs when they model the time naturally. We address these issues by adopting a non-parametric Bayesian approach. We assume the data is partially exchangeable and divide it into consecutive epochs. Then, by fixing the topics in a recurrent Chinese restaurant franchise, we impose a static topical structure on the corpus such that the topics are shared across epochs and the documents within epochs. We demonstrate the effectiveness of the proposed framework on a collection of medical literature related to autism spectrum disorder. We collect a large corpus of publications and carefully examine two important research issues of the domain as case studies. Moreover, we make the results of our experiment and the source code of the model, freely available to the public. This aids other researchers to analyse our results or apply the model to their data collections.
- Published
- 2016
21. Forecasting Patient Outflow from Wards having No Real-Time Clinical Data
- Author
-
Truyen Tran, Dinh Phung, Shivapratap Gopakumar, Svetha Venkatesh, and Wei Luo
- Subjects
Mean squared error ,business.industry ,030208 emergency & critical care medicine ,01 natural sciences ,Data modeling ,Random forest ,010104 statistics & probability ,03 medical and health sciences ,0302 clinical medicine ,Moving average ,Statistics ,Medicine ,Outflow ,Autoregressive integrated moving average ,0101 mathematics ,Time series ,Symmetric mean absolute percentage error ,business - Abstract
Modelling patient flow is crucial in understanding resource demand and prioritization. To date, there has been limited work in predicting ward-level discharges. Our study investigates forecasting total next-day discharges from an open ward. In the absence of real-time clinical data, we propose to construct a feature set from patient demographics, ward data and discharge time series to derive a random forest model for forecasting daily discharge. Using data from a general ward of a large regional Australian hospital, we compared our random forest model with a classical auto-regressive integrated moving average (ARIMA) for 12,141 patient visits over 1826 days. Forecasting quality was measured using Mean Forecast Error, Mean Absolute Error, symmetric Mean Absolute Percentage Error and Root Mean Square Error. When compared to the baseline model, next day discharge forecasts using random forests achieved 17.4 % improvement in Mean Absolute Error, for all days in the year 2014.
- Published
- 2016
22. Discovering latent affective dynamics among individuals in online mental health-related communities
- Author
-
Dinh Phung, Thin Nguyen, Svetha Venkatesh, and Bo Dao
- Subjects
Computer science ,02 engineering and technology ,Affect (psychology) ,medicine.disease ,Mental health ,Developmental psychology ,Mood ,Dynamics (music) ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,medicine ,Anxiety ,Autism ,020201 artificial intelligence & image processing ,Social media ,medicine.symptom - Abstract
Discovering dynamics of emotion and mood changes for individuals has the potential to enhance the diagnosis and treatment of mental disorders. In this paper we study affective transitions and dynamics among individuals in online mental health communities. Using social media as form of ‘sensor’, we crawl a large dataset of blogs posted by online communities whose descriptions declared to be associated with affective disorder conditions such as depression, anxiety, or autism. We then apply nonnegative matrix factorization model to extract the common and individual factors of affective transitions across groups of individuals in different levels of affective disorders. We examine the latent patterns of emotional transitions and investigate the effects of emotional transitions across the cohorts. Our framework is novel as it utilizes social media as an online sensing platform of mood and emotional dynamics. Hence, our work has implication in constructing systems to screen individuals and communities at high risks of mental health problems in online settings.
- Published
- 2016
23. SECC: Simultaneous extraction of context and community from pervasive signals
- Author
-
Thuong Nguyen, Dinh Phung, Vu Nguyen, and Flora D. Salim
- Subjects
Context model ,Ubiquitous computing ,Computer science ,Context (language use) ,02 engineering and technology ,Mixture model ,computer.software_genre ,01 natural sciences ,Data modeling ,Dirichlet process ,010104 statistics & probability ,Human dynamics ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Data mining ,0101 mathematics ,Cluster analysis ,computer - Abstract
Understanding user contexts and group structures plays a central role in pervasive computing. These contexts and community structures are complex to mine from data collected in the wild due to the unprecedented growth of data, noise, uncertainties and complexities. Typical existing approaches would first extract the latent patterns to explain the human dynamics or behaviors and then use them as the way to consistently formulate numerical representations for community detection, often via a clustering method. While being able to capture highorder and complex representations, these two steps are performed separately. More importantly, they face a fundamental difficulty in determining the correct number of latent patterns and communities. This paper presents an approach that seamlessly addresses these challenges to simultaneously discover latent patterns and communities in a unified Bayesian nonparametric framework. Our Simultaneous Extraction of Context and Community (SECC) model roots in the nested Dirichlet process theory which allows nested structure to be built to explain data at multiple levels. We demonstrate our framework on three public datasets where the advantages of the proposed approach are validated.
- Published
- 2016
24. Multi-View Subspace Clustering for Face Images
- Author
-
Wanquan Liu, Dinh Phung, Duc-Son Pham, Xin Zhang, and Svetha Venkatesh
- Subjects
Clustering high-dimensional data ,Biclustering ,CURE data clustering algorithm ,Computer science ,business.industry ,Correlation clustering ,Consensus clustering ,Canopy clustering algorithm ,Affinity propagation ,Pattern recognition ,Artificial intelligence ,Cluster analysis ,business - Abstract
In many real-world computer vision applications, such as multi-camera surveillance, the objects of interest are captured by visual sensors concurrently, resulting in multi-view data. These views usually provide complementary information to each other. One recent and powerful computer vision method for clustering is sparse subspace clustering (SSC); however, it was not designed for multi-view data, which break down its linear separability assumption. To integrate complementary information between views, multi-view clustering algorithms are required to improve the clustering performance. In this paper, we propose a novel multi-view subspace clustering by searching for an unified latent structure as a global affinity matrix in subspace clustering. Due to the integration of affinity matrices for each view, this global affinity matrix can best represent the relationship between clusters. This could help us achieve better performance on face clustering. We derive a provably convergent algorithm based on the alternating direction method of multipliers (ADMM) framework, which is computationally efficient, to solve the formulation. We demonstrate that this formulation outperforms other alternatives based on state-of-the-arts on challenging multi-view face datasets.
- Published
- 2015
25. Nonparametric discovery of online mental health-related communities
- Author
-
Thin Nguyen, Svetha Venkatesh, Bo Dao, and Dinh Phung
- Subjects
Topic model ,Multimedia ,Online participation ,Computer science ,Mood swing ,Sentiment analysis ,Cognition ,computer.software_genre ,Mental health ,Mood ,medicine ,Social media ,medicine.symptom ,computer ,Cognitive psychology - Abstract
People are increasingly using social media, especially online communities, to discuss mental health issues and seek supports. Understanding topics, interaction, sentiment and clustering structures of these communities informs important aspects of mental health. It can potentially add knowledge to the underlying cognitive dynamics, mood swings patterns, shared interests, and interaction. There has been growing research interest in analyzing online mental health communities; however sentiment analysis of these communities has been largely under-explored. This study presents an analysis of online Live Journal communities with and without mental health-related conditions including depression and autism. Latent topics for mood tags, affective words, and generic words in the content of the posts made in these communities were learned using nonparametric topic modelling. These representations were then input into a nonparametric clustering to discover meta-groups among the communities. The best performance results can be achieved on clustering communities with latent mood-based representation for such communities. The study also found significant differences in usage latent topics for mood tags and affective features between online communities with and without affective disorders. The findings reveal useful insights into hyper-group detection of online mental health-related communities.
- Published
- 2015
26. Exploiting feature relationships towards stable feature selection
- Author
-
Svetha Venkatesh, Dinh Phung, Sunil Gupta, and Iman Kamkar
- Subjects
Linear programming ,Covariance matrix ,Computer science ,business.industry ,Stability (learning theory) ,Regular polygon ,Feature selection ,Pattern recognition ,Machine learning ,computer.software_genre ,Correlation ,Lasso (statistics) ,Feature (computer vision) ,Artificial intelligence ,business ,computer - Abstract
Feature selection is an important step in building predictive models for most real-world problems. One of the popular methods in feature selection is Lasso. However, it shows instability in selecting features when dealing with correlated features. In this work, we propose a new method that aims to increase the stability of Lasso by encouraging similarities between features based on their relatedness, which is captured via a feature covariance matrix. Besides modeling positive feature correlations, our method can also identify negative correlations between features. We propose a convex formulation for our model along with an alternating optimization algorithm that can learn the weights of the features as well as the relationship between them. Using both synthetic and real-world data, we show that the proposed method is more stable than Lasso and many state-of-the-art shrinkage and feature selection methods. Also, its predictive performance is comparable to other methods.
- Published
- 2015
27. Visual Object Clustering via Mixed-Norm Regularization
- Author
-
Dinh Phung, Svetha Venkatesh, Duc-Son Pham, Wanquan Liu, Budhaditya Saha, and Xin Zhang
- Subjects
Biclustering ,Clustering high-dimensional data ,K-SVD ,Brown clustering ,business.industry ,Computer science ,CURE data clustering algorithm ,Correlation clustering ,Canopy clustering algorithm ,Pattern recognition ,Artificial intelligence ,business ,Cluster analysis - Abstract
Many vision problems deal with high-dimensional data, such as motion segmentation and face clustering. However, these high-dimensional data usually lie in a low-dimensional structure. Sparse representation is a powerful principle for solving a number of clustering problems with high-dimensional data. This principle is motivated from an ideal modeling of data points according to linear algebra theory. However, real data in computer vision are unlikely to follow the ideal model perfectly. In this paper, we exploit the mixed norm regularization for sparse subspace clustering. This regularization term is a convex combination of the l1norm, which promotes sparsity at the individual level and the block norm l2/1 which promotes group sparsity. Combining these powerful regularization terms will provide a more accurate modeling, subsequently leading to a better solution for the affinity matrix used in sparse subspace clustering. This could help us achieve better performance on motion segmentation and face clustering problems. This formulation also caters for different types of data corruptions. We derive a provably convergent algorithm based on the alternating direction method of multipliers (ADMM) framework, which is computationally efficient, to solve the formulation. We demonstrate that this formulation outperforms other state-of-arts on both motion segmentation and face clustering.
- Published
- 2015
28. Individualized arrhythmia detection with ECG signals from wearable devices
- Author
-
Dinh Phung, Terry Caelli, Svetha Venkatesh, Thanh-Binh Nguyen, Wei Lou, Nguyen, Thanh-Binh, Lou, Wei, Caelli, Terry, Venkatesh, Svetha, Phung, Dinh, and 2014 International Conference on Data Science and Advanced Analytics (DSAA) Shanghai, China 30 October - 1 November 2014
- Subjects
Arrhythmia detection ,ECG ,business.industry ,Computer science ,Individual difference ,Robust statistics ,Wearable computer ,Pattern recognition ,arrhythmia detection ,wearable devices ,ComputingMethodologies_PATTERNRECOGNITION ,classification ,Robustness (computer science) ,ComputerSystemsOrganization_SPECIAL-PURPOSEANDAPPLICATION-BASEDSYSTEMS ,cardiovascular diseases ,Artificial intelligence ,Ecg signal ,Fiducial marker ,business ,Wearable technology - Abstract
Low cost pervasive electrocardiogram (ECG) monitors is changing how sinus arrhythmia are diagnosed among patients with mild symptoms. With the large amount of data generated from long-term monitoring, come new data science and analytical challenges. Although traditional rule-based detection algorithms still work on relatively short clinical quality ECG, they are not optimal for pervasive signals collected from wearable devices - they don't adapt to individual difference and assume accurate identification of ECG fiducial points. To overcome these short-comings of the rule-based methods, this paper introduces an arrhythmia detection approach for low quality pervasive ECG signals. To achieve the robustness needed, two techniques were applied. First, a set of ECG features with minimal reliance on fiducial point identification were selected. Next, the features were normalized using robust statistics to factors out baseline individual differences and clinically irrelevant temporal drift that is common in pervasive ECG. The proposed method was evaluated using pervasive ECG signals we collected, in combination with clinician validated ECG signals from Physiobank. Empirical evaluation confirms accuracy improvements of the proposed approach over the traditional clinical rules. Refereed/Peer-reviewed
- Published
- 2014
29. Analysis of circadian rhythms from online communities of individuals with affective disorders
- Author
-
Bo Dao, Dinh Phung, Thin Nguyen, and Svetha Venkatesh
- Subjects
Rhythm ,Mood ,Cohort ,Data collecting ,Chronic stress ,Social media ,Circadian rhythm ,Psychology ,Affect (psychology) ,Developmental psychology - Abstract
The circadian system regulates 24 hour rhythms in biological creatures. It impacts mood regulation. The disruptions of circadian rhythms cause destabilization in individuals with affective disorders, such as depression and bipolar disorders. Previous work has examined the role of the circadian system on effects of light interactions on mood-related systems, the effects of light manipulation on brain, the impact of chronic stress on rhythms. However, such studies have been conducted in small, preselected populations. The deluge of data is now changing the landscape of research practice. The unprecedented growth of social media data allows one to study individual behavior across large and diverse populations. In particular, individuals with affective disorders from online communities have not been examined rigorously. In this paper, we aim to use social media as a sensor to identify circadian patterns for individuals with affective disorders in online communities.We use a large scale study cohort of data collecting from online affective disorder communities. We analyze changes in hourly, daily, weekly and seasonal affect of these clinical groups in contrast with control groups of general communities. By comparing the behaviors between the clinical groups and the control groups, our findings show that individuals with affective disorders show a significant distinction in their circadian rhythms across the online activity. The results shed light on the potential of using social media for identifying diurnal individual variation in affective state, providing key indicators and risk factors for noninvasive wellbeing monitoring and prediction.
- Published
- 2014
30. Nonparametric Discovery of Learning Patterns and Autism Subgroups from Therapeutic Data
- Author
-
Thi Duong, Dinh Phung, Pratibha Vellanki, and Svetha Venkatesh
- Subjects
Computer science ,business.industry ,media_common.quotation_subject ,Bayesian probability ,Nonparametric statistics ,Novelty ,medicine.disease ,Machine learning ,computer.software_genre ,Non-negative matrix factorization ,ComputingMethodologies_PATTERNRECOGNITION ,Autism spectrum disorder ,medicine ,Autism ,Unsupervised learning ,Artificial intelligence ,Cluster analysis ,Imitation ,business ,computer ,media_common - Abstract
Autism Spectrum Disorder (ASD) is growing at a staggering rate, but, little is known about the cause of this condition. Inferring learning patterns from therapeutic performance data, and subsequently clustering ASD children into subgroups, is important to understand this domain, and more importantly to inform evidence-based intervention. However, this data-driven task was difficult in the past due to insufficiency of data to perform reliable analysis. For the first time, using data from a recent application for early intervention in autism (TOBY Play pad), whose download count is now exceeding 4500, we present in this paper the automatic discovery of learning patterns across 32 skills in sensory, imitation and language. We use unsupervised learning methods for this task, but a notorious problem with existing methods is the correct specification of number of patterns in advance, which in our case is even more difficult due to complexity of the data. To this end, we appeal to recent Bayesian nonparametric methods, in particular the use of Bayesian Nonparametric Factor Analysis. This model uses Indian Buffet Process (IBP) as prior on a binary matrix of infinite columns to allocate groups of intervention skills to children. The optimal number of learning patterns as well as subgroup assignments are inferred automatically from data. Our experimental results follow an exploratory approach, present different newly discovered learning patterns. To provide quantitative results, we also report the clustering evaluation against K-means and Nonnegative matrix factorization (NMF). In addition to the novelty of this new problem, we were able to demonstrate the suitability of Bayesian nonparametric models over parametric rivals.
- Published
- 2014
31. Regularizing Topic Discovery in EMRs with Side Information by Using Hierarchical Bayesian Models
- Author
-
Cheng Li, Santu Rana, Dinh Phung, and Svetha Venkatesh
- Subjects
Structure (mathematical logic) ,Hierarchical Dirichlet process ,symbols.namesake ,Tree structure ,Computer science ,Bayesian probability ,symbols ,Inference ,Markov chain Monte Carlo ,Data mining ,Diagnosis code ,computer.software_genre ,computer - Abstract
We propose a novel hierarchical Bayesian framework, word-distance-dependent Chinese restaurant franchise (wd-dCRF) for topic discovery from a document corpus regularized by side information in the form of word-to-word relations, with an application on Electronic Medical Records (EMRs). Typically, a EMRs dataset consists of several patients (documents) and each patient contains many diagnosis codes (words). We exploit the side information available in the form of a semantic tree structure among the diagnosis codes for semantically-coherent disease topic discovery. We introduce novel functions to compute word-to-word distances when side information is available in the form of tree structures. We derive an efficient inference method for the wddCRF using MCMC technique. We evaluate on a real world medical dataset consisting of about 1000 patients with PolyVascular disease. Compared with the popular topic analysis tool, hierarchical Dirichlet process (HDP), our model discovers topics which are superior in terms of both qualitative and quantitative measures.
- Published
- 2014
32. A Bayesian Nonparametric Framework for Activity Recognition Using Accelerometer Data
- Author
-
Thuong Nguyen, Svetha Venkatesh, Dinh Phung, and Sunil Gupta
- Subjects
Hierarchical Dirichlet process ,Ground truth ,Computer science ,business.industry ,Feature extraction ,Machine learning ,computer.software_genre ,Bayesian nonparametrics ,Data modeling ,Activity recognition ,Data mining ,Artificial intelligence ,business ,computer - Abstract
Monitoring daily physical activity of human plays an important role in preventing diseases as well as improving health. In this paper, we demonstrate a framework for monitoring the physical activity levels in daily life. We collect the data using accelerometer sensors in a realistic setting without any supervision. The ground truth of activities is provided by the participants themselves using an experience sampling application running on mobile phones. The original data is discretized by the hierarchical Dirichlet process (HDP) into different activity levels and the number of levels is inferred automatically. We validate the accuracy of the extracted patterns by using them for the multi-label classification of activities and demonstrate the high performances in various standard evaluation metrics. We further show that the extracted patterns are highly correlated to the daily routine of users.
- Published
- 2014
33. Data-mining twitter and the autism spectrum disorder: A Pilot study
- Author
-
Adham Beykikhoshk, Ognjen Arandjelovic, Dinh Phung, Svetha Venkatesh, Terry Caelli, Beykikhoshk, Adham, Arandjelović, Ognjen, Phung, Dinh, Venkatesh, Svetha, Caelli, Terry, and 2014 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM 2014) Beijing, China 17-20 August 2014
- Abstract
The autism spectrum disorder (ASD) is increasingly being recognized as a major public health issue which affects approximately 0.5-0.6% of the population. Promoting the general awareness of the disorder, increasing the engagement with the affected individuals and their carers, and understanding the success of penetration of the current clinical recommendations in the target communities, is crucial in driving research as well as policy. The aim of the present work is to investigate if Twitter, as a highly popular platform for information exchange, can be used as a data-mining source which could aid in the aforementioned challenges. Specifically, using a large data set of harvested tweets, we present a series of experiments which examine a range of linguistic and semantic aspects of messages posted by individuals interested in ASD. Our findings, the first of their nature in the published scientific literature, strongly motivate additional research on this topic and present a methodological basis for further work. Refereed/Peer-reviewed
- Published
- 2014
34. Fixed-lag particle filter for continuous context discovery using Indian Buffet Process
- Author
-
Svetha Venkatesh, Thuong Nguyen, Sunil Gupta, and Dinh Phung
- Subjects
Exponential distribution ,Computer science ,Nonparametric statistics ,Inference ,Context (language use) ,computer.software_genre ,symbols.namesake ,Matrix (mathematics) ,symbols ,Data mining ,Particle filter ,Coefficient matrix ,computer ,Gibbs sampling - Abstract
Exploiting context from stream data in pervasive environments remains a challenge. We aim to extract proximal context from Bluetooth stream data, using an incremental, Bayesian nonparametric framework that estimates the number of contexts automatically. Unlike current approaches that can only provide final proximal grouping, our method provides proximal grouping and membership of users over time. Additionally, it provides an efficient online inference. We construct co-location matrix over time using Bluetooth data. A Poisson-exponential model is used to factorize this matrix into a factor matrix, interpreted as proximal groups, and a coefficient matrix that indicates factor usage. The coefficient matrix follows the Indian Buffet Process prior, which estimates the number of factors automatically. The non-negativity and sparsity of factors are enforced by using the exponential distribution to generate the factors. We propose a fixed-lag particle filter algorithm to process data incrementally. We compare the incremental inference (particle filter) with full batch inference (Gibbs sampling) in terms of normalized factorization error and execution time. The normalized error obtained through our incremental inference is comparable to that of full batch inference, whilst the execution time is more than 100 times faster. The discovered factors have similar meaning to the results of the popular Louvain method for community detection.
- Published
- 2014
35. Learning sparse latent representation and distance metric for image retrieval
- Author
-
Dinh Phung, Svetha Venkatesh, Truyen Tran, and Tu Dinh Nguyen
- Subjects
Similarity (geometry) ,Computer science ,business.industry ,Representation (systemics) ,Pattern recognition ,Machine learning ,computer.software_genre ,Feature (computer vision) ,Metric (mathematics) ,Learning to rank ,Visual Word ,Artificial intelligence ,business ,Distance transform ,computer ,Image retrieval - Abstract
The performance of image retrieval depends critically on the semantic representation and the distance function used to estimate the similarity of two images. A good representation should integrate multiple visual and textual (e.g., tag) features and offer a step closer to the true semantics of interest (e.g., concepts). As the distance function operates on the representation, they are interdependent, and thus should be addressed at the same time. We propose a probabilistic solution to learn both the representation from multiple feature types and modalities and the distance metric from data. The learning is regularised so that the learned representation and information-theoretic metric will (i) preserve the regularities of the visual/textual spaces, (ii) enhance structured sparsity, (iii) encourage small intra-concept distances, and (iv) keep inter-concept images separated. We demonstrate the capacity of our method on the NUS-WIDE data. For the well-studied 13 animal subset, our method outperforms state-of-the-art rivals. On the subset of single-concept images, we gain 79:5% improvement over the standard nearest neighbours approach on the MAP score, and 45.7% on the NDCG.
- Published
- 2013
36. Analysis of psycholinguistic processes and topics in online autism communities
- Author
-
Thin Nguyen, Dinh Phung, and Svetha Venkatesh
- Subjects
Computer science ,business.industry ,media_common.quotation_subject ,Pragmatics ,medicine.disease ,computer.software_genre ,Popularity ,Friendship ,Social support ,Autism spectrum disorder ,medicine ,Autism ,Social media ,Artificial intelligence ,Computational linguistics ,business ,computer ,Natural language processing ,media_common ,Cognitive psychology - Abstract
Current growth of individuals on the autism spectrum disorder (ASD) requires continuous support and care. With the popularity of social media, online communities of people affected by ASD emerge. This paper presents an analysis of these online communities through understanding aspects that differentiate such communities. In this paper, the aspects given are not expressed in terms of friendship, exchange of information, social support or recreation, but rather with regard to the topics and linguistic styles that people express in their on-line writing. Using data collected unobtrusively from LiveJournal, we analyze posts made by ten autism communities in conjunction with those made by a control group of standard communities. Significant differences have been found between autism and control communities when characterized by latent topics of discussion and psycholinguistic features. Latent topics are found to have greater predictive power than linguistic features when classifying blog posts as either autism or control community. This study suggests that data mining of online blogs has the potential to detect clinically meaningful data. It opens the door to possibilities including sentinel risk surveillance and harnessing the power in diverse large datasets.
- Published
- 2013
37. Exploiting side information in distance dependent Chinese restaurant processes for data clustering
- Author
-
Santu Rana, Svetha Venkatesh, Cheng Li, and Dinh Phung
- Subjects
Set (abstract data type) ,Structure (mathematical logic) ,Measure (data warehouse) ,Computer science ,Chinese restaurant process ,Mutual information ,Data mining ,Cluster analysis ,computer.software_genre ,computer - Abstract
Multimedia contents often possess weakly annotated data such as tags, links and interactions. The weakly annotated data is called side information. It is the auxiliary information of data and provides hints for exploring the link structure of data. Most clustering algorithms utilize pure data for clustering. A model that combines pure data and side information, such as images and tags, documents and keywords, can perform better at understanding the underlying structure of data. We demonstrate how to incorporate different types of side information into a recently proposed Bayesian nonparametric model, the distance dependent Chinese restaurant process (DD-CRP). Our algorithm embeds the affinity of this information into the decay function of the DD-CRP when side information is in the form of subsets of discrete labels. It is flexible to measure distance based on arbitrary side information instead of only the spatial layout or time stamp of observations. At the same time, for noisy and incomplete side information, we set the decay function so that the DD-CRP reduces to the traditional Chinese restaurant process, thus not inducing side effects of noisy and incomplete side information. Experimental evaluations on two real-world datasets NUS WIDE and 20 Newsgroups show exploiting side information in DD-CRP significantly improves the clustering performance.
- Published
- 2013
38. Interactive browsing system for anomaly video surveillance
- Author
-
Sunil Gupta, Svetha Venkatesh, Dinh Phung, and Tien-Vu Nguyen
- Subjects
Data set ,Computer science ,Interface (Java) ,Pattern recognition (psychology) ,Key (cryptography) ,Rare events ,Anomaly detection ,Data mining ,User interface ,computer.software_genre ,computer ,Motion (physics) - Abstract
Existing anomaly detection methods in video surveillance exhibit lack of congruence between rare events detected by algorithms and what is considered anomalous by users. This paper introduces a novel browsing model to address this issue, allowing users to interactively examine rare events in an intuitive manner. Introducing a novel way to compute rare motion patterns, we estimate latent factors of foreground motion patterns through Bayesian Nonparametric Factor analysis. Each factor corresponds to a typical motion pattern. A rarity score for each factor is computed, and ordered in decreasing order of rarity, permitting users to browse events using any proportion of rare factors. Rare events correspond to frames that contain the rare factors chosen. We present the user with an interface to inspect events that incorporate these rarest factors in a spatial-temporal manner. We demonstrate the system on a public video data set, showing key aspects of the browsing paradigm.
- Published
- 2013
39. Extraction of latent patterns and contexts from social honest signals using hierarchical Dirichlet processes
- Author
-
Svetha Venkatesh, Thuong Nguyen, Dinh Phung, and Sunil Gupta
- Subjects
Hierarchical Dirichlet process ,Ubiquitous computing ,business.industry ,Computer science ,Reality mining ,computer.software_genre ,Machine learning ,Data modeling ,Unsupervised learning ,Artificial intelligence ,Data mining ,business ,Raw data ,Cluster analysis ,computer ,Parametric statistics - Abstract
A fundamental task in pervasive computing is reliable acquisition of contexts from sensor data. This is crucial to the operation of smart pervasive systems and services so that they might behave efficiently and appropriately upon a given context. Simple forms of context can often be extracted directly from raw data. Equally important, or more, is the hidden context and pattern buried inside the data, which is more challenging to discover. Most of existing approaches borrow methods and techniques from machine learning, dominantly employ parametric unsupervised learning and clustering techniques. Being parametric, a severe drawback of these methods is the requirement to specify the number of latent patterns in advance. In this paper, we explore the use of Bayesian nonparametric methods, a recent data modelling framework in machine learning, to infer latent patterns from sensor data acquired in a pervasive setting. Under this formalism, nonparametric prior distributions are used for data generative process, and thus, they allow the number of latent patterns to be learned automatically and grow with the data - as more data comes in, the model complexity can grow to explain new and unseen patterns. In particular, we make use of the hierarchical Dirichlet processes (HDP) to infer atomic activities and interaction patterns from honest signals collected from sociometric badges. We show how data from these sensors can be represented and learned with HDP. We illustrate insights into atomic patterns learned by the model and use them to achieve high-performance clustering. We also demonstrate the framework on the popular Reality Mining dataset, illustrating the ability of the model to automatically infer typical social groups in this dataset. Finally, our framework is generic and applicable to a much wider range of problems in pervasive computing where one needs to infer high-level, latent patterns and contexts from sensor data.
- Published
- 2013
40. Sparse Subspace Representation for Spectral Document Clustering
- Author
-
Dinh Phung, Budhaditya Saha, Svetha Venkatesh, and Duc-Son Pham
- Subjects
Clustering high-dimensional data ,Fuzzy clustering ,K-SVD ,business.industry ,Computer Science::Information Retrieval ,Correlation clustering ,Pattern recognition ,Sparse approximation ,Document clustering ,computer.software_genre ,ComputingMethodologies_PATTERNRECOGNITION ,Affinity propagation ,Data mining ,Artificial intelligence ,Cluster analysis ,business ,computer ,Mathematics - Abstract
We present a novel method for document clustering using sparse representation of documents in conjunction with spectral clustering. An \ell_{1} - norm optimization formulation is posed to learn the sparse representation of each document, allowing us to characterize the affinity between documents by considering the overall information instead of traditional pair wise similarities. This document affinity is encoded through a graph on which spectral clustering is performed. The decomposition into multiple subspaces allows documents to be part of a sub-group that shares a smaller set of similar vocabulary, thus allowing for cleaner clusters. Extensive experimental evaluations on two real-world datasets from Reuters-21578 and 20Newsgroup corpora show that our proposed method consistently outperforms state-of-the-art algorithms. Significantly, the performance improvement over other methods is prominent for this datasets.
- Published
- 2012
41. Learning Boltzmann Distance Metric for Face Recognition
- Author
-
Dinh Phung, Svetha Venkatesh, and Truyen Tran
- Subjects
Restricted Boltzmann machine ,Computer science ,business.industry ,Feature extraction ,Boltzmann machine ,Pattern recognition ,Information theory ,Facial recognition system ,symbols.namesake ,Face space ,Face (geometry) ,Metric (mathematics) ,Boltzmann constant ,symbols ,Feature (machine learning) ,Computer vision ,Artificial intelligence ,business - Abstract
We introduce a new method for face recognition using a versatile probabilistic model known as Restricted Boltzmann Machine (RBM). In particular, we propose to regularise the standard data likelihood learning with an information-theoretic distance metric defined on intra-personal images. This results in an effective face representation which captures the regularities in the face space and minimises the intra-personal variations. In addition, our method allows easy incorporation of multiple feature sets with controllable level of sparsity. Our experiments on a high variation dataset show that the proposed method is competitive against other metric learning rivals. We also investigated the RBM method under a variety of settings, including fusing facial parts and utilising localised feature detectors under varying resolutions. In particular, the accuracy is boosted from 71.8% with the standard whole-face pixels to 99.2% with combination of facial parts, localised feature extractors and appropriate resolutions.
- Published
- 2012
42. Improved subspace clustering via exploitation of spatial constraints
- Author
-
Duc-Son Pham, Saha Budhaditya, Dinh Phung, and Svetha Venkatesh
- Subjects
Computer science ,business.industry ,Regression analysis ,Pattern recognition ,Sparse approximation ,Image segmentation ,Data set ,Kernel (linear algebra) ,Robustness (computer science) ,Embedding ,Segmentation ,Artificial intelligence ,Cluster analysis ,business ,Sparse matrix - Abstract
We present a novel approach to improving subspace clustering by exploiting the spatial constraints. The new method encourages the sparse solution to be consistent with the spatial geometry of the tracked points, by embedding weights into the sparse formulation. By doing so, we are able to correct sparse representations in a principled manner without introducing much additional computational cost. We discuss alternative ways to treat the missing and corrupted data using the latest theory in robust lasso regression and suggest numerical algorithms so solve the proposed formulation. The experiments on the benchmark Johns Hopkins 155 dataset demonstrate that exploiting spatial constraints significantly improves motion segmentation.
- Published
- 2012
43. Detection of Cross-Channel Anomalies from Multiple Data Channels
- Author
-
Budhaditya Saha, Svetha Venkatesh, Duc-Son Pham, and Dinh Phung
- Subjects
business.industry ,Computer science ,Cross channel ,Pattern recognition ,computer.software_genre ,Constant false alarm rate ,Support vector machine ,Multiple data ,Principal component analysis ,Anomaly detection ,Artificial intelligence ,Data mining ,business ,computer ,Communication channel - Abstract
We identify and formulate a novel problem: cross channel anomaly detection from multiple data channels. Cross channel anomalies are common amongst the individual channel anomalies, and are often portent of significant events. Using spectral approaches, we propose a two-stage detection method: anomaly detection at a single-channel level, followed by the detection of cross-channel anomalies from the amalgamation of single channel anomalies. Our mathematical analysis shows that our method is likely to reduce the false alarm rate. We demonstrate our method in two applications: document understanding with multiple text corpora, and detection of repeated anomalies in video surveillance. The experimental results consistently demonstrate the superior performance of our method compared with related state-of-art methods, including the one-class SVM and principal component pursuit. In addition, our framework can be deployed in a decentralized manner, lending itself for large scale data stream analysis.
- Published
- 2011
44. High accuracy context recovery using clustering mechanisms
- Author
-
Dinh Phung, Brett Adams, Kha Tran, Mohan Kumar, and Svetha Venkatesh
- Subjects
DBSCAN ,Ubiquitous computing ,Computer science ,business.industry ,Mobile computing ,Context (language use) ,computer.software_genre ,Machine learning ,Latent Dirichlet allocation ,symbols.namesake ,symbols ,Unsupervised learning ,Data mining ,Artificial intelligence ,Noise (video) ,business ,Cluster analysis ,computer - Abstract
This paper examines the recovery of user context in indoor environmnents with existing wireless infrastructures to enable assistive systems. We present a novel approach to the extraction of user context, casting the problem of context recovery as an unsupervised, clustering problem. A well known density-based clustering technique, DBSCAN, is adapted to recover user context that includes user motion state, and significant places the user visits from WiFi observations consisting of access point id and signal strength. Furthermore, user rhythms or sequences of places the user visits periodically are derived from the above low level contexts by employing a state-of-the-art probabilistic clustering technique, the Latent Dirichlet Allocation (LDA), to enable a variety of application services. Experimental results with real data are presented to validate the proposed unsupervised learning approach and demonstrate its applicability.
- Published
- 2009
45. A probabilistic model with parsinomious representation for sensor fusion in recognizing activity in pervasive environment
- Author
-
Svetha Venkatesh, Dinh Phung, D.T. Tran, and Hung Bui
- Subjects
Ubiquitous computing ,Computer science ,business.industry ,Softmax function ,Statistical model ,Pattern recognition ,State (computer science) ,Artificial intelligence ,Representation (mathematics) ,business ,Sensor fusion - Abstract
To tackle the problem of increasing numbers of state transition parameters when the number of sensors increases, we present a probabilistic model together with several parsinomious representations for sensor fusion. These include context specific independence (CSI), mixtures of smaller multinomials and softmax function representations to compactly represent the state transitions of a large number of sensors. The model is evaluated on real-world data acquired through ubiquitous sensors in recognizing daily morning activities. The results show that the combination of CSI and mixtures of smaller multinomials achieves comparable performance with much fewer parameters.
- Published
- 2006
46. Human Behavior Recognition with Generic Exponential Family Duration Modeling in the Hidden Semi-Markov Model
- Author
-
Svetha Venkatesh, Dinh Phung, Thi Duong, and Hung Bui
- Subjects
Computer science ,business.industry ,Inference ,Poisson distribution ,Machine learning ,computer.software_genre ,Inverse Gaussian distribution ,symbols.namesake ,Exponential family ,Pattern recognition (psychology) ,symbols ,Probability distribution ,Multinomial distribution ,Hidden semi-Markov model ,Artificial intelligence ,Duration (project management) ,Hidden Markov model ,business ,computer ,Dynamic Bayesian network - Abstract
The ability to learn and recognize human activities of daily living (ADLs) is important in building pervasive and smart environments. In this paper, we tackle this problem using the hidden semi-Markov model. We discuss the stateof- the-art duration modeling choices and then address a large class of exponential family distributions to model state durations. Inference and learning are efficiently addressed by providing a graphical representation for the model in terms of a dynamic Bayesian network (DBN). We investigate both discrete and continuous distributions from the exponential family (Poisson and Inverse Gaussian respectively) for the problem of learning and recognizing ADLs. A full comparison between the exponential family duration models and other existing models including the traditional multinomial and the new Coxian are also presented. Our work thus completes a thorough investigation into the aspect of duration modeling and its application to human activities recognition in a real-world smart home surveillance scenario.
- Published
- 2006
47. Learning and Detecting Activities from Movement Trajectories Using the Hierarchical Hidden Markov Models
- Author
-
Dinh Phung, N.T. Nguyen, Svetha Venkatesh, and Hung Bui
- Subjects
business.industry ,Stochastic process ,Stochastic modelling ,Estimation theory ,Computer science ,Hierarchical hidden Markov model ,Pattern recognition ,Machine learning ,computer.software_genre ,Activity recognition ,Approximate inference ,Tree (data structure) ,Robustness (computer science) ,Artificial intelligence ,business ,Particle filter ,Hidden Markov model ,computer ,Time complexity - Abstract
Directly modeling the inherent hierarchy and shared structures of human behaviors, we present an application of the hierarchical hidden Markov model (HHMM) for the problem of activity recognition. We argue that to robustly model and recognize complex human activities, it is crucial to exploit both the natural hierarchical decomposition and shared semantics embedded in the movement trajectories. To this end, we propose the use of the HHMM, a rich stochastic model that has been recently extended to handle shared structures, for representing and recognizing a set of complex indoor activities. Furthermore, in the need of real-time recognition, we propose a Rao-Blackwellised particle filter (RBPF) that efficiently computes the filtering distribution at a constant time complexity for each new observation arrival. The main contributions of this paper lie in the application of the shared-structure HHMM, the estimation of the model's parameters at all levels simultaneously, and a construction of an RBPF approximate inference scheme. The experimental results in a real-world environment have confirmed our belief that directly modeling shared structures not only reduces computational cost, but also improves recognition accuracy when compared with the tree HHMM and the flat HMM.
- Published
- 2005
48. Activity Recognition and Abnormality Detection with the Switching Hidden Semi-Markov Model
- Author
-
Svetha Venkatesh, Thi Duong, Hung Bui, and Dinh Phung
- Subjects
Sequence ,Training set ,Computer science ,business.industry ,Hierarchical hidden Markov model ,Machine learning ,computer.software_genre ,Generalization error ,Activity recognition ,Multinomial distribution ,Artificial intelligence ,Hidden semi-Markov model ,Duration (project management) ,Hidden Markov model ,business ,computer - Abstract
This paper addresses the problem of learning and recognizing human activities of daily living (ADL), which is an important research issue in building a pervasive and smart environment. In dealing with ADL, we argue that it is beneficial to exploit both the inherent hierarchical organization of the activities and their typical duration. To this end, we introduce the switching hidden semi-markov model (S-HSMM), a two-layered extension of the hidden semi-Markov model (HSMM) for the modeling task. Activities are modeled in the S-HSMM in two ways: the bottom layer represents atomic activities and their duration using HSMMs; the top layer represents a sequence of high-level activities where each high-level activity is made of a sequence of atomic activities. We consider two methods for modeling duration: the classic explicit duration model using multinomial distribution, and the novel use of the discrete Coxian distribution. In addition, we propose an effective scheme to detect abnormality without the need for training on abnormal data. Experimental results show that the S-HSMM performs better than existing models including the flat HSMM and the hierarchical hidden Markov model in both classification and abnormality detection tasks, alleviating the need for presegmented training data. Furthermore, our discrete Coxian duration model yields better computation time and generalization error than the classic explicit duration model.
- Published
- 2005
49. Automatically learning structural units in educational videos with the hierarchical hidden markov models
- Author
-
Hung Bui, Dinh Phung, and Svetha Venkatesh
- Subjects
Structure (mathematical logic) ,Hierarchy (mathematics) ,business.industry ,Computer science ,Artificial intelligence ,business ,Hidden Markov model ,Machine learning ,computer.software_genre ,computer ,Abstraction (linguistics) - Abstract
In this paper we present a coherent approach using the hierarchical HMM with shared structures to extract the structural units form the building blocks of an education/training video. Rather than using hand-crafted approaches to define the structural units, we use the data from nine training videos to learn the parameters of the HHMM, and thus naturally extract the hierarchy. We then study this hierarchy and examine the nature of the structure at different levels of abstraction. Since the observable is continuous, we also show how to extend the parameter learning in the HHMM to deal with continuous observations.
- Published
- 2005
50. Efficient Coxian Duration Modelling for Activity Recognition in Smart Environments with the Hidden semi-Markov Model
- Author
-
H. H. Bui, Thi Duong, Svetha Venkatesh, and Dinh Phung
- Subjects
Stochastic process ,Stochastic modelling ,business.industry ,Computer science ,Machine learning ,computer.software_genre ,Activity recognition ,Smart environment ,Multinomial distribution ,Hidden semi-Markov model ,Artificial intelligence ,Hidden Markov model ,business ,computer ,Free parameter - Abstract
In this paper, we exploit the discrete Coxian distribution and propose a novel form of stochastic model, termed as the Coxian hidden semi-Makov model (Cox-HSMM), and apply it to the task of recognising activities of daily living (ADLs) in a smart house environment. The use of the Coxian has several advantages over traditional parameterization (e.g. multinomial or continuous distributions) including the low number of free parameters needed, its computational efficiency, and the existing of closed-form solution. To further enrich the model in real-world applications, we also address the problem of handling missing observation for the proposed Cox-HSMM. In the domain of ADLs, we emphasize the importance of the duration information and model it via the Cox-HSMM. Our experimental results have shown the superiority of the Cox-HSMM in all cases when compared with the standard HMM. Our results have further shown that outstanding recognition accuracy can be achieved with relatively low number of phases required in the Coxian, thus making the Cox-HSMM particularly suitable in recognizing ADLs whose movement trajectories are typically very long in nature.
- Published
- 2005
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.