9,055 results on '"TASK analysis"'
Search Results
2. Collusion-Resistant Worker Recruitment in Crowdsourcing Systems
- Author
-
Ming Li, Arun Thapa, Pan Li, Lei Yang, Wenqiang Jin, and Mingyan Xiao
- Subjects
Computer Networks and Communications ,Computer science ,business.industry ,Internet privacy ,Collusion ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,020206 networking & telecommunications ,02 engineering and technology ,Electrical and Electronic Engineering ,Crowdsourcing ,business ,Software - Published
- 2023
3. Multi-Label Emotion Detection via Emotion-Specified Feature Extraction and Emotion Correlation Learning
- Author
-
Jiawen Deng and Fuji Ren
- Subjects
Emotion Correlation ,0209 industrial biotechnology ,Context model ,Computer science ,Speech recognition ,Sentiment analysis ,Feature extraction ,02 engineering and technology ,Task (project management) ,Emotion Detection ,Human-Computer Interaction ,Correlation ,chemistry.chemical_compound ,Multi-label Focal Loss ,020901 industrial engineering & automation ,chemistry ,Feature (computer vision) ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Multi-label ,020201 artificial intelligence & image processing ,Software ,MEDA - Abstract
Textual emotion detection is an attractive task while previous studies mainly focused on polarity or single-emotion classification. However, human expressions are complex, and multiple emotions often occur simultaneously with non-negligible emotion correlations. In this paper, a Multi-label Emotion Detection Architecture (MEDA) is proposed to detect all associated emotions expressed in a given piece of text. MEDA is mainly composed of two modules: Multi-Channel Emotion-Specified Feature Extractor (MC-ESFE) and Emotion Correlation Learner (ECorL). MEDA captures underlying emotion-specified features through MC-ESFE module in advance. MC-ESFE is composed of multiple channel-wise ESFE networks. Each channel is devoted to the feature extraction of a specified emotion from sentence-level to context-level through a hierarchical structure. Based on obtained features, emotion correlation learning is implemented through an emotion sequence predictor in ECorL. During model training, we define a new loss function, which is called multi-label focal loss. With this loss function, the model can focus more on misclassified positive-negative emotion pairs and improve the overall performance by balancing the prediction of positive and negative emotions. The evaluation of proposed MEDA architecture is carried out on emotional corpus: RenCECps and NLPCC2018 datasets. The experimental results indicate that the proposed method can achieve better performance than state-of-the-art methods in this task.
- Published
- 2023
4. UrbanRama: Navigating Cities in Virtual Reality
- Author
-
Michael Koutsoubis, Nivan Ferreira, Corinne Brenner, Ken Perlin, Luc Wilson, Cláudio T. Silva, Harish Doraiswamy, Shaoyu Chen, Fábio Rodrigues de Miranda, Marcos Lage, and Connor DeFanti
- Subjects
FOS: Computer and information sciences ,Point of interest ,Computer science ,Interface (computing) ,Computer Science - Human-Computer Interaction ,Context (language use) ,02 engineering and technology ,Virtual reality ,Human-Computer Interaction (cs.HC) ,Task (project management) ,User-Computer Interface ,Computer Science - Graphics ,Human–computer interaction ,Urban planning ,11. Sustainability ,Computer Graphics ,0202 electrical engineering, electronic engineering, information engineering ,Cities ,Perspective (graphical) ,Virtual Reality ,020207 software engineering ,Computer Graphics and Computer-Aided Design ,Graphics (cs.GR) ,Signal Processing ,Task analysis ,Computer Vision and Pattern Recognition ,Software - Abstract
Exploring large virtual environments, such as cities, is a central task in several domains, such as gaming and urban planning. VR systems can greatly help this task by providing an immersive experience; however, a common issue with viewing and navigating a city in the traditional sense is that users can either obtain a local or a global view, but not both at the same time, requiring them to continuously switch between perspectives, losing context and distracting them from their analysis. In this paper, our goal is to allow users to navigate to points of interest without changing perspectives. To accomplish this, we design an intuitive navigation interface that takes advantage of the strong sense of spatial presence provided by VR. We supplement this interface with a perspective that warps the environment, called UrbanRama, based on a cylindrical projection, providing a mix of local and global views. The design of this interface was performed as an iterative process in collaboration with architects and urban planners. We conducted a qualitative and a quantitative pilot user study to evaluate UrbanRama and the results indicate the effectiveness of our system in reducing perspective changes, while ensuring that the warping doesn't affect distance and orientation perception., Comment: Video: https://www.youtube.com/watch?v=M8BFZnxq-Qg
- Published
- 2022
5. Point-of-Interest Recommendation With Global and Local Context
- Author
-
Aixin Sun, Kai Zheng, Shuo Shang, Xiangliang Zhang, Peng Han, and Peilin Zhao
- Subjects
Context model ,Information retrieval ,Point of interest ,Computer science ,Context (language use) ,02 engineering and technology ,Computer Science Applications ,Task (project management) ,Matrix decomposition ,Computational Theory and Mathematics ,020204 information systems ,Similarity (psychology) ,0202 electrical engineering, electronic engineering, information engineering ,Collaborative filtering ,Task analysis ,Information Systems - Abstract
The task of point of interest (POI) recommendation aims to recommend unvisited places to users based on their check-in history. A major challenge in POI recommendation is data sparsity, because a user typically visits only a very small number of POIs among all available POIs. In this paper, we propose AUC-MF to address the POI recommendation problem by maximizing Area Under the ROC curve (AUC). AUC has been widely used for measuring classification performance with imbalanced data distributions. To optimize AUC, we transform the recommendation task to a classification problem, where the visited locations are positive examples and the unvisited are negative ones. We define a new lambda for AUC to utilize the LambdaMF model, which combines the lambda-based method and matrix factorization model in collaborative filtering. Many studies have shown that geographic information plays an important role in POI recommendation. In this study, we focus on two levels geographic information: local similarity and global similarity. We further show that AUC-MF can be easily extended to incorporate geographical contextual information for POI recommendation.
- Published
- 2022
6. Unsupervised Person Re-Identification via Multi-Label Classification
- Author
-
Dongkai Wang and Shiliang Zhang
- Subjects
Multi-label classification ,FOS: Computer and information sciences ,Computer science ,business.industry ,Computer Vision and Pattern Recognition (cs.CV) ,Computer Science - Computer Vision and Pattern Recognition ,Pattern recognition ,02 engineering and technology ,010501 environmental sciences ,01 natural sciences ,Re identification ,ComputingMethodologies_PATTERNRECOGNITION ,Discriminative model ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,020201 artificial intelligence & image processing ,Artificial intelligence ,Computer Vision and Pattern Recognition ,Transfer of learning ,business ,Classifier (UML) ,Software ,0105 earth and related environmental sciences - Abstract
The challenge of unsupervised person re-identification (ReID) lies in learning discriminative features without true labels. This paper formulates unsupervised person ReID as a multi-label classification task to progressively seek true labels. Our method starts by assigning each person image with a single-class label, then evolves to multi-label classification by leveraging the updated ReID model for label prediction. The label prediction comprises similarity computation and cycle consistency to ensure the quality of predicted labels. To boost the ReID model training efficiency in multi-label classification, we further propose the memory-based multi-label classification loss (MMCL). MMCL works with memory-based non-parametric classifier and integrates multi-label classification and single-label classification in a unified framework. Our label prediction and MMCL work iteratively and substantially boost the ReID performance. Experiments on several large-scale person ReID datasets demonstrate the superiority of our method in unsupervised person ReID. Our method also allows to use labeled person images in other domains. Under this transfer learning setting, our method also achieves state-of-the-art performance., Comment: CVPR2020
- Published
- 2022
7. Entity Alignment for Knowledge Graphs With Multi-Order Convolutional Networks
- Author
-
Bolong Zheng, Thanh Trung Huynh, Vinh Van Tong, Tam Thanh Nguyen, Quoc Viet Hung Nguyen, Darnbi Sakong, and Hongzhi Yin
- Subjects
Theoretical computer science ,Computational Theory and Mathematics ,Knowledge graph ,Computer science ,020204 information systems ,Knowledge engineering ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,02 engineering and technology ,Computer Science Applications ,Information Systems ,Data modeling - Abstract
Knowledge graphs (KGs) have become popular structures for unifying real-world entities by modelling the relationships between them and their attributes. Entity alignment -- the task of identifying corresponding entities across different KGs -- has attracted a great deal of attention in both academia and industry. However, existing alignment techniques often require large amounts of labelled data, are unable to encode multi-modal data simultaneously, and enforce only few consistency constraints. In this paper, we propose an end-to-end, unsupervised entity alignment framework for cross-lingual KGs that fuses different types of information in order to fully exploit the richness of KG data. The model captures the relation-based correlation between entities by using a multi-order graph convolutional neural (GCN) model that is designed to satisfy the consistency constraints, while incorporating the attribute-based correlation via a translation machine. We adopt a late-fusion mechanism to combine all the information together, which allows these approaches to complement each other and thus enhances the final alignment result, and makes the model more robust to consistency violations. Empirical results show that our model is more accurate and orders of magnitude faster than existing baselines. We also demonstrate its sensitivity to hyper-parameters, effort saving in terms of labelling, and the robustness against adversarial conditions.
- Published
- 2022
8. Incentive Mechanism Design for Truth Discovery in Crowdsourcing With Copiers
- Author
-
Lijie Xu, Jia Xu, Dejun Yang, Xiaofu Niu, and Lingyun Jiang
- Subjects
Mechanism design ,Information Systems and Management ,Computer Networks and Communications ,business.industry ,Computer science ,TheoryofComputation_GENERAL ,020206 networking & telecommunications ,02 engineering and technology ,Machine learning ,computer.software_genre ,Semantics ,Crowdsourcing ,Computer Science Applications ,Reverse auction ,Knowledge extraction ,Hardware and Architecture ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,020201 artificial intelligence & image processing ,Artificial intelligence ,Greedy algorithm ,business ,Cluster analysis ,computer - Abstract
Crowdsourcing has become an effective tool to utilize human intelligence to perform tasks that are challenging for machines. Many truth discovery methods and incentive mechanisms for crowdsourcing have been proposed. However, most of them cannot deal with the crowdsourcing with copiers, who copy a part (or all) of data from other workers. This paper aims at designing crowdsourcing incentive mechanism for truth discovery of textual answers with copiers. We formulate the problem of maximizing the social welfare such that all tasks can be completed with the least confidence for truth discovery and design an three-stage incentive mechanism. In contextual embedding and clustering stage, we construct and cluster the content vector representations of textual crowdsourced answers at the semantic level. In truth discovery stage, we estimate the truth for each task based on the dependence and accuracy of workers. In reverse auction stage, we design a greedy algorithm to select the winners and determine the payment. Through both rigorous theoretical analysis and extensive simulations, we demonstrate that the proposed mechanisms achieve computational efficiency, individual rationality, truthfulness, and guaranteed approximation. Moreover, our truth discovery methods show prominent advantage in terms of precision when there are copiers in the crowdsourcing systems.
- Published
- 2022
9. Few-Shot Named Entity Recognition via Meta-Learning
- Author
-
Hao Wang, Jing Li, Billy Chiu, and Shanshan Feng
- Subjects
Meta learning (computer science) ,Contextual image classification ,Computer science ,business.industry ,02 engineering and technology ,Overfitting ,computer.software_genre ,Relationship extraction ,Sequence labeling ,Computer Science Applications ,Task (project management) ,Computational Theory and Mathematics ,Named-entity recognition ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Artificial intelligence ,business ,computer ,Natural language processing ,Information Systems - Abstract
Few-shot learning under the N-way K-shot setting (i.e., K annotated samples for each of N classes) has been widely studied in relation extraction (e.g., FewRel) and image classification (e.g., Mini-ImageNet). Named entity recognition (NER) is typically framed as a sequence labeling problem where the entity classes are inherently entangled together because the entity number and classes in a sentence are not known in advance, leaving the N-way K-shot NER problem so far unexplored. In this paper, we first formally define a more suitable N-way K-shot setting for NER. Then we propose FewNER, a novel meta-learning approach for few-shot NER. FewNER separates the entire network into a task-independent part and a task-specific part. During training in FewNER, the task-independent part is meta-learned across multiple tasks and a task-specific part is learned for each single task in a low-dimensional space. At test time, FewNER keeps the task-independent part fixed and adapts to a new task via gradient descent by updating only the task-specific part, resulting in it being less prone to overfitting and more computationally efficient. The results demonstrate that FewNER achieves state-of-the-art performance against nine baseline methods by significant margins on three adaptation experiments.
- Published
- 2022
10. Joint Representation Learning and Clustering: A Framework for Grouping Partial Multiview Data
- Author
-
Wenzhang Zhuge, Chenping Hou, Hong Tao, Dongyun Yi, Tingjin Luo, and Ling-Li Zeng
- Subjects
Optimization problem ,Theoretical computer science ,Computer science ,Iterative method ,02 engineering and technology ,Computer Science Applications ,Matrix (mathematics) ,Computational Theory and Mathematics ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Embedding ,Graph (abstract data type) ,Cluster analysis ,Feature learning ,Information Systems - Abstract
Partial multi-view clustering has attracted various attentions from diverse fields. Most existing methods adopt separate steps to obtain unified representations and extract clustering indicators. This separate manner prevents two learning processes to negotiate to achieve optimal performance. In this paper, we propose the Joint Representation Learning and Clustering (JRLC) framework to address this issue. The JRLC framework employs representation matrices to extract view-specific clustering information directly from the presence of partial similarity matrices, and rotates them to learn a common probability label matrix simultaneously, which connects representation learning and clustering seamlessly to achieve better clustering performance. Under the guidance of JRLC framework, several new incomplete multi-view clustering methods can be developed by extending existing single-view graph-based representation learning methods. For illustration, within the framework, we propose two specific methods, JRLC with spectral embedding (JRLC-SE) and JRLC via integrating nonnegative embedding and spectral embedding (JRLC-NS). Two iterative algorithms with guaranteed convergence are designed to solve the resultant optimization problems of JRLC-SE and JRLC-NS. Experimental results on various datasets and news topic clustering application demonstrate the effectiveness of the proposed algorithms.
- Published
- 2022
11. Adapted Dynamic Memory Network for Emotion Recognition in Conversation
- Author
-
Sijie Mai, Haifeng Hu, and Songlong Xing
- Subjects
Dependency (UML) ,Computer science ,media_common.quotation_subject ,Speech recognition ,Process (computing) ,020206 networking & telecommunications ,Context (language use) ,02 engineering and technology ,Human-Computer Interaction ,0202 electrical engineering, electronic engineering, information engineering ,Benchmark (computing) ,Task analysis ,020201 artificial intelligence & image processing ,Conversation ,Representation (mathematics) ,Episodic memory ,Software ,media_common - Abstract
In this paper, we address Emotion Recognition in Conversation (ERC) where conversational data are presented in a multimodal setting. Psychological evidence shows that self and inter-speaker influence are two central factors to emotion dynamics in conversation. State-of-the-art models do not effectively synthesise these two factors. Therefore, we propose an Adapted Dynamic Memory Network (A-DMN) where self and inter-speaker influences are modelled individually and further synthesised oriented towards the current utterance. Specifically, we model the dependency of the constituent utterances in a dialogue video using a global RNN to capture inter-speaker influence. Likewise, each speaker is assigned an RNN to capture their self influence. Afterwards, an Episodic Memory Module is devised to extract contexts for self and inter-speaker influence and synthesise them to update the memory. This process repeats itself for multiple passes until a refined representation is obtained and used for final prediction. Additionally, we explore cross-modal fusion in the context of multimodal ERC, and propose a convolution-based method which proves effective in extracting local interactions and computationally efficient. Extensive experiments demonstrate that A-DMN outperforms the state-of-the-art models on benchmark datasets.
- Published
- 2022
12. Evolutionary Multi-Objective Workflow Scheduling for Volatile Resources in the Cloud
- Author
-
Thanh-Phuong Pham, Thomas Fahringer, Department of Computer Science, University of Innsbruck, Aalto-yliopisto, and Aalto University
- Subjects
0209 industrial biotechnology ,Job shop scheduling ,Computer Networks and Communications ,Computer science ,business.industry ,Distributed computing ,Evolutionary algorithm ,Cloud computing ,02 engineering and technology ,Computer Science Applications ,Scheduling (computing) ,Set (abstract data type) ,020901 industrial engineering & automation ,Workflow ,Hardware and Architecture ,Economic cost ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,020201 artificial intelligence & image processing ,Workflow Scheduling , Workflow Makespan and Cost , Evolutionary Algorithm , Spot Instance ,business ,Software ,Information Systems - Abstract
The cloud has been widely used as a distributed computing platform for running scientific workflow applications. Most of the cloud providers encourage the use of their underutilized resources as spot instances for much cheaper prices compared with common resources as on-demand instances, however, the promise of lower costs for resources results in the volatility such that spot instances can be interrupted at any time by cloud providers. Many workflow scheduling algorithms have been proposed to deal with volatile resources. In this work, we consider the two most important features of the volatile resources namely fulfillment and interruption rates to fully model the instability of the cloud infrastructure. Subsequently, we propose a novel evolutionary multi-objective workflow scheduling approach to generate a set of trade-off solutions that outperform state-of-the-art algorithms both with respect to makespan and economic costs. In addition, we explore the fluctuation of makespan and costs for our obtained schedules under different levels of fulfillment and interruption rates. Experimental results with the five well-known real-world workflows demonstrate that our evolutionary multi-objective workflow scheduling algorithm is competitive in terms of makespan and cost compared with state-of-the-art on-demand scheduling techniques.
- Published
- 2022
13. Affine Transformation-Enhanced Multifactorial Optimization for Heterogeneous Problems
- Author
-
Kai Zhang, Xinggang Zhao, Guodong Chen, Liming Zhang, Jun Yao, Liang Feng, Jian Wang, Xiaoming Xue, and Kay Chen Tan
- Subjects
Domain adaptation ,Boosting (machine learning) ,Theoretical computer science ,Fitness landscape ,020209 energy ,0211 other engineering and technologies ,02 engineering and technology ,Evolutionary computation ,Computer Science Applications ,Human-Computer Interaction ,Control and Systems Engineering ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Human multitasking ,Affine transformation ,Electrical and Electronic Engineering ,Knowledge transfer ,Software ,021106 design practice & management ,Information Systems - Abstract
Evolutionary multitasking (EMT) is a newly emerging research topic in the community of evolutionary computation, which aims to improve the convergence characteristic across multiple distinct optimization tasks simultaneously by triggering knowledge transfer among them. Unfortunately, most of the existing EMT algorithms are only capable of boosting the optimization performance for homogeneous problems which explicitly share the same (or similar) fitness landscapes. Seldom efforts have been devoted to generalize the EMT for solving heterogeneous problems. A few preliminary studies employ domain adaptation techniques to enhance the transferability between two distinct tasks. However, almost all of these methods encounter a severe issue which is the so-called degradation of intertask mapping. Keeping this in mind, a novel rank loss function for acquiring a superior intertask mapping is proposed in this article. In particular, with an evolutionary-path-based representation model for optimization instance, an analytical solution of affine transformation for bridging the gap between two distinct problems is mathematically derived from the proposed rank loss function. It is worth mentioning that the proposed mapping-based transferability enhancement technique can be seamlessly embedded into an EMT paradigm. Finally, the efficacy of our proposed method against several state-of-the-art EMTs is verified experimentally on a number of synthetic multitasking and many-tasking benchmark problems, as well as a practical case study.
- Published
- 2022
14. Latency and Alphabet Size in the Context of Multicast Network Coding
- Author
-
Alex Sprintson, Michael Langberg, and Mira Gone
- Subjects
Discrete mathematics ,Network packet ,Computer science ,020206 networking & telecommunications ,0102 computer and information sciences ,02 engineering and technology ,Library and Information Sciences ,Network topology ,01 natural sciences ,Telecommunications network ,Computer Science Applications ,010201 computation theory & mathematics ,Bounded function ,Linear network coding ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Latency (engineering) ,Coding (social sciences) ,Information Systems - Abstract
We study the relation between latency and alphabet size in the context of Multicast Network Coding. Given a graph $G =$ (${V}$, E) representing a communication network, a subset $S \subseteq V$ of sources, each of which initially holds a set of information messages, and a set $ T\subseteq V$ of terminals; we consider the problem in which one wishes to design a communication scheme that eventually allows all terminals to obtain all the messages held by the sources. In this study we assume that communication is performed in rounds, where in each round each network node may transmit a single (possibly encoded) information packet on any of its outgoing edges. The objective is to minimize the communication latency, i.e., number of communication rounds needed until all terminals have all the messages of the source nodes.For sufficiently large alphabet sizes (i.e., large block length, packet sizes), it is known that traditional linear multicast network coding techniques (such as random linear network coding)) minimize latency. In this work we seek to study the task of minimizing latency in the setting of limited alphabet sizes $(\mathrm {i}.\mathrm {e}.,$ finite block length), and alternatively, the task of minimizing the alphabet size in the setting of bounded latency. Through reductive arguments, we prove that it is NP-hard to (i) approximate (and in particular to determine) the minimum alphabet size given a latency constraint; (ii) to approximate (and in particular to determine) the minimum latency of communication schemes in the setting of limited alphabet sizes.
- Published
- 2022
15. B4SDC: A Blockchain System for Security Data Collection in MANETs
- Author
-
Xiaokang Zhou, Zheng Yan, Huidong Dong, Shohei Shimizu, Gao Liu, Department of Communications and Networking, Xidian University, Shiga University, Aalto-yliopisto, and Aalto University
- Subjects
blockchain ,Ad hoc networks ,Information Systems and Management ,Spoofing attack ,Computer science ,Wireless ad hoc network ,MANETs ,Mobile computing ,02 engineering and technology ,Digital signature ,0202 electrical engineering, electronic engineering, information engineering ,Data collection ,business.industry ,Node (networking) ,020206 networking & telecommunications ,Mobile ad hoc network ,security-related data collection ,Task analysis ,Collusion ,incentive mechanism ,020201 artificial intelligence & image processing ,business ,Bitcoin ,Information Systems ,Computer network - Abstract
Security-related data collection is an essential part for attack detection and security measurement in Mobile Ad Hoc Networks (MANETs). A detection node (i.e., collector) should discover available routes to a collection node for data collection and collect security-related data during route discovery for determining reliable routes. However, few studies provide incentives for security-related data collection in MANETs. In this paper, we propose B4SDC, a blockchain system for security-related data collection in MANETs. Through controlling the scale of Route REQuest (RREQ) forwarding in route discovery, the collector can constrain its payment and simultaneously make each forwarder of control information (namely RREQs and Route REPlies, in short RREPs) obtain rewards as much as possible to ensure fairness. At the same time, B4SDC avoids collusion attacks with cooperative receipt reporting, and spoofing attacks by adopting a secure digital signature. Based on a novel Proof-of-Stake consensus mechanism by accumulating stakes through message forwarding, B4SDC not only provides incentives for all participating nodes, but also avoids forking and ensures high efficiency and real decentralization. We analyze B4SDC in terms of incentives and security, and evaluate its performance through simulations. The thorough analysis and experimental results show the efficacy and effectiveness of B4SDC.
- Published
- 2022
16. Migration-Aware Network Services With Edge Computing
- Author
-
Atri Mukhopadhyay, George Iosifidis, and Marco Ruffini
- Subjects
generalized assignment problem ,Computer Networks and Communications ,Markov processes ,Resource management ,learning automata ,markov decision process ,Servers ,020206 networking & telecommunications ,02 engineering and technology ,service migration ,multi-access edge computing ,Minimization ,Costs ,Task analysis ,0202 electrical engineering, electronic engineering, information engineering ,Electrical and Electronic Engineering ,Passive optical networks - Abstract
—The development of Multi-access edge computing (MEC) has resulted from the requirement for supporting next generation mobile services, which need high capacity, high reliability and low latency. The key issue in such MEC architectures is to decide which edge nodes will be employed for serving the needs of the different end users. Here, we take a fresh look into this problem by focusing on the minimization of migration events rather than focusing on maximizing usage of resources. This is important because service migrations can create significant service downtime to applications that need low latency and high reliability, in addition to increasing traffic congestion in the underlying network. This paper introduces a priority induced service migration minimization (PrISMM) algorithm, which aims at minimizing service migration for both high and low priority services, through the use of Markov decision process, learning automata and combinatorial optimization. We carry out extensive simulations and produce results showing its effectiveness in reducing the mean service downtime of lower priority services and the mean admission time of the higher priority services.
- Published
- 2022
17. Efficient and effective multi-modal queries through heterogeneous network embedding
- Author
-
Quoc Viet Hung Nguyen, Son T. Mai, Chi Thang Duong, Karl Aberer, Tam Thanh Nguyen, Matthias Weidlich, and Hongzhi Yin
- Subjects
fusion ,Computer science ,Graph embedding ,videos ,Information needs ,02 engineering and technology ,heterogeneous networks ,Data modeling ,information ,Set (abstract data type) ,query embedding ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Relevance (information retrieval) ,information retrieval ,semantics ,retrieval ,games ,graph embedding ,Information retrieval ,heterogeneous information network ,Computer Science Applications ,data models ,Computational Theory and Mathematics ,task analysis ,Embedding ,Heterogeneous network ,Information Systems - Abstract
The heterogeneity of today’s Web sources requires information retrieval (IR) systems to handle multi-modal queries. Such queries define a user’s information needs by different data modalities, such as keywords, hashtags, user profiles, and other media. Recent IR systems answer such a multi-modal query by considering it as a set of separate uni-modal queries. However, depending on the chosen operationalisation, such an approach is inefficient or ineffective. It either requires multiple passes over the data or leads to inaccuracies since the relations between data modalities are neglected in the relevance assessment. To mitigate these challenges, we present an IR system that has been designed to answer genuine multi-modal queries. It relies on a heterogeneous network embedding, so that features from diverse modalities can be incorporated when representing both, a query and the data over which it shall be evaluated. By embedding a query and the data in the same vector space, the relations across modalities are made explicit and exploited for more accurate query evaluation. At the same time, multi-modal queries are answered with a single pass over the data. An experimental evaluation using diverse real-world and synthetic datasets illustrates that our approach returns twice the amount of relevant information compared to baseline techniques, while scaling to large multi-modal databases.
- Published
- 2023
18. Role-Based Graph Embeddings
- Author
-
Hoda Eldardiry, Xiangnan Kong, Ryan A. Rossi, Rong Zhou, Theodore L. Willke, Nesreen K. Ahmed, and John Boaz Lee
- Subjects
Context model ,Theoretical computer science ,Computer science ,02 engineering and technology ,Random walk ,Graph ,Computer Science Applications ,Vertex (geometry) ,Computational Theory and Mathematics ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Embedding ,Information Systems - Abstract
Random walks are at the heart of many existing node embedding and network representation learning methods. However, such methods have many limitations that arise from the use of traditional random walks, e.g., the embeddings resulting from these methods capture proximity (communities) among the vertices as opposed to structural similarity (roles). Furthermore, the embeddings are unable to transfer to new nodes and graphs as they are tied to node identity. To overcome these limitations, we introduce the Role2Vec framework based on the proposed notion of attributed random walks to learn structural role-based embeddings. Notably, the framework serves as a basis for generalizing any walk-based method. The Role2Vec framework enables these methods to be more widely applicable by learning inductive functions that capture the structural roles in the graph. Furthermore, the original methods are recovered as a special case of the framework when each vertex is mapped to its own function that uniquely identifies it. Finally, the Role2Vec framework is shown to be effective with an average AUC improvement of 17.8% for link prediction while requiring on average 853x less space than existing methods on a variety of graphs from different domains.
- Published
- 2022
19. Automatic Detection of Reflective Thinking in Mathematical Problem Solving Based on Unconstrained Bodily Exploration
- Author
-
Radoslaw Niewiadomski, Erica Volta, Joseph W. Newbold, Gualtiero Volpe, Rose Johnson, Temitayo A. Olugbade, Max Dillon, Paolo Alborno, and Nadia Bianchi-Berthouze
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Mathematical problem ,Computer science ,Problem-solving ,Computer Science - Human-Computer Interaction ,Neural nets ,Machine Learning (stat.ML) ,02 engineering and technology ,Affect sensing and analysis ,Education ,Human-Computer Interaction (cs.HC) ,Machine Learning (cs.LG) ,Interactive Learning ,Statistics - Machine Learning ,Emotional corpora ,Annotations ,0202 electrical engineering, electronic engineering, information engineering ,0501 psychology and cognitive sciences ,Observers ,050107 human factors ,Artificial neural network ,Movement (music) ,G400 ,05 social sciences ,020207 software engineering ,Body movement ,Human-Computer Interaction ,Binary classification ,Task analysis ,Games ,Neural networks ,F1 score ,Software ,Cognitive psychology - Abstract
For technology (like serious games) that aims to deliver interactive learning, it is important to address relevant mental experiences such as reflective thinking during problem solving. To facilitate research in this direction, we present the weDraw-1 Movement Dataset of body movement sensor data and reflective thinking labels for 26 children solving mathematical problems in unconstrained settings where the body (full or parts) was required to explore these problems. Further, we provide qualitative analysis of behaviours that observers used in identifying reflective thinking moments in these sessions. The body movement cues from our compilation informed features that lead to average F1 score of 0.73 for binary classification of problem-solving episodes by reflective thinking based on Long Short-Term Memory neural networks. We further obtained 0.79 average F1 score for end-to-end classification, i.e. based on raw sensor data. Finally, the algorithms resulted in 0.64 average F1 score for subsegments of these episodes as short as 4 seconds. Overall, our results show the possibility of detecting reflective thinking moments from body movement behaviours of a child exploring mathematical concepts bodily, such as within serious game play.
- Published
- 2022
20. Quantifying the Alignment of Graph and Features in Deep Learning
- Author
-
Pietro Panzarasa, Tom Rieu, Yifan Qian, Paul Expert, Mauricio Barahona, and Engineering & Physical Science Research Council (EPSRC)
- Subjects
FOS: Computer and information sciences ,principal angles ,Computer Science - Machine Learning ,Technology ,Data alignment ,Computer science ,cs.LG ,02 engineering and technology ,Computer Science, Artificial Intelligence ,Machine Learning (cs.LG) ,Engineering ,Statistics - Machine Learning ,Chordal graph ,0202 electrical engineering, electronic engineering, information engineering ,Artificial Intelligence & Image Processing ,physics.soc-ph ,Computer Science - Neural and Evolutionary Computing ,Computer Science - Social and Information Networks ,SCIENCE ,stat.ML ,Linear subspace ,Graph ,Computer Science Applications ,Task analysis ,Graph (abstract data type) ,020201 artificial intelligence & image processing ,graph subspaces ,cs.SI ,Subspace topology ,Physics - Physics and Society ,Computer Networks and Communications ,Matrix norm ,FOS: Physical sciences ,Machine Learning (stat.ML) ,Physics and Society (physics.soc-ph) ,Measure (mathematics) ,Symmetric matrices ,Deep Learning ,Computer Science, Theory & Methods ,Artificial Intelligence ,Training ,Neural and Evolutionary Computing (cs.NE) ,cs.NE ,Computer Science, Hardware & Architecture ,Social and Information Networks (cs.SI) ,Science & Technology ,Nonhomogeneous media ,Learning systems ,business.industry ,Deep learning ,Engineering, Electrical & Electronic ,Pattern recognition ,Convolution ,graph convolutional networks (GCNs) ,Computer Science ,Neural Networks, Computer ,Artificial intelligence ,business ,Software - Abstract
We show that the classification performance of graph convolutional networks (GCNs) is related to the alignment between features, graph, and ground truth, which we quantify using a subspace alignment measure (SAM) corresponding to the Frobenius norm of the matrix of pairwise chordal distances between three subspaces associated with features, graph, and ground truth. The proposed measure is based on the principal angles between subspaces and has both spectral and geometrical interpretations. We showcase the relationship between the SAM and the classification performance through the study of limiting cases of GCNs and systematic randomizations of both features and graph structure applied to a constructive example and several examples of citation networks of different origins. The analysis also reveals the relative importance of the graph and features for classification purposes., Comment: Published in IEEE Transactions on Neural Networks and Learning Systems; Date of Publication: 11 January 2021
- Published
- 2022
21. Issues and Challenges of Aspect-based Sentiment Analysis: A Comprehensive Survey
- Author
-
Ling Sun, Yuan Rao, Ambreen Nazir, and Lianwei Wu
- Subjects
0209 industrial biotechnology ,Computer science ,business.industry ,Deep learning ,Sentiment analysis ,02 engineering and technology ,Data science ,Field (computer science) ,Domain (software engineering) ,Human-Computer Interaction ,020901 industrial engineering & automation ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,020201 artificial intelligence & image processing ,Social media ,Artificial intelligence ,business ,Data objects ,Software - Abstract
The domain of Aspect-based Sentiment Analysis, in which aspects are extracted, their sentiments are analyzed and sentiments are evolved over time, is getting much attention with increasing feedback of public and customers on social media. The immense advancements in the field urged researchers to devise new techniques and approaches, each sermonizing a different research analysis/question, that cope with upcoming issues and complex scenarios of Aspect-based Sentiment Analysis. Therefore, this survey emphasized on the issues and challenges that are related to extraction of different aspects and their relevant sentiments, relational mapping between aspects, interactions, dependencies and contextual-semantic relationships between different data objects for improved sentiment accuracy, and prediction of sentiment evolution dynamicity. A rigorous overview of the recent progress is summarized based on whether they contributed towards highlighting and mitigating the issue of Aspect Extraction, Aspect Sentiment Analysis or Sentiment Evolution. The reported performance for each scrutinized study of Aspect Extraction and Aspect Sentiment Analysis is also given, showing the quantitative evaluation of the proposed approach. Future research directions are proposed and discussed, by critically analysing the presented recent solutions, that will be helpful for researchers and beneficial for improving sentiment classification at aspect-level.
- Published
- 2022
22. Effort-Aware Just-in-Time Bug Prediction for Mobile Apps Via Cross-Triplet Deep Feature Embedding
- Author
-
Zhiwen Xie, Chunlei Fu, Meng Yan, Gemma Catolino, Xiaohong Zhang, Kunsong Zhao, Tao Zhang, and Zhou Xu
- Subjects
Computer science ,Feature vector ,mobile app bug prediction ,metric learning ,02 engineering and technology ,Commit ,Machine learning ,computer.software_genre ,just-in-time bug prediction ,Task (project management) ,Predictive models ,Mobile applications ,020204 information systems ,mental disorders ,0202 electrical engineering, electronic engineering, information engineering ,Leverage (statistics) ,Electrical and Electronic Engineering ,Safety, Risk, Reliability and Quality ,Computer bugs ,Cross-triplet feature embedding ,Data models ,Deep learning ,effort-aware performance ,Feature extraction ,Task analysis ,Artificial neural network ,business.industry ,020207 software engineering ,Feature (computer vision) ,Benchmark (computing) ,Artificial intelligence ,business ,Feature learning ,computer - Abstract
Just-in-time (JIT) bug prediction is an effective quality assurance activity that identifies whether a code commit will introduce bugs into the mobile app, aiming to provide prompt feedback to practitioners for priority review. Since collecting sufficient labeled bug data is not always feasible for some mobile apps, one possible approach is to leverage cross-app models. In this work, we propose a new cross-triplet deep feature embedding method, called CDFE, for cross-app JIT bug prediction task. The CDFE method incorporates a state-of-the-art cross-triplet loss function into a deep neural network to learn high-level feature representation for the cross-app data. This loss function adapts to the cross-app feature learning task and aims to learn a new feature space to shorten the distance of commit instances with the same label and enlarge the distance of commit instances with different labels. In addition, this loss function assigns higher weights to losses caused by cross-app instance pairs than that by intra-app instance pairs, aiming to narrow the discrepancy of cross-app bug data. We evaluate our CDFE method on a benchmark bug dataset from 19 mobile apps with two effort-aware indicators. The experimental results on 342 cross-app pairs show that our proposed CDFE method performs better than 14 baseline methods.
- Published
- 2022
23. RefactoringMiner 2.0
- Author
-
Nikolaos Tsantalis, Ameya Ketkar, and Danny Dig
- Subjects
Code review ,Computer science ,business.industry ,020207 software engineering ,02 engineering and technology ,computer.software_genre ,Machine learning ,Maintenance engineering ,Code refactoring ,Application domain ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Code (cryptography) ,Software system ,Artificial intelligence ,business ,computer ,Software - Abstract
Refactoring detection is crucial for a variety of applications and tasks: (i) empirical studies about code evolution, (ii) tools for library API migration, (iii) code reviews and change comprehension. However, recent research has questioned the accuracy of the state-of-the-art refactoring mining tools, which poses threats to the reliability of the detected refactorings. Moreover, the majority of refactoring mining tools depend on code similarity thresholds. Finding universal threshold values that can work well for all projects, regardless of their architectural style, application domain, and development practices is extremely challenging. Therefore, in a previous work [1], we introduced the first refactoring mining tool that does not require any code similarity thresholds to operate. In this work, we extend our tool to support low-level refactorings that take place within the body of methods. To evaluate our tool, we created one of the most accurate, complete, and representative refactoring oracles to date, including 7,223 true instances for 40 different refactoring types detected by one (minimum) up to six (maximum) different tools, and validated by one up to four refactoring experts. Our evaluation showed that our approach achieves the highest average precision (99.6%) and recall (94%) among all competitive tools, and on median is 2.6 times faster than the second faster competitive tool.
- Published
- 2022
24. A Novel RL-Assisted Deep Learning Framework for Task-Informative Signals Selection and Classification for Spontaneous BCIs
- Author
-
Heung-Il Suk, Eunjin Jeon, and Wonjun Ko
- Subjects
Signal Processing (eess.SP) ,FOS: Computer and information sciences ,Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer science ,Feature extraction ,02 engineering and technology ,Machine learning ,computer.software_genre ,Machine Learning (cs.LG) ,Task (project management) ,FOS: Electrical engineering, electronic engineering, information engineering ,0202 electrical engineering, electronic engineering, information engineering ,Reinforcement learning ,Electrical Engineering and Systems Science - Signal Processing ,Electrical and Electronic Engineering ,Brain–computer interface ,business.industry ,Deep learning ,I.2.8 ,020208 electrical & electronic engineering ,Computer Science Applications ,Identification (information) ,Artificial Intelligence (cs.AI) ,Control and Systems Engineering ,Task analysis ,Artificial intelligence ,Markov decision process ,business ,computer ,Information Systems - Abstract
In this work, we formulate the problem of estimating and selecting task-relevant temporal signal segments from a single EEG trial in the form of a Markov decision process and propose a novel reinforcement-learning mechanism that can be combined with the existing deep-learning based BCI methods. To be specific, we devise an actor-critic network such that an agent can determine which timepoints need to be used (informative) or discarded (uninformative) in composing the intention-related features in a given trial, and thus enhancing the intention identification performance. To validate the effectiveness of our proposed method, we conducted experiments with a publicly available big MI dataset and applied our novel mechanism to various recent deep-learning architectures designed for MI classification. Based on the exhaustive experiments, we observed that our proposed method helped achieve statistically significant improvements in performance., 8 pages, 6 figures, 2 tables, and under review
- Published
- 2022
25. Multi-Task Learning With Coarse Priors for Robust Part-Aware Person Re-Identification
- Author
-
Dacheng Tao, Kan Wang, Pengfei Wang, and Changxing Ding
- Subjects
FOS: Computer and information sciences ,Databases, Factual ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,Feature vector ,Feature extraction ,Computer Science - Computer Vision and Pattern Recognition ,Multi-task learning ,02 engineering and technology ,Electronic mail ,Task (project management) ,Deep Learning ,Artificial Intelligence ,Robustness (computer science) ,0202 electrical engineering, electronic engineering, information engineering ,Humans ,Pedestrians ,business.industry ,Applied Mathematics ,Pattern recognition ,Computational Theory and Mathematics ,Feature (computer vision) ,Task analysis ,020201 artificial intelligence & image processing ,Computer Vision and Pattern Recognition ,Artificial intelligence ,business ,Algorithms ,Software - Abstract
Part-level representations are important for robust person re-identification (ReID), but in practice feature quality suffers due to the body part misalignment problem. In this paper, we present a robust, compact, and easy-to-use method called the Multi-task Part-aware Network (MPN), which is designed to extract semantically aligned part-level features from pedestrian images. MPN solves the body part misalignment problem via multi-task learning (MTL) in the training stage. More specifically, it builds one main task (MT) and one auxiliary task (AT) for each body part on the top of the same backbone model. The ATs are equipped with a coarse prior of the body part locations for training images. ATs then transfer the concept of the body parts to the MTs via optimizing the MT parameters to identify part-relevant channels from the backbone model. Concept transfer is accomplished by means of two novel alignment strategies: namely, parameter space alignment via hard parameter sharing and feature space alignment in a class-wise manner. With the aid of the learned high-quality parameters, MTs can independently extract semantically aligned part-level features from relevant channels in the testing stage. MPN has three key advantages: 1) it does not need to conduct body part detection in the inference stage; 2) its model is very compact and efficient for both training and testing; 3) in the training stage, it requires only coarse priors of body part locations, which are easy to obtain. Systematic experiments on four large-scale ReID databases demonstrate that MPN consistently outperforms state-of-the-art approaches by significant margins. Code is available at https://github.com/WangKan0128/MPN., Accepted Version to IEEE Transactions on Pattern Analysis and Machine Intelligence
- Published
- 2022
26. Pegasus: Performance Engineering for Software Applications Targeting HPC Systems
- Author
-
Pedro Pinto, Gianluca Palermo, João Bispo, Cristina Silvano, Jorge G. Barbosa, Jan Martinovič, Joao M. P. Cardoso, Katerina Slaninova, Davide Gadioli, Martin Golasowski, Radim Cmar, and Faculdade de Engenharia
- Subjects
Profiling (computer programming) ,Multi-core processor ,Application programming interface ,Computer science ,business.industry ,Performance tuning ,020207 software engineering ,02 engineering and technology ,Tuning ,computer.software_genre ,Toolchain ,020202 computer hardware & architecture ,Tools ,Software ,Runtime ,Task analysis ,Performance engineering ,0202 electrical engineering, electronic engineering, information engineering ,Power demand ,Compiler ,Software engineering ,business ,computer - Abstract
Developing and optimizing software applications for high performance and energy efficiency is a very challenging task, even when considering a single target machine. For instance, optimizing for multicore-based computing systems requires in-depth knowledge about programming languages, application programming interfaces, compilers, performance tuning tools, and computer architecture and organization. Many of the tasks of performance engineering methodologies require manual efforts and the use of different tools not always part of an integrated toolchain. This paper presents Pegasus, a performance engineering approach supported by a framework that consists of a source-to-source compiler, controlled and guided by strategies programmed in a Domain-Specific Language, and an autotuner. Pegasus is a holistic and versatile approach spanning various decision layers composing the software stack, and exploiting the system capabilities and workloads effectively through the use of runtime autotuning. The Pegasus approach helps developers by automating tasks regarding the efficient implementation of software applications in multicore computing systems. These tasks focus on application analysis, profiling, code transformations, and the integration of runtime autotuning. Pegasus allows developers to program their strategies or to automatically apply existing strategies to software applications in order to ensure the compliance of non-functional requirements, such as performance and energy efficiency. We show how to apply Pegasus and demonstrate its applicability and effectiveness in a complex case study, which includes tasks from a smart navigation system.
- Published
- 2022
27. User Identity Linkage via Co-Attentive Neural Network From Heterogeneous Mobility Data
- Author
-
Depeng Jin, Zeyu Yang, Mingyang Zhang, Huandong Wang, Han Cao, Yong Li, and Jie Feng
- Subjects
Artificial neural network ,business.industry ,Computer science ,Deep learning ,02 engineering and technology ,Service provider ,Computer Science Applications ,Data modeling ,Computational Theory and Mathematics ,Human–computer interaction ,020204 information systems ,Business intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Identity (object-oriented programming) ,Task analysis ,Artificial intelligence ,business ,Information Systems - Abstract
Online services are playing critical roles in almost all aspects of users' life. Users usually have multiple online identities (IDs) in different online services. In order to fuse the separated user data in multiple services for better business intelligence, it is critical for service providers to link online IDs belonging to the same user. On the other hand, the popularity of mobile networks and GPS-equipped smart devices have provided a generic way to link IDs, i.e., utilizing the mobility traces of IDs. However, linking IDs based on their mobility traces has been a challenging problem due to the highly heterogeneous, incomplete and noisy mobility data across services. In this paper, we propose DPLink, an end-to-end deep learning based framework, to complete the user identity linkage task for heterogeneous mobility data collected from different services with different properties. DPLink is made up by a feature extractor including a location encoder and a trajectory encoder to extract representative features from trajectory and a comparator to compare and decide whether to link two trajectories as the same user. Particularly, we propose a pre-training strategy with a simple task to train the DPLink model to overcome the training difficulties introduced by the highly heterogeneous nature of different source mobility data. Besides, we introduce a multi-modal embedding network and a co-attention mechanism in DPLink to deal with the low-quality problem of mobility data. By conducting extensive experiments on two real-life ground-truth mobility datasets with eight baselines, we demonstrate that DPLink outperforms the state-of-the-art solutions by more than 15% in terms of hit-precision. Moreover, it is expandable to add external geographical context data and works stably with heterogeneous noisy mobility traces.
- Published
- 2022
28. ConfusionFlow: A Model-Agnostic Visualization for Temporal Analysis of Classifier Confusion
- Author
-
Hendrik Strobelt, Andreas Hinterreiter, Martin Ennemoser, Peter Ruch, Marc Streit, Jürgen Bernard, and Holger Stitz
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,Computer Science - Computer Vision and Pattern Recognition ,Machine Learning (stat.ML) ,02 engineering and technology ,Machine learning ,computer.software_genre ,Machine Learning (cs.LG) ,Data modeling ,Data visualization ,Statistics - Machine Learning ,Classifier (linguistics) ,0202 electrical engineering, electronic engineering, information engineering ,Instance selection ,Artificial neural network ,business.industry ,Model selection ,Confusion matrix ,020207 software engineering ,Computer Graphics and Computer-Aided Design ,Visualization ,Signal Processing ,Active learning ,Task analysis ,Computer Vision and Pattern Recognition ,Artificial intelligence ,business ,computer ,Software - Abstract
Classifiers are among the most widely used supervised machine learning algorithms. Many classification models exist, and choosing the right one for a given task is difficult. During model selection and debugging, data scientists need to assess classifiers' performances, evaluate their learning behavior over time, and compare different models. Typically, this analysis is based on single-number performance measures such as accuracy. A more detailed evaluation of classifiers is possible by inspecting class errors. The confusion matrix is an established way for visualizing these class errors, but it was not designed with temporal or comparative analysis in mind. More generally, established performance analysis systems do not allow a combined temporal and comparative analysis of class-level information. To address this issue, we propose ConfusionFlow, an interactive, comparative visualization tool that combines the benefits of class confusion matrices with the visualization of performance characteristics over time. ConfusionFlow is model-agnostic and can be used to compare performances for different model types, model architectures, and/or training and test datasets. We demonstrate the usefulness of ConfusionFlow in a case study on instance selection strategies in active learning. We further assess the scalability of ConfusionFlow and present a use case in the context of neural network pruning., Changes compared to previous version: Reintroduced NN pruning use case; restructured Evaluation section; several additional minor revisions. Submitted as Minor Revision to IEEE TVCG on 2020-07-02
- Published
- 2022
29. When Information Freshness Meets Service Latency in Federated Learning: A Task-Aware Incentive Scheme for Smart Industries
- Author
-
Jiawen Kang, Cyril Leung, Dusit Niyato, Chunyan Miao, Wei Yang Bryan Lim, Zehui Xiong, Xuemin Shen, School of Computer Science and Engineering, Alibaba-NTU Joint Research Institute, and Joint NTU-UBC Research Centre of Excellence in Active Living for the Elderly (LILY)
- Subjects
Data collection ,business.industry ,Computer science ,Age of Information ,020208 electrical & electronic engineering ,02 engineering and technology ,Computer Science Applications ,System model ,Data modeling ,Incentive ,Control and Systems Engineering ,Incentive compatibility ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Computer science and engineering [Engineering] ,Electrical and Electronic Engineering ,Latency (engineering) ,business ,Federated Learning ,Information Systems ,Computer network - Abstract
For several industrial applications, a sole data owner may lack sufficient training samples to train effective machine learning based models. As such, we propose a Federated Learning (FL) based approach to promote privacy-preserving collaborative machine learning for applications in smart industries. In our system model, a model owner initiates an FL task involving a group of workers, i.e., data owners, to perform model training on their locally stored data before transmitting the model updates for aggregation. There exists a tradeoff between service latency, i.e., the time taken for the training request to be completed, and Age of Information (AoI), i.e., the time elapsed between data aggregation from the deployed IIoT devices to completion of the FL based training. On one hand, if the data is collected only upon the model owner's request, the AoI is low. On the other hand, the service latency incurred is more significant. Furthermore, given that different training tasks may have varying AoI requirements, we propose a contract-theoretic task-aware incentive scheme that can be calibrated based on the weighted preferences of the model owner towards AoI and service latency. Performance evaluation validates the incentive compatibility of our contract amid information asymmetry, and shows the flexibility of our proposed scheme towards satisfying varying preferences of AoI and service latency. AI Singapore Energy Market Authority (EMA) Ministry of Education (MOE) Nanyang Technological University National Research Foundation (NRF) Accepted version This research is supported, in part, by Alibaba Group through Alibaba Innovative Research (AIR) Program and Alibaba-NTU Singapore Joint Research Institute (JRI), National Research Foundation, Singapore, under its AI Singapore Programme (AISG Award No: AISG-GC-2019-003), Singapore Energy Market Authority (EMA), Energy Resilience, under Grant NRF2017EWT-EP003-041; and in part by the Singapore NRF2015-NRF-ISF001-2277. Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not reflect the views of National Research Foundation, Singapore. This research is also supported by WASP/NTU grant M4082187 (4080) and Singapore Ministry of Education (MOE) Tier 1 (RG16/20).
- Published
- 2022
30. Transfer Meta-Learning: Information- Theoretic Bounds and Information Meta-Risk Minimization
- Author
-
Giuseppe Durisi, Osvaldo Simeone, and Sharu Theresa Jose
- Subjects
Kullback–Leibler divergence ,Meta learning (computer science) ,Loss measurement ,Computer science ,Initialization ,02 engineering and technology ,Library and Information Sciences ,Upper and lower bounds ,single-draw bounds ,Task (project management) ,0202 electrical engineering, electronic engineering, information engineering ,Training ,PAC-Bayesian bounds ,Inductive bias ,Transfer meta-learning ,information risk minimization ,020206 networking & telecommunications ,Hospitals ,information-theoretic generalization bounds ,Transfer learning ,Computer Science Applications ,Risk management ,Task analysis ,Transfer of learning ,Algorithm ,Upper bound ,Information Systems - Abstract
Meta-learning automatically infers an inductive bias by observing data from a number of related tasks. The inductive bias is encoded by hyperparameters that determine aspects of the model class or training algorithm, such as initialization or learning rate. Meta-learning assumes that the learning tasks belong to a task environment, and that tasks are drawn from the same task environment both during meta-training and meta-testing. This, however, may not hold true in practice. In this paper, we introduce the problem of transfer meta-learning, in which tasks are drawn from a target task environment during meta-testing that may differ from the source task environment observed during meta-training. Novel information-theoretic upper bounds are obtained on the transfer meta-generalization gap, which measures the difference between the meta-training loss, available at the meta-learner, and the average loss on meta-test data from a new, randomly selected, task in the target task environment. The first bound, on the average transfer meta-generalization gap, captures the meta-environment shift between source and target task environments via the KL divergence between source and target data distributions. The second, PAC-Bayesian bound, and the third, single-draw bound, account for this shift via the log-likelihood ratio between source and target task distributions. Furthermore, two transfer meta-learning solutions are introduced. For the first, termed Empirical Meta-Risk Minimization (EMRM), we derive bounds on the average optimality gap. The second, referred to as Information Meta-Risk Minimization (IMRM), is obtained by minimizing the PAC-Bayesian bound. IMRM is shown via experiments to potentially outperform EMRM.
- Published
- 2022
31. Deep Learning for HDD Health Assessment: An Application Based on LSTM
- Author
-
Aniello De Santo, Vincenzo Moscato, Antonio Galli, Giancarlo Sperlì, Michela Gravina, de Santo, A., Galli, A., Gravina, M., Moscato, V., and Sperlì, G.
- Subjects
Downtime ,Service (systems architecture) ,business.industry ,Computer science ,Reliability (computer networking) ,Deep learning ,02 engineering and technology ,Data loss ,Machine learning ,computer.software_genre ,Predictive maintenance ,020202 computer hardware & architecture ,Theoretical Computer Science ,Data modeling ,Hard drive failure prediction, SMART Health degree, Long short-term memory ,Computational Theory and Mathematics ,Hardware and Architecture ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Artificial intelligence ,business ,computer ,Software - Abstract
Hard disk drive failures are one of the most common causes of service downtime in data centers. Predictive maintenance techniques have been adopted to extend the Remaining Useful Life (RUL) of these drives, and minimize service shortage and data loss. Several approaches based on machine and deep learning techniques have been proposed to address these issues, mostly exploiting models based on Self-Monitoring analysis and Reporting Technology (SMART) attributes. While these models have proven to be reliable, their performance is affected by the lack of information about the proximity of disk failure in time. Moreover, many of these techniques are sensitive to the highly unbalanced nature of existing data-sets, in terms of good to failed hard disk ratio. In this paper, we propose a LSTM based model combining SMART attributes and temporal analysis for estimating a hard drive health status according to its time to failure. Our approach outperforms state-of-the-art methods when evaluated on two data-sets, one containing hourly samples from 23,395 disks and the other reporting daily samples from 29,878 disks. Experimental results showed that our approach is well suited to data-sets with different sampling periods, being able to predict hard drive health status up to 45 days before failure.
- Published
- 2022
32. Multiple Instance Learning for Emotion Recognition Using Physiological Signals
- Author
-
Andrea Cavallo, Nadia Berthouze, Lucia Pepa, Luca Romeo, and Massimiliano Pontil
- Subjects
Support Vector Machine ,Diverse Density ,Computer science ,media_common.quotation_subject ,Reliability (computer networking) ,Multiple Instance Learning ,02 engineering and technology ,Physiological signals ,Machine learning ,computer.software_genre ,050105 experimental psychology ,0202 electrical engineering, electronic engineering, information engineering ,Natural (music) ,0501 psychology and cognitive sciences ,Affective computing ,media_common ,business.industry ,05 social sciences ,Ambiguity ,Emotion Recognition ,Time Ambiguity ,Human-Computer Interaction ,Support vector machine ,Task analysis ,020201 artificial intelligence & image processing ,State (computer science) ,Artificial intelligence ,Sequence learning ,business ,computer ,Software - Abstract
The problem of continuous emotion recognition has been the subject of several studies. The proposed affective computing approaches employ sequential machine learning algorithms for improving the classification stage, accounting for the time ambiguity of emotional responses. Modeling and predicting the affective state over time is not a trivial problem because continuous data labeling is costly and not always feasible. This is a crucial issue in real-life applications, where data labeling is sparse and possibly captures only the most important events rather than the typical continuous subtle affective changes that occur. In this work, we introduce a framework from the machine learning literature called Multiple Instance Learning, which is able to model time intervals by capturing the presence or absence of relevant states, without the need to label the affective responses continuously (as required by standard sequential learning approaches). This choice offers a viable and natural solution for learning in a weakly supervised setting, taking into account the ambiguity of affective responses. We demonstrate the reliability of the proposed approach in a gold-standard scenario and towards real-world usage by employing an existing dataset (DEAP) and a purposely built one (Consumer). We also outline the advantages of this method with respect to standard supervised machine learning algorithms.
- Published
- 2022
33. Approximate Transverse Feedback Linearization Under Digital Control
- Author
-
Dorothée Normand-Cyrot, Salvatore Monaco, Mohamed Elobaid, Dipartimento di Ingegneria informatica automatica e gestionale (DIAG), Università degli Studi di Roma 'La Sapienza' = Sapienza University [Rome], Laboratoire des signaux et systèmes (L2S), and CentraleSupélec-Université Paris-Saclay-Centre National de la Recherche Scientifique (CNRS)
- Subjects
0209 industrial biotechnology ,Control and Optimization ,Computer science ,Trajectory ,Aerospace electronics ,02 engineering and technology ,[SPI.AUTO]Engineering Sciences [physics]/Automatic ,Algebraic/geometric methods ,020901 industrial engineering & automation ,Control theory ,0202 electrical engineering, electronic engineering, information engineering ,[INFO.INFO-SY]Computer Science [cs]/Systems and Control [cs.SY] ,MIMO communication ,Feedback linearization ,Digital control ,Manifolds ,020208 electrical & electronic engineering ,Stability analysis ,Mimo communication ,Transverse plane ,Order (business) ,Control and Systems Engineering ,Sampled-data control ,Task analysis - Abstract
International audience; Thanks to a suitable redesign of the maps involved in the continuous-time solution, a digital design procedure preserving transverse feedback linearization up to a prefixed order of approximation in the sampling period is described. Simulated examples illustrate the results.
- Published
- 2022
34. Improving Machine Vision Using Human Perceptual Representations: The Case of Planar Reflection Symmetry for Object Classification
- Author
-
Arun Sp and R. T. Pramod
- Subjects
Computer science ,Machine vision ,media_common.quotation_subject ,Computational models of Vision ,02 engineering and technology ,Convolutional neural network ,Article ,Reflection symmetry ,Artificial Intelligence ,Perception and Psychophysics ,Perception ,0202 electrical engineering, electronic engineering, information engineering ,Humans ,Object perception ,media_common ,Object Recognition ,business.industry ,Applied Mathematics ,Cognitive neuroscience of visual object recognition ,Pattern recognition ,Object (computer science) ,Visualization ,Pattern Recognition, Visual ,Computational Theory and Mathematics ,Task analysis ,020201 artificial intelligence & image processing ,Neural Networks, Computer ,Computer Vision and Pattern Recognition ,Artificial intelligence ,business ,Algorithms ,Software - Abstract
Achieving human-like visual abilities is a holy grail for machine vision, yet precisely how insights from human vision can improve machines has remained unclear. Here, we demonstrate two key conceptual advances: First, we show that most machine vision models are systematically different from human object perception. To do so, we collected a large dataset of perceptual distances between isolated objects in humans and asked whether these perceptual data can be predicted by many common machine vision algorithms. We found that while the best algorithms explain ∼ 70 percent of the variance in the perceptual data, all the algorithms we tested make systematic errors on several types of objects. In particular, machine algorithms underestimated distances between symmetric objects compared to human perception. Second, we show that fixing these systematic biases can lead to substantial gains in classification performance. In particular, augmenting a state-of-the-art convolutional neural network with planar/reflection symmetry scores along multiple axes produced significant improvements in classification accuracy (1-10 percent) across categories. These results show that machine vision can be improved by discovering and fixing systematic differences from human vision.
- Published
- 2022
35. Wasserstein Adversarial Regularization for learning with label noise
- Author
-
Nicolas Courty, Devis Tuia, Rémi Flamary, Bharath Bhushan Damodaran, Sylvain Lobry, Kilian Fatras, Université de Bretagne Sud (UBS), Institut de Recherche en Informatique et Systèmes Aléatoires (IRISA), Université de Rennes 1 (UR1), Université de Rennes (UNIV-RENNES)-Université de Rennes (UNIV-RENNES)-Institut National des Sciences Appliquées - Rennes (INSA Rennes), Institut National des Sciences Appliquées (INSA)-Université de Rennes (UNIV-RENNES)-Institut National des Sciences Appliquées (INSA)-Université de Bretagne Sud (UBS)-École normale supérieure - Rennes (ENS Rennes)-Institut National de Recherche en Informatique et en Automatique (Inria)-CentraleSupélec-Centre National de la Recherche Scientifique (CNRS)-IMT Atlantique Bretagne-Pays de la Loire (IMT Atlantique), Institut Mines-Télécom [Paris] (IMT)-Institut Mines-Télécom [Paris] (IMT), Observation de l’environnement par imagerie complexe (OBELIX), SIGNAUX ET IMAGES NUMÉRIQUES, ROBOTIQUE (IRISA-D5), Institut Mines-Télécom [Paris] (IMT)-Institut Mines-Télécom [Paris] (IMT)-Université de Rennes 1 (UR1), Institut Mines-Télécom [Paris] (IMT)-Institut Mines-Télécom [Paris] (IMT)-Institut de Recherche en Informatique et Systèmes Aléatoires (IRISA), Ecole Polytechnique Fédérale de Lausanne (EPFL), Centre de Mathématiques Appliquées - Ecole Polytechnique (CMAP), École polytechnique (X)-Centre National de la Recherche Scientifique (CNRS), École polytechnique (X), ANR-17-CE23-0012,OATMIL,Apprentissage statistique avec transport optimal(2017), ANR-20-CHIA-0030,OTTOPIA,Observation de la Terre par Transport Optimal pour l'Intelligence Artificielle(2020), ANR-18-EURE-0006,E4C,Energy for Climate Interdisciplinary Instute(2018), ANR-19-P3IA-0002,3IA@cote d'azur,3IA Côte d'Azur(2019), Université de Bretagne Sud (UBS)-Institut National des Sciences Appliquées - Rennes (INSA Rennes), Institut National des Sciences Appliquées (INSA)-Université de Rennes (UNIV-RENNES)-Institut National des Sciences Appliquées (INSA)-Université de Rennes (UNIV-RENNES)-Institut National de Recherche en Informatique et en Automatique (Inria)-École normale supérieure - Rennes (ENS Rennes)-Centre National de la Recherche Scientifique (CNRS)-Université de Rennes 1 (UR1), Université de Rennes (UNIV-RENNES)-CentraleSupélec-IMT Atlantique Bretagne-Pays de la Loire (IMT Atlantique), Université de Rennes (UR)-Institut National des Sciences Appliquées - Rennes (INSA Rennes), Institut National des Sciences Appliquées (INSA)-Institut National des Sciences Appliquées (INSA)-Université de Bretagne Sud (UBS)-École normale supérieure - Rennes (ENS Rennes)-Institut National de Recherche en Informatique et en Automatique (Inria)-CentraleSupélec-Centre National de la Recherche Scientifique (CNRS)-IMT Atlantique (IMT Atlantique), and Institut Mines-Télécom [Paris] (IMT)-Institut Mines-Télécom [Paris] (IMT)-Université de Rennes (UR)-Institut National des Sciences Appliquées - Rennes (INSA Rennes)
- Subjects
Computer science ,Noise measurement ,Entropy ,Adversarial regularization ,Context (language use) ,02 engineering and technology ,Signal-To-Noise Ratio ,Regularization (mathematics) ,Smoothing methods ,[INFO.INFO-LG]Computer Science [cs]/Machine Learning [cs.LG] ,Label noise ,Artificial Intelligence ,Classifier (linguistics) ,0202 electrical engineering, electronic engineering, information engineering ,Optimal transport ,Learning ,Entropy (information theory) ,Training ,Wasserstein distance ,ComputingMilieux_MISCELLANEOUS ,Artificial neural network ,business.industry ,Applied Mathematics ,Pattern recognition ,Semantics ,Computational Theory and Mathematics ,Task analysis ,Decision boundary ,020201 artificial intelligence & image processing ,Neural Networks, Computer ,Computer Vision and Pattern Recognition ,Artificial intelligence ,Noise (video) ,business ,Algorithms ,Software ,Neural networks - Abstract
Noisy labels often occur in vision datasets, especially when they are obtained from crowdsourcing or Web scraping. We propose a new regularization method, which enables learning robust classifiers in presence of noisy data. To achieve this goal, we propose a new adversarial regularization {scheme} based on the Wasserstein distance. Using this distance allows taking into account specific relations between classes by leveraging the geometric properties of the labels space. {Our Wasserstein Adversarial Regularization (WAR) encodes a selective regularization, which promotes smoothness of the classifier between some classes, while preserving sufficient complexity of the decision boundary between others. We first discuss how and why adversarial regularization can be used in the context of noise and then show the effectiveness of our method on five datasets corrupted with noisy labels: in both benchmarks and real datasets, WAR outperforms the state-of-the-art competitors.
- Published
- 2022
36. Social Media Data in an Augmented Reality System for Situation Awareness Support in Emergency Control Rooms
- Author
-
Jennifer Fromm, Melina Baßfeld, Tim A. Majchrzak, Stefan Stieglitz, and Kaan Eyilmez
- Subjects
Situation awareness ,Computer Networks and Communications ,Computer science ,business.industry ,Interface (computing) ,05 social sciences ,Internet privacy ,Information access ,02 engineering and technology ,Filter (software) ,VDP::Matematikk og Naturvitenskap: 400::Informasjons- og kommunikasjonsvitenskap: 420 ,Social media analytics ,Theoretical Computer Science ,Angewandte Kognitionswissenschaft ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,0501 psychology and cognitive sciences ,Social media ,Augmented reality ,business ,050107 human factors ,Software ,Information Systems - Abstract
During crisis situations, emergency operators require fast information access to achieve situation awareness and make the best possible decisions. Augmented reality could be used to visualize the wealth of user-generated content available on social media and enable context-adaptive functions for emergency operators. Although emergency operators agree that social media analytics will be important for their future work, it poses a challenge to filter and visualize large amounts of social media data. We conducted a goal-directed task analysis to identify the situation awareness requirements of emergency operators. By collecting tweets during two storms in Germany we evaluated the usefulness of Twitter data for achieving situation awareness and conducted interviews with emergency operators to derive filter strategies for social media data. We synthesized the results by discussing how the unique interface of augmented reality can be used to integrate social media data into emergency control rooms for situation awareness support.
- Published
- 2023
37. Stable Task Assignment for Mobile Crowdsensing With Budget Constraint
- Author
-
Pan Zhou, Deyu Qi, Xiumin Wang, Weiwei Lin, Chenxin Dai, and Kai Liu
- Subjects
Service quality ,Matching (statistics) ,Computer Networks and Communications ,Computer science ,Distributed computing ,Mobile computing ,Stability (learning theory) ,020206 networking & telecommunications ,02 engineering and technology ,Task (project management) ,Incentive ,Crowdsensing ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Electrical and Electronic Engineering ,Software ,Budget constraint - Abstract
In mobile crowdsensing, it is a challenge to assign tasks to appropriate smartphones. Existing task allocation mechanisms mainly aim at optimizing the global system performance, while ignoring the personal preferences of individual crowdsensing tasks and smartphone users. Nevertheless, in an open crowdsensing system, a task assignment is prone to be unstable if smartphone users or tasks have incentives to deviate from the global assignment, and seek for alternative choices to improve their own utilities. Besides that, during task competition, the rational smartphone users might choose to adjust their payments after the first few failures, which however, brings new challenges in achieving the stability. To address these issues, this paper constructs a distributed many-to-many matching model to capture the interaction between crowdsensing tasks and smartphone users, taking into account the budget constraints of tasks. Then, we design a stable matching algorithm to allocate the tasks to the users, and determine their payments. We prove that the proposed algorithm achieves several desirable properties including individual rationality, stability, and convergency. It is also proved that the proposed scheme achieves at least half of the optimal system efficiency when each smartphone provides homogeneous service quality. Finally, simulation results confirm the effectiveness of the proposed scheme.
- Published
- 2021
38. Multi-View Representation Learning With Deep Gaussian Processes
- Author
-
Qiuyang Liu, Wenbo Dong, and Shiliang Sun
- Subjects
Artificial neural network ,business.industry ,Computer science ,Applied Mathematics ,Deep learning ,02 engineering and technology ,Machine learning ,computer.software_genre ,Data modeling ,Data set ,symbols.namesake ,Computational Theory and Mathematics ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,symbols ,Task analysis ,020201 artificial intelligence & image processing ,Computer Vision and Pattern Recognition ,Artificial intelligence ,business ,Feature learning ,Gaussian process ,computer ,Classifier (UML) ,Software - Abstract
Multi-view representation learning is a promising and challenging research topic, which aims to integrate multiple data information from different views to improve the learning performance. The recent deep Gaussian processes (DGPs) have the advantages of good uncertainty estimates, powerful non-linear mapping ability and great generalization capability, which can be used as an excellent data representation learning method. However, DGPs only focus on single view data and are rarely applied to the multi-view scenario. In this paper, we propose a multi-view representation learning algorithm with deep Gaussian processes (named MvDGPs), which inherits the advantages of deep Gaussian processes and multi-view representation learning, and can learn more effective representation of multi-view data. The MvDGPs consist of two stages. The first stage is multi-view data representation learning, which is mainly used to learn more comprehensive representations of multi-view data. The second stage is classifier design, which aims to select an appropriate classifier to better employ the representations obtained in the first stage. In contrast with DGPs, MvDGPs support asymmetrical modeling depths for different views of data, resulting in better characterizations of the discrepancies among different views. Experimental results on real-world multi-view data sets verify the effectiveness of the proposed algorithm, which indicates that MvDGPs can integrate the complementary information in multiple views to discover a good representation of the data.
- Published
- 2021
39. Where2Change: Change Request Localization for App Reviews
- Author
-
David Lo, Xiapu Luo, Jiachi Chen, Xian Zhan, He Jiang, and Tao Zhang
- Subjects
Source code ,Information retrieval ,End user ,Computer science ,media_common.quotation_subject ,Change request ,Cosine similarity ,020207 software engineering ,02 engineering and technology ,Software maintenance ,Empirical research ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Software ,media_common - Abstract
Million of mobile apps have been released to the market. Developers need to maintain these apps so that they can continue to benefit end users. Developers usually extract useful information from user reviews to maintain and evolve mobile apps. One of the important activities that developers need to do while reading user reviews is to locate the source code related to requested changes. Unfortunately, this manual work is costly and time consuming since: (1) an app can receive thousands of reviews, and (2) a mobile app can consist of hundreds of source code files. To address this challenge, Palomba et al. recently proposed CHANGEADVISOR that utilizes user reviews to locate source code to be changed. However, we find that it cannot identify real source code to be changed for part of reviews. In this work, we aim to advance Palomba et al. 's work by proposing a novel approach that can achieve higher accuracy in change localization. Our approach first extracts the informative sentences (i.e., user feedback) from user reviews and identifies user feedback related to various problems and feature requests, and then cluster the corresponding user feedback into groups. Each group reports the similar users’ needs. Next, these groups are mapped to issue reports by using $Word2Vec$ W o r d 2 V e c . The resultant enriched text consisting of user feedback and their corresponding issue reports is used to identify source code classes that should be changed by using our novel weight selection -based cosine similarity metric. We have evaluated the new proposed change request localization approach ( Where2Change ) on 31,597 user reviews and 3,272 issue reports of 10 open source mobile apps. The experiments demonstrate that Where2Change can successfully locate more source code classes related to the change requests for more user feedback clusters than CHANGEADVISOR as demonstrated by higher Top-N and Recall values. The differences reach up to 17 for Top-1, 18.1 for Top-3, 17.9 for Top-5, and 50.08 percent for Recall. In addition, we also compare the performance of Where2Change and two previous Information Retrieval (IR)-based fault localization technologies: BLUiR and BLIA . The results showed that our approach performs better than them. As an important part of our work, we conduct an empirical study to investigate the value of using both user reviews and historical issue reports for change request localization; the results shown that historical issue reports can help to improve the performance of change localization.
- Published
- 2021
40. QoS-Based Budget Constrained Stable Task Assignment in Mobile Crowdsensing
- Author
-
Eyuphan Bulut, Murat Yuksel, and Fatih Yucel
- Subjects
Matching (statistics) ,Computer Networks and Communications ,Computer science ,Quality of service ,Distributed computing ,Stability (learning theory) ,Mobile computing ,020206 networking & telecommunications ,02 engineering and technology ,Task (project management) ,0202 electrical engineering, electronic engineering, information engineering ,Key (cryptography) ,Task analysis ,Leverage (statistics) ,Electrical and Electronic Engineering ,Software - Abstract
One of the key problems in mobile crowdsensing (MCS) systems is the assignment of tasks to users. Most of the existing work aim to maximize a predefined system utility (e.g., quality of service or sensing), however, users (i.e., task requesters and performers/workers) may value different parameters and hence find an assignment unsatisfying if it is produced disregarding these parameters that define their preferences. While several studies utilize incentive mechanisms to motivate user participation in different ways, they do not take individual user preferences into account either. To address this issue, we leverage Stable Matching Theory which can help obtain a satisfying matching between two groups of entities based on their preferences. However, the existing approaches to find stable matchings do not work in MCS systems due to the many-to-one nature of task assignments and the budget constraints of task requesters. Thus, we first define two different stability conditions for user happiness in MCS systems. Then, we propose three efficient stable task assignment algorithms and discuss their stability guarantees in four different MCS scenarios. Finally, we evaluate the performance of the proposed algorithms through extensive simulations using a real dataset, and show that they outperform the state-of-the-art solutions.
- Published
- 2021
41. Attention-Based Multilevel Co-Occurrence Graph Convolutional LSTM for 3-D Action Recognition
- Author
-
Shihao Xu, Bin Hu, Haocong Rao, Xin Jiang, Yi Guo, Hong Peng, and Xiping Hu
- Subjects
Computer Networks and Communications ,business.industry ,Computer science ,Feature extraction ,Co-occurrence ,Inter frame ,020206 networking & telecommunications ,Pattern recognition ,02 engineering and technology ,Computer Science Applications ,Hardware and Architecture ,Signal Processing ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Graph (abstract data type) ,RGB color model ,020201 artificial intelligence & image processing ,The Internet ,Artificial intelligence ,business ,Feature learning ,Information Systems - Abstract
Action recognition is essential for many human-centered applications in the Internet of Things (IoT). Especially, in the Internet of Medical Things (IoMT), action recognition shows great importance in surgical assistance, patient monitoring, etc. Recently, 3-D skeleton sequence-based action recognition draws broad attention. It is a challenging task that needs effective modeling on intraframe skeleton representations and interframe temporal dynamics. Standard long short-term memory (LSTM)-based models are widely used for sequence modeling due to its long-term memory, yet they are unable to fully model the relationship between different body joints or persons to extract crucial co-occurrence features from different levels. To handle this shortcoming, we propose an attention-based multilevel co-occurrence graph convolutional LSTM (AMCGC-LSTM). By integrating graph convolutional networks (GCNs) into LSTM, the proposed model is capable of leveraging body structural information from skeletons and strengthening the multilevel co-occurrence (MC) feature learning. Specifically, we first design the spatial attention module for feature enhancement of key joints from skeleton inputs. Second, we design MC memory units coupled with GCN to automatically model the spatial relationship between joints, and simultaneously capture the co-occurrence features from different joints, persons, and frames. Finally, we construct aggregated features of MCs (AFMCs) from MC memory units to better represent the intraframe action context encoding, and leverage a concurrent LSTM (Co-LSTM) to further model their temporal dynamics for action recognition. Our model significantly outperforms mainstream methods on NTU RGB+D 60/120 data set, mutual action subset of NTU RGB+D 60/120 data set, and Northewestern-UCLA data set.
- Published
- 2021
42. Adversarial Learning for Multiscale Crowd Counting Under Complex Scenes
- Author
-
Jianxing Yang, Tao Cao, Yuan Zhou, Hongru Li, and Sun-Yuan Kung
- Subjects
Similarity (geometry) ,business.industry ,Feature extraction ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,020206 networking & telecommunications ,Pattern recognition ,02 engineering and technology ,Convolutional neural network ,Computer Science Applications ,Human-Computer Interaction ,Control and Systems Engineering ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,020201 artificial intelligence & image processing ,Artificial intelligence ,Electrical and Electronic Engineering ,Crowd density ,business ,Software ,Crowd counting ,Information Systems ,Generator (mathematics) - Abstract
In this article, a multiscale generative adversarial network (MS-GAN) is proposed for generating high-quality crowd density maps of arbitrary crowd density scenes. The task of crowd counting has many challenges, such as severe occlusions in extremely dense crowd scenes, perspective distortion, and high visual similarity between the pedestrians and background elements. To address these problems, the proposed MS-GAN combines a multiscale convolutional neural network (generator) and an adversarial network (discriminator) to generate a high-quality density map and accurately estimate the crowd count in complex crowd scenes. The multiscale generator utilizes the fusion features from multiple hierarchical layers to detect people with large-scale variation. The resulting density map produced by the multiscale generator is processed by a discriminator network trained to solve a binary classification task between a poor quality density map and real ground-truth ones. The additional adversarial loss can improve the quality of the density map, which is critical to accurately estimate the crowd counts. The experiments were conducted on multiple datasets with different crowd scenes and densities. The results showed that the proposed method provided better performance compared to current state-of-the-art methods.
- Published
- 2021
43. Data-Efficient Hierarchical Reinforcement Learning for Robotic Assembly Control Applications
- Author
-
Zhimin Hou, Yuelin Deng, Jing Xu, and Jiajun Fei
- Subjects
Computer science ,business.industry ,020208 electrical & electronic engineering ,Control (management) ,02 engineering and technology ,DUAL (cognitive architecture) ,Machine learning ,computer.software_genre ,Control and Systems Engineering ,Resampling ,Softmax function ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Robot ,Reinforcement learning ,Markov decision process ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,computer - Abstract
Hierarchical reinforcement learning (HRL) can learn the decomposed subpolicies corresponding to the local state-space; therefore, it is a promising solution to complex robotic assembly control tasks with fewer interactions with environments. Most existing HRL algorithms often require on-policy learning, where resampling is necessary for every training step. In this article, we propose a data-efficient HRL via off-policy learning with three main contributions. First, two augmented MDPs (Markov decision processes) are reformulated to learn the higher level policy and lower level policy from the same samples. Second, to learn higher level policy that leads to efficient exploration, a softmax gating policy is derived to determine the lower level policy for interacting with the environment. Third, to learn the lower level policies via off-policy samples from one lower level replay buffer, the higher level policy derived by the option-value network is adopted to select the appropriate option for learning the corresponding lower level policy. The data-efficiency performance of our algorithm is validated on two simulations and real-world robotic dual peg-in-hole assembly tasks.
- Published
- 2021
44. Auxiliary Information-Guided Industrial Data Augmentation for Any-Shot Fault Learning and Diagnosis
- Author
-
Zhiqiang Ge and Yue Zhuo
- Subjects
Computer science ,media_common.quotation_subject ,020208 electrical & electronic engineering ,Process (computing) ,02 engineering and technology ,Space (commercial competition) ,computer.software_genre ,Fault (power engineering) ,Computer Science Applications ,Term (time) ,Control and Systems Engineering ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Train ,Quality (business) ,Data mining ,Electrical and Electronic Engineering ,Hydraulic machinery ,computer ,Information Systems ,media_common - Abstract
The label scarcity problem widely exists in industrial processes. In particular, samples of some fault types are extremely rare; even worse, the samples of certain faults cannot be accessed, but they may appear in the actual process. These two kinds of challenges together can be termed as any-shot learning problem in industrial fault diagnosis. In this article, taking the advantages of generative adversarial network, a generative approach is proposed to tackle the any-shot learning problem, which generates the abundant samples for those rare and inaccessible faults, and trains a strong diagnosis model. To reach this, an attribute space is built to introduce the auxiliary information, which achieves the diagnosis of unseen faults and makes the generated samples more resembled to the real data. Besides, an auxiliary loss of triplet form is introduced as a joint training loss term, further improving the quality of augmented data and diagnosis accuracy. Finally, the performance of model is verified by the experiments of a hydraulic system and Tennessee–Eastman process, the results of which show that our method performs excellently for both zero-shot and few-shot fault diagnosis problems.
- Published
- 2021
45. MP-Coopetition: Competitive and Cooperative Mechanism for Multiple Platforms in Mobile Crowd Sensing
- Author
-
Kashif Sharif, Youqi Li, Yue Wu, Yu Wang, Song Yang, Huijie Chen, and Fan Li
- Subjects
Service (systems architecture) ,Information Systems and Management ,Computer Networks and Communications ,Heuristic (computer science) ,Computer science ,Distributed computing ,Stochastic game ,020206 networking & telecommunications ,Coopetition ,02 engineering and technology ,Computer Science Applications ,Task (project management) ,Incentive ,Hardware and Architecture ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Stackelberg competition ,020201 artificial intelligence & image processing - Abstract
Mobile Crowd Sensing (MCS) enables the platform to offer data-based service by incentivizing mobile users to perform sensing task and then collecting sensing data from them. Most of the existing works on MCS only consider designing incentive mechanisms for a single MCS platform. In this paper, we study the incentive mechanism in MCS with multiple platforms under two scenarios: with competitive platforms and with cooperative platforms, and correspondingly propose new competitive or cooperative mechanisms for each scenario. In the competitive platform scenario, platforms decide their prices on rewards to attract more participants, while the users choose which platform to work for. We model such a competitive platform scenario as a two-stage Stackelberg game. In the cooperative platform scenario, platforms cooperate to share sensing data with each other. We model it as many-to-many bargaining. Moreover, we prove the NP-hardness of exact bargaining and then propose heuristic bargaining. Finally, numerical results show that (1) platforms in the competitive platform scenario can guarantee their payoff by optimally pricing on rewards and participants can select the best platform to contribute; (2) platforms in the cooperative platform scenario can further improve their payoff by bargaining with other platforms for cooperatively sharing collected sensing data.
- Published
- 2021
46. Recent Advances in Open Set Recognition: A Survey
- Author
-
Songcan Chen, Chuanxing Geng, and Sheng-Jun Huang
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Computer science ,Open set ,Machine Learning (stat.ML) ,02 engineering and technology ,One-shot learning ,Machine learning ,computer.software_genre ,Facial recognition system ,Machine Learning (cs.LG) ,Data visualization ,Statistics - Machine Learning ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,business.industry ,Applied Mathematics ,Computational Theory and Mathematics ,Task analysis ,020201 artificial intelligence & image processing ,Computer Vision and Pattern Recognition ,Artificial intelligence ,business ,computer ,Classifier (UML) ,Software - Abstract
In real-world recognition/classification tasks, limited by various objective factors, it is usually difficult to collect training samples to exhaust all classes when training a recognizer or classifier. A more realistic scenario is open set recognition (OSR), where incomplete knowledge of the world exists at training time, and unknown classes can be submitted to an algorithm during testing, requiring the classifiers to not only accurately classify the seen classes, but also effectively deal with the unseen ones. This paper provides a comprehensive survey of existing open set recognition techniques covering various aspects ranging from related definitions, representations of models, datasets, evaluation criteria, and algorithm comparisons. Furthermore, we briefly analyze the relationships between OSR and its related tasks including zero-shot, one-shot (few-shot) recognition/learning techniques, classification with reject option, and so forth. Additionally, we also overview the open world recognition which can be seen as a natural extension of OSR. Importantly, we highlight the limitations of existing approaches and point out some promising subsequent research directions in this field., Accepted by IEEE TPAMI
- Published
- 2021
47. Over-Sampling Emotional Speech Data Based on Subjective Evaluations Provided by Multiple Individuals
- Author
-
Carlos Busso and Reza Lotfian
- Subjects
Majority rule ,Artificial neural network ,business.industry ,Computer science ,media_common.quotation_subject ,020207 software engineering ,02 engineering and technology ,computer.software_genre ,Agreement ,Human-Computer Interaction ,030507 speech-language pathology & audiology ,03 medical and health sciences ,Perception ,Classifier (linguistics) ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Artificial intelligence ,0305 other medical science ,business ,Categorical variable ,computer ,Software ,Sentence ,Natural language processing ,media_common - Abstract
A common step in the area of speech emotion recognition is to obtain ground-truth labels describing the emotional content of a sentence. The underlying emotion of a given recording is usually unknown, so perceptual evaluations are conducted to annotate its perceived emotion. Each sentence is often annotated by multiple raters, which are aggregated with methods such as majority vote rules. This paper argues that several labels provided by different individuals convey more information than the consensus labels. We demonstrate that leveraging the information provided by separate evaluations collected by multiple raters can help in building more robust classifiers which maximize the utilization of labeled data. Motivated by the synthetic minority over-sampling technique(SMOTE), we present a novel over-sampling approach during training, where the samples with categorical emotion labels are over-sampled according to the labels assigned by multiple individuals. This approach (1)increases the number of sentences from classes with underrepresented consensus labels, and (2)utilizes sentences with ambiguous emotional content even if they do not reach consensus agreement. The experimental evaluation shows the benefits of the approach over a baseline classifier trained with consensus labels, which increases the F1-score by 5.2% (absolute) for the USC-IEMOCAP corpus, and 5.4% (absolute) for the MSP-IMPROV corpus.
- Published
- 2021
48. IAUnet: Global Context-Aware Feature Learning for Person Reidentification
- Author
-
Xilin Chen, Ruibing Hou, Hong Chang, Xinqian Gu, Shiguang Shan, and Bingpeng Ma
- Subjects
Context model ,Source code ,Computer Networks and Communications ,business.industry ,Computer science ,media_common.quotation_subject ,Feature extraction ,02 engineering and technology ,Machine learning ,computer.software_genre ,Convolutional neural network ,Computer Science Applications ,Visualization ,Text mining ,Categorization ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Feature learning ,computer ,Software ,media_common - Abstract
Person reidentification (reID) by convolutional neural network (CNN)-based networks has achieved favorable performance in recent years. However, most of existing CNN-based methods do not take full advantage of spatial–temporal context modeling. In fact, the global spatial–temporal context can greatly clarify local distractions to enhance the target feature representation. To comprehensively leverage the spatial–temporal context information, in this work, we present a novel block, interaction–aggregation-update (IAU), for high-performance person reID. First, the spatial–temporal IAU (STIAU) module is introduced. STIAU jointly incorporates two types of contextual interactions into a CNN framework for target feature learning. Here, the spatial interactions learn to compute the contextual dependencies between different body parts of a single frame, while the temporal interactions are used to capture the contextual dependencies between the same body parts across all frames. Furthermore, a channel IAU (CIAU) module is designed to model the semantic contextual interactions between channel features to enhance the feature representation, especially for small-scale visual cues and body parts. Therefore, the IAU block enables the feature to incorporate the globally spatial, temporal, and channel context. It is lightweight, end-to-end trainable, and can be easily plugged into existing CNNs to form IAUnet. The experiments show that IAUnet performs favorably against state of the art on both image and video reID tasks and achieves compelling results on a general object categorization task. The source code is available at https://github.com/blue-blue272/ImgReID-IAnet .
- Published
- 2021
49. Distributed Device-to-Device Offloading System: Design and Performance Optimization
- Author
-
Sangheon Pack and Haneul Ko
- Subjects
Computer Networks and Communications ,business.industry ,Computer science ,ComputerSystemsOrganization_COMPUTER-COMMUNICATIONNETWORKS ,Stochastic game ,Mobile computing ,020206 networking & telecommunications ,02 engineering and technology ,Task (project management) ,symbols.namesake ,Nash equilibrium ,Best response ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,symbols ,Resource allocation ,Electrical and Electronic Engineering ,business ,Game theory ,Software ,Computer network - Abstract
In task offloading systems, it is imperative to guarantee that an offloaded task is completed within a pre-specified deadline. In this paper, we propose a distributed device-to-device (D2D) offloading system (DDOS) in which a task owner opportunistically broadcasts an offloading request that includes its mobility level and task completion deadline. After receiving the request, mobile devices in the vicinity of the task owner employ a constraint stochastic game to decide, in a distributed manner, whether to accept the request or not. We devise a best response dynamics-based algorithm (BRDA) to obtain a multi-policy constrained Nash equilibrium. Evaluation results demonstrate that DDOS can guarantee a high on-time task completion probability, as well as a low energy consumption.
- Published
- 2021
50. dmTP: A Deep Meta-Learning Based Framework for Mobile Traffic Prediction
- Author
-
Fuyou Li, Xiaoli Chu, Jie Zhang, Yuguang Fang, and Zitian Zhang
- Subjects
Structure (mathematical logic) ,Meta learning (computer science) ,business.industry ,Computer science ,Deep learning ,020206 networking & telecommunications ,02 engineering and technology ,Machine learning ,computer.software_genre ,Computer Science Applications ,Time–frequency analysis ,Task (project management) ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Artificial intelligence ,Electrical and Electronic Engineering ,Time series ,business ,computer ,Predictive modelling - Abstract
In the big data era, deep learning technologies have been widely exploited to mine increasingly available traffic data for mobile traffic prediction. Proactive management and optimization of mobile network resources for various wireless services require accurate mobile traffic prediction on different time and spacial scales. However, training deep learning models for different traffic prediction tasks individually is not only time consuming but also sometimes unrealistic as there are not always sufficient historical traffic records available. In this paper, we propose a novel mobile traffic prediction framework based on deep meta-learning (MTPFoDML), which can adaptively learn to learn the proper prediction model for each distinct prediction task from accumulated meta-knowledge of previous prediction tasks. In MTPFoDML, we regard each mobile traffic prediction task as a base-task and adopt a long short-term memory (LSTM) network with a fixed structure as the base-learner for each base-task. By transforming real-world mobile traffic data into the frequency domain, we find that the five main frequency components can characterize the mobile traffic variation over hours, days, and weeks, hence can be used as meta-features of a base-task. We employ a multi-layer perceptron (MLP) as the meta-learner to find the optimal super-parameter value and initial training status for the base-learner of each new base-task according to its meta-features, thus improving the base-learner's prediction accuracy and learning efficiency. Extensive experiments using real-world mobile traffic datasets demonstrate that our framework outperforms the existing prediction models for the same size of base-task training sets. Moreover, while guaranteeing a similar or even better prediction accuracy, meta-learning in MTPFoDML can lead to about 75% and 81%reduction in epoches and base-samples needed to train the base-learners, respectively.
- Published
- 2021
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.