268 results on '"Zhou, Joey Tianyi"'
Search Results
252. Structured AutoEncoders for Subspace Clustering
- Author
-
Peng, Xi, primary, Feng, Jiashi, additional, Xiao, Shijie, additional, Yau, Wei-Yun, additional, Zhou, Joey Tianyi, additional, and Yang, Songfan, additional
- Published
- 2018
- Full Text
- View/download PDF
253. SC2Net: Sparse LSTMs for Sparse Coding
- Author
-
Zhou, Joey Tianyi, primary, Di, Kai, additional, Du, Jiawei, additional, Peng, Xi, additional, Yang, Hao, additional, Pan, Sinno Jialin, additional, Tsang, Ivor, additional, Liu, Yong, additional, Qin, Zheng, additional, and Goh, Rick Siow Mong, additional
- Published
- 2018
- Full Text
- View/download PDF
254. ‘Who Likes What and, Why?’ Insights into Modeling Users’ Personality Based on Image ‘Likes’
- Author
-
Guntuku, Sharath Chandra, primary, Zhou, Joey Tianyi, additional, Roy, Sujoy, additional, Lin, Weisi, additional, and Tsang, Ivor W., additional
- Published
- 2018
- Full Text
- View/download PDF
255. MIML-FCN+: Multi-Instance Multi-Label Learning via Fully Convolutional Networks with Privileged Information
- Author
-
Yang, Hao, primary, Zhou, Joey Tianyi, additional, Cai, Jianfei, additional, and Ong, Yew Soon, additional
- Published
- 2017
- Full Text
- View/download PDF
256. RoSeq: Robust Sequence Labeling.
- Author
-
Zhou, Joey Tianyi, Zhang, Hao, Jin, Di, Peng, Xi, Xiao, Yang, and Cao, Zhiguo
- Subjects
- *
LABELS , *RANDOM fields , *DRUG labeling , *NOISE measurement , *MARKOV processes , *TASK analysis , *ROBUST control - Abstract
In this paper, we mainly investigate two issues for sequence labeling, namely, label imbalance and noisy data that are commonly seen in the scenario of named entity recognition (NER) and are largely ignored in the existing works. To address these two issues, a new method termed robust sequence labeling (RoSeq) is proposed. Specifically, to handle the label imbalance issue, we first incorporate label statistics in a novel conditional random field (CRF) loss. In addition, we design an additional loss to reduce the weights of overwhelming easy tokens for augmenting the CRF loss. To address the noisy training data, we adopt an adversarial training strategy to improve model generalization. In experiments, the proposed RoSeq achieves the state-of-the-art performances on CoNLL and English Twitter NER—88.07% on CoNLL-2002 Dutch, 87.33% on CoNLL-2002 Spanish, 52.94% on WNUT-2016 Twitter, and 43.03% on WNUT-2017 Twitter without using the additional data. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
257. Understanding Deep Representations Learned in Modeling Users Likes
- Author
-
Guntuku, Sharath Chandra, primary, Zhou, Joey Tianyi, additional, Roy, Sujoy, additional, Lin, Weisi, additional, and Tsang, Ivor W., additional
- Published
- 2016
- Full Text
- View/download PDF
258. Exploit Bounding Box Annotations for Multi-Label Object Recognition
- Author
-
Yang, Hao, primary, Zhou, Joey Tianyi, additional, Zhang, Yu, additional, Gao, Bin-Bin, additional, Wu, Jianxin, additional, and Cai, Jianfei, additional
- Published
- 2016
- Full Text
- View/download PDF
259. Good practices on building effective CNN baseline model for person re-identification
- Author
-
Li, Chunming, Yu, Hui, Pan, Zhigeng, Pu, Yifei, Xiong, Fu, Xiao, Yang, Cao, Zhiguo, Gong, Kaicheng, Fang, Zhiwen, and Zhou, Joey Tianyi
- Published
- 2019
- Full Text
- View/download PDF
260. Retinal multi-lesion segmentation by reinforcing single-lesion guidance with multi-view learning.
- Author
-
Zhang, Liyun, Fang, Zhiwen, Li, Ting, Xiao, Yang, Zhou, Joey Tianyi, and Yang, Feng
- Subjects
RETINAL blood vessels ,INCORPORATION - Abstract
The primary prerequisite for multi-lesion segmentation is the simultaneous detection of multiple lesions. Numerous techniques based on the frameworks of simultaneous multi-lesion segmentation and repetitive single-lesion segmentation are continually refining their models to meet this requirement. While simultaneous multi-lesion segmentation techniques may fully exploit interactions between distinct lesions, single-lesion segmentation methods can concentrate only on a particular lesion. However, since different lesions exhibit distinct patterns, it is challenging for these sophisticated models to work properly when several lesions are present simultaneously. We propose a retinal multi-lesion segmentation method by reinforcing single-lesion guidance with multi-view learning. To the best of our knowledge, this is the first work to formulate the retinal multi-lesion segmentation task as a multi-view task. In the multi-view method, each segmentation branch incorporates context guidance of the particular lesion at the input to focus more attention on specific lesions while producing multi-lesion outcomes. For implementing the multi-view method, we design a two-level hierarchical heterogeneous network, whose core element is the multi-view segmentation branches at the second level. The first level provides the context guidance as multi-view cues. Concretely, our network incorporates two complementary models (i.e., U-Net and TransU-Net). Compared with U-Net focusing on local receptive fields, TransU-Net augments U-Net by Transformer, which is adept at establishing long-distance dependence through global attention. The entire network consists of TransU-Net at the first level and a subsequent multi-view U-Net group. It is denoted as Trans2U-Net. Extensive experiments on three datasets demonstrate the effectiveness of the proposed method. • This study treats the retinal multi-lesion segmentation task as a multi-view task. • Each branch of multi-lesion segmentation pays more attention to a particular lesion. • A two-level hierarchical heterogeneous network is designed for context guidance. • Experiments on three public datasets verify the effectiveness of the proposed method. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
261. Deep Representations to Model User `Likes'.
- Author
-
Guntuku, Sharath Chandra, Zhou, Joey Tianyi, Roy, Sujoy, Weisi, Lin, and Tsang, Ivor W.
- Published
- 2015
- Full Text
- View/download PDF
262. SAR: Sharpness-Aware minimization for enhancing DNNs' Robustness against bit-flip errors.
- Author
-
Zhou, Changbao, Du, Jiawei, Yan, Ming, Yue, Hengshan, Wei, Xiaohui, and Zhou, Joey Tianyi
- Subjects
- *
ARTIFICIAL neural networks - Abstract
As Deep Neural Networks (DNNs) are increasingly deployed in safety-critical scenarios, there is a growing need to address bit-flip errors occurring in hardware, such as memory. These errors can lead to changes in DNN weights, potentially degrading the performance of deployed models and causing catastrophic consequences. Existing methods improve DNNs' fault tolerance or robustness by modifying network size, structure, or inference and training processes. Unfortunately, these methods often enhance robustness at the expense of clean accuracy and introduce additional overhead during inference. To address these issues, we propose S harpness- A ware Minimization for enhancing DNNs' R obustness against bit-flip errors (SAR), which aims to leverage the intrinsic robustness of DNNs. We begin with a comprehensive investigation of DNNs under bit-flip errors, yielding insightful observations regarding the intensity and occurrence of such errors. Based on these insights, we identify that Sharpness-Aware Minimization (SAM) has the potential to enhance DNN robustness. We further analyze this potential through the relationship between SAM formulation and our observations, building a robustness-enhancing framework based on SAM. Experimental validation across various models and datasets demonstrates that SAR can effectively improve DNN robustness against bit-flip errors without sacrificing clean accuracy or introducing additional inference costs, making it a "double-win" method compared to existing approaches. • The first work to uncover the DNNs' intrinsic robustness against bit-flip errors. • The first work to adopt Sharpness-Aware Minimization to resist bit-flip errors. • A valuable and lightweight framework for security-critical scenarios. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
263. Blessing few-shot segmentation via semi-supervised learning with noisy support images.
- Author
-
Zhang, Runtong, Zhu, Hongyuan, Zhang, Hanwang, Gong, Chen, Zhou, Joey Tianyi, and Meng, Fanman
- Subjects
- *
IMAGE segmentation , *CAUSAL models , *SUPERVISED learning , *STRUCTURAL models , *SEARCH engines , *PROBLEM solving , *CAUSAL inference - Abstract
Mainstream few-shot segmentation methods meet performance bottleneck due to the data scarcity of novel classes with insufficient intra-class variations, which results in a biased model primarily favoring the base classes. Fortunately, owing to the evolution of the Internet, an extensive repository of unlabeled images has become accessible from diverse sources such as search engines and publicly available datasets. However, such unlabeled images are not a free lunch. There are noisy inter-class and intra-class samples causing severe feature bias and performance degradation. Therefore, we propose a semi-supervised few-shot segmentation framework named F4S , which incorporates a ranking algorithm designed to eliminate noisy samples and select superior pseudo-labeled images, thereby fostering the improvement of few-shot segmentation within a semi-supervised paradigm. The proposed F4S framework can not only enrich the intra-class variations of novel classes during the test phase, but also enhance meta-learning of the network during the training phase. Furthermore, it can be readily implemented with ease on any off-the-shelf few-shot segmentation methods. Additionally, based on a Structural Causal Model (SCM), we further theoretically explain why the proposed method can solve the noise problem: the severe noise effects are removed by cutting off the backdoor path between pseudo labels and noisy support images via causal intervention. On PASCAL-5 i and COCO-20 i datasets, we show that the proposed F4S can boost various popular few-shot segmentation methods to new state-of-the-art performances. [Display omitted] • Incorporating semi-supervised learning into few-shot task to tackle data scarcity. • A ranking algorithm to identify and remove noisy samples in pseudo labels. • Explaining the algorithm using a Structural Causal Model to reduce confounding bias. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
264. Multi-spectral template matching based object detection in a few-shot learning manner.
- Author
-
Feng, Chen, Cao, Zhiguo, Xiao, Yang, Fang, Zhiwen, and Zhou, Joey Tianyi
- Subjects
- *
OBJECT recognition (Computer vision) , *KNOWLEDGE management , *FEATURE selection , *KNOWLEDGE transfer , *ONLINE education , *LEARNING - Abstract
Multi-spectral template matching (MSTM) based object detection approaches can be widely used in robotics and aerospace systems for fine-grained object discovery. However, the performance of existing nearest neighbor search based nonparametric paradigms (e.g. , correlation coefficient and l p -norm) turns out to be unsatisfactory. These paradigms tend to suffer from two defects: 1) they fail to select from the raw features the discriminative ones that can help distinguish between the target and background; 2) the domain shift between the template and search spectra has not been well addressed within the feature space. In this work, we propose a data-driven MSTM method to address these two issues. First, Exemplar-SVM (E-SVM) is applied to execute feature selection and target/background categorization jointly, which is facilitated by its max-margin mechanism. To enable the learning process where the template is regarded as a single positive sample, knowledge transfer is executed to attain negative samples from other domains, e.g. , large-scale public datasets. Then, the hard negative samples are mined to help train a discriminative classifier. Concerning practical applications, we also augment the template with different image degradations and extend E-SVM from the original one-shot learning approach to its few-shot version. Second, a multi-domain adaptation approach via unsupervised multi-domain subspace alignment is proposed to tackle multi-domain shift problem. Here the multiple domains relate to template, search, and negative ones considering both offline learning and online matching. The wide-range experimental results on two multi-spectral datasets demonstrate the effectiveness of our method. The tailored dataset and code will be released publicly. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
265. A concise but high-performing network for image guided depth completion in autonomous driving.
- Author
-
Liu, Moyun, Chen, Bing, Chen, Youping, Xie, Jingming, Yao, Lei, Zhang, Yang, and Zhou, Joey Tianyi
- Abstract
Depth completion is a crucial task in autonomous driving, aiming to convert a sparse depth map into a dense depth prediction. The sparse depth map serves as a partial reference for the actual depth, and the fusion of RGB images is frequently employed to augment the completion process owing to its inherent richness in semantic information. Image-guided depth completion confronts three principal challenges: (1) the effective fusion of the two modalities; (2) the enhancement of depth information recovery; and (3) the realization of real-time predictive capabilities requisite for practical autonomous driving scenarios. In response to these challenges, we propose a concise but high-performing network, named CHNet, to achieve high-performance depth completion with an elegant and straightforward architecture. Firstly, we use a fast guidance module to fuse the two sensor features, harnessing abundant auxiliary information derived from the color space. Unlike the prevalent complex guidance modules, our approach adopts an intuitive and cost-effective strategy. In addition, we find and analyze the optimization inconsistency problem for observed and unobserved positions. To mitigate this challenge, we introduce a decoupled depth prediction head, tailored to better discern and predict depth values for both valid and invalid positions, incurring minimal additional inference time. Capitalizing on the dual-encoder and single-decoder architecture, the simplicity of CHNet facilitates an optimal balance between accuracy and computational efficiency. In benchmark evaluations on the KITTI depth completion dataset, CHNet demonstrates competitive performance metrics and inference speeds relative to contemporary state-of-the-art methodologies. To assess the generalizability of our approach, we extend our evaluations to the indoor NYUv2 dataset, where CHNet continues to yield impressive outcomes. The code of this work will be available at https://github.com/lmomoy/CHNet. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
266. A principled framework for explainable multimodal disentanglement.
- Author
-
Han, Zongbo, Luo, Tao, Fu, Huazhu, Hu, Qinghua, Zhou, Joey Tianyi, and Zhang, Changqing
- Subjects
- *
MACHINE learning , *INFORMATION sharing - Abstract
Learning effective representations for data from multiple modalities is crucial in machine learning. Recent efforts focus on learning latent representations that integrate information from various modalities. These approaches generally assume simple or implicit relationships between different modalities and as a result are not able to accurately and explicitly depict the correlations among these modalities and lack explainability. To address this, we propose definitions and conditions for unsupervised multimodal disentanglement, offering guidelines for explicit disentanglement between modalities to enhance explainability. Furthermore, we have derived a novel objective function to explicitly separate multimodal data into components shared across modalities and components exclusive to each modality. The explicit guaranteed disentanglement is of great potential for downstream tasks. Benefiting from a cleverly designed network structure, we can visualize these disentangled representations, providing intuitive explainability. Experiments on a variety of multimodal datasets demonstrate that our objective can effectively disentangle information from different modalities while satisfying the disentangling conditions. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
267. TC-SEPM: Characterizing soft error resilience of CNNs on Tensor Cores from program and microarchitecture perspectives.
- Author
-
Wei, Xiaohui, Zhou, Changbao, Yue, Hengshan, and Zhou, Joey Tianyi
- Subjects
- *
SOFT errors , *CONVOLUTIONAL neural networks , *MACHINE learning - Abstract
As an architectural CNN accelerator integrated into NVIDIA's GPUs, existing research mainly focuses on improving the performance of Tensor Cores. However, the highly integrated Tensor Cores are vulnerable to transient faults (i.e., soft errors), causing catastrophic consequences in safety-critical applications like automatic driving. Thus, it is imperative to estimate the reliability of CNNs on Tensor Cores. However, obtaining a statistically significant resilience profile of CNNs on Tensor Cores with the existing fault injection (FI)-based reliability estimation methods is expensive. To this end, we build TC-SEPM to predict the error resilience of CNNs on Tensor Cores instead of FI methods. To ensure the accuracy of TC-SEPM, we first investigate resilience-related features from program and microarchitecture perspectives. Then, leveraging these heuristic features, we train machine learning models to learn the hidden relationship between error resilience and the investigated features, enabling us to predict the impact of soft errors in Tensor Cores on CNN output. Experimental results show that TC-SEPM achieves high accuracy for individual soft error resiliency prediction and overall program resilience estimation while its overhead is only 1/27 of FI methods. Additionally, TC-SEPM can provide valuable insights for programmers or architects to design more robust CNN models on Tensor Cores. • The first work to predict the soft error resilience of CNNs on Tensor Cores. • The first work to make a hardware–software co-exploration on reliability-relevant features. • A valuable tool for programmers or architects due to its high efficiency and accuracy on soft error reliability evaluation. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
268. Contrastive domain adaptation with consistency match for automated pneumonia diagnosis.
- Author
-
Feng, Yangqin, Wang, Zizhou, Xu, Xinxing, Wang, Yan, Fu, Huazhu, Li, Shaohua, Zhen, Liangli, Lei, Xiaofeng, Cui, Yingnan, Sim Zheng Ting, Jordan, Ting, Yonghan, Zhou, Joey Tianyi, Liu, Yong, Siow Mong Goh, Rick, and Heng Tan, Cher
- Subjects
- *
ARTIFICIAL neural networks , *PNEUMONIA , *X-ray imaging , *COVID-19 testing , *MEDICAL screening - Abstract
Pneumonia can be difficult to diagnose since its symptoms are too variable, and the radiographic signs are often very similar to those seen in other illnesses such as a cold or influenza. Deep neural networks have shown promising performance in automated pneumonia diagnosis using chest X-ray radiography, allowing mass screening and early intervention to reduce the severe cases and death toll. However, they usually require many well-labelled chest X-ray images for training to achieve high diagnostic accuracy. To reduce the need for training data and annotation resources, we propose a novel method called Contrastive Domain Adaptation with Consistency Match (CDACM). It transfers the knowledge from different but relevant datasets to the unlabelled small-size target dataset and improves the semantic quality of the learnt representations. Specifically, we design a conditional domain adversarial network to exploit discriminative information conveyed in the predictions to mitigate the domain gap between the source and target datasets. Furthermore, due to the small scale of the target dataset, we construct a feature cloud for each target sample and leverage contrastive learning to extract more discriminative features. Lastly, we propose adaptive feature cloud expansion to push the decision boundary to a low-density area. Unlike most existing transfer learning methods that aim only to mitigate the domain gap, our method instead simultaneously considers the domain gap and the data deficiency problem of the target dataset. The conditional domain adaptation and the feature cloud generation of our method are learning jointly to extract discriminative features in an end-to-end manner. Besides, the adaptive feature cloud expansion improves the model's generalisation ability in the target domain. Extensive experiments on pneumonia and COVID-19 diagnosis tasks demonstrate that our method outperforms several state-of-the-art unsupervised domain adaptation approaches, which verifies the effectiveness of CDACM for automated pneumonia diagnosis using chest X-ray imaging. • A novel domain adaptation method for small-scale target dataset. • A contrastive learning-based strategy to solve data deficiency problem. • An adaptive feature cloud expansion mechanism to improve model's generalisation ability. • A new SOTA result for automated pneumonia and COVID-19 diagnosis. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.