60 results on '"Ngoc Trung, Tran"'
Search Results
2. Revisiting Label Smoothing and Knowledge Distillation Compatibility: What was Missing?
- Author
-
Keshigeyan Chandrasegaran, Ngoc-Trung Tran, Yunqing Zhao, and Ngai-Man Cheung
- Published
- 2022
3. A Closer Look at Fourier Spectrum Discrepancies for CNN-Generated Images Detection.
- Author
-
Keshigeyan Chandrasegaran, Ngoc-Trung Tran, and Ngai-Man Cheung
- Published
- 2021
- Full Text
- View/download PDF
4. InfoMax-GAN: Improved Adversarial Image Generation via Information Maximization and Contrastive Learning.
- Author
-
Kwot Sin Lee, Ngoc-Trung Tran, and Ngai-Man Cheung
- Published
- 2021
- Full Text
- View/download PDF
5. On Data Augmentation for GAN Training.
- Author
-
Ngoc-Trung Tran, Viet-Hung Tran, Ngoc-Bao Nguyen, Trung-Kien Nguyen, and Ngai-Man Cheung
- Published
- 2021
- Full Text
- View/download PDF
6. Self-supervised GAN: Analysis and Improvement with Multi-class Minimax Game.
- Author
-
Ngoc-Trung Tran, Viet-Hung Tran, Ngoc-Bao Nguyen, Linxiao Yang, and Ngai-Man Cheung
- Published
- 2019
7. Improving GAN with Neighbors Embedding and Gradient Matching.
- Author
-
Ngoc-Trung Tran, Tuan-Anh Bui, and Ngai-Man Cheung
- Published
- 2019
- Full Text
- View/download PDF
8. DOPING: Generative Data Augmentation for Unsupervised Anomaly Detection with GAN.
- Author
-
Swee Kiat Lim, Yi Loo, Ngoc-Trung Tran, Ngai-Man Cheung, Gemma Roig, and Yuval Elovici
- Published
- 2018
- Full Text
- View/download PDF
9. Dist-GAN: An Improved GAN Using Distance Constraints.
- Author
-
Ngoc-Trung Tran, Tuan-Anh Bui, and Ngai-Man Cheung
- Published
- 2018
- Full Text
- View/download PDF
10. On-Device Scalable Image-Based Localization via Prioritized Cascade Search and Fast One-Many RANSAC.
- Author
-
Ngoc-Trung Tran, Dang-Khoa Le Tan, Anh-Dzung Doan, Thanh-Toan Do, Tuan-Anh Bui, Mengxuan Tan, and Ngai-Man Cheung
- Published
- 2019
- Full Text
- View/download PDF
11. Challenging 3D Head Tracking and Evaluation Using Unconstrained Test Data Set.
- Author
-
Fakhreddine Ababsa, Ngoc-Trung Tran, and Maurice Charbit
- Published
- 2017
- Full Text
- View/download PDF
12. InfoMax-GAN: Improved Adversarial Image Generation via Information Maximization and Contrastive Learning.
- Author
-
Kwot Sin Lee, Ngoc-Trung Tran, and Ngai-Man Cheung
- Published
- 2020
13. Towards Good Practices for Data Augmentation in GAN Training.
- Author
-
Ngoc-Trung Tran, Viet-Hung Tran, Ngoc-Bao Nguyen, Trung-Kien Nguyen, and Ngai-Man Cheung
- Published
- 2020
14. A Non-rigid Face Tracking Method for Wide Rotation Using Synthetic Data.
- Author
-
Ngoc-Trung Tran, Fakhreddine Ababsa, and Maurice Charbit
- Published
- 2015
- Full Text
- View/download PDF
15. Towards Pose-free Tracking of Non-rigid Face using Synthetic Data.
- Author
-
Ngoc-Trung Tran, Fakhreddine Ababsa, and Maurice Charbit
- Published
- 2015
16. Cascaded Regressions of Learning Features for Face Alignment.
- Author
-
Ngoc-Trung Tran, Fakhreddine Ababsa, Sarra Ben Fredj, and Maurice Charbit
- Published
- 2015
- Full Text
- View/download PDF
17. U3PT: A New Dataset for Unconstrained 3D Pose Tracking Evaluation.
- Author
-
Ngoc-Trung Tran, Fakhreddine Ababsa, and Maurice Charbit
- Published
- 2015
- Full Text
- View/download PDF
18. 3D Face Pose and Animation Tracking via Eigen-Decomposition based Bayesian Approach.
- Author
-
Ngoc-Trung Tran, Fakhr-Eddine Ababsa, Maurice Charbit, Jacques Feldmar, Dijana Petrovska-Delacrétaz, and Gérard Chollet
- Published
- 2019
19. An Improved Self-supervised GAN via Adversarial Training.
- Author
-
Ngoc-Trung Tran, Viet-Hung Tran, Ngoc-Bao Nguyen, and Ngai-Man Cheung
- Published
- 2019
20. 3D Face Pose and Animation Tracking via Eigen-Decomposition Based Bayesian Approach.
- Author
-
Ngoc-Trung Tran, Fakhr-Eddine Ababsa, Maurice Charbit, Jacques Feldmar, Dijana Petrovska-Delacrétaz, and Gérard Chollet
- Published
- 2013
- Full Text
- View/download PDF
21. Combining Descriptors Extracted from Feature Maps of Deconvolutional Networks and SIFT Descriptors in Scene Image Classification.
- Author
-
Dung A. Doan, Ngoc-Trung Tran, Dinh-Phong Vo, Bac Le, and Atsuo Yoshitaka
- Published
- 2013
- Full Text
- View/download PDF
22. Learned and designed features for sparse coding in image classification.
- Author
-
Dung A. Doan, Ngoc-Trung Tran, Dinh-Phong Vo, and Bac Le
- Published
- 2013
- Full Text
- View/download PDF
23. 3D Face Pose Tracking from Monocular Camera via Sparse Representation of Synthesized Faces.
- Author
-
Ngoc-Trung Tran, Jacques Feldmar, Maurice Charbit, Dijana Petrovska-Delacrétaz, and Gérard Chollet
- Published
- 2013
24. Improving GAN with neighbors embedding and gradient matching.
- Author
-
Ngoc-Trung Tran, Tuan-Anh Bui, and Ngai-Man Cheung
- Published
- 2018
25. Generative Adversarial Autoencoder Networks.
- Author
-
Ngoc-Trung Tran, Tuan-Anh Bui, and Ngai-Man Cheung
- Published
- 2018
26. On-device Scalable Image-based Localization.
- Author
-
Ngoc-Trung Tran, Dang-Khoa Le Tan, Anh-Dzung Doan, Thanh-Toan Do, Tuan-Anh Bui, and Ngai-Man Cheung
- Published
- 2018
27. A robust framework for tracking simultaneously rigid and non-rigid face using synthesized data.
- Author
-
Ngoc-Trung Tran, Fakhreddine Ababsa, and Maurice Charbit
- Published
- 2015
- Full Text
- View/download PDF
28. On-Device Scalable Image-Based Localization via Prioritized Cascade Search and Fast One-Many RANSAC
- Author
-
Mengxuan Tan, Anh-Dzung Doan, Ngai-Man Cheung, Thanh-Toan Do, Ngoc-Trung Tran, Dang-Khoa Le Tan, and Tuan-Anh Bui
- Subjects
FOS: Computer and information sciences ,business.industry ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,Hash function ,Computer Science - Computer Vision and Pattern Recognition ,02 engineering and technology ,RANSAC ,Computer Graphics and Computer-Aided Design ,Scalability ,0202 electrical engineering, electronic engineering, information engineering ,Benchmark (computing) ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business ,Pose ,Image retrieval ,Software - Abstract
We present the design of an entire on-device system for large-scale urban localization using images. The proposed design integrates compact image retrieval and 2D-3D correspondence search to estimate the location in extensive city regions. Our design is GPS agnostic and does not require network connection. In order to overcome the resource constraints of mobile devices, we propose a system design that leverages the scalability advantage of image retrieval and accuracy of 3D model-based localization. Furthermore, we propose a new hashing-based cascade search for fast computation of 2D-3D correspondences. In addition, we propose a new one-many RANSAC for accurate pose estimation. The new one-many RANSAC addresses the challenge of repetitive building structures (e.g. windows and balconies) in urban localization. Extensive experiments demonstrate that our 2D-3D correspondence search achieves the state-of-the-art localization accuracy on multiple benchmark datasets. Furthermore, our experiments on a large Google street view image dataset show the potential of large-scale localization entirely on a typical mobile device.
- Published
- 2019
- Full Text
- View/download PDF
29. A Bifunctional Copper Catalyst Enables Ester Reduction with H2: Expanding the Reactivity Space of Nucleophilic Copper Hydrides
- Author
-
Zimmermann, Birte M., primary, Ngoc, Trung Tran, additional, Tzaras, Dimitrios-Ioannis, additional, Kaicharla, Trinadh, additional, and Teichert, Johannes F., additional
- Published
- 2021
- Full Text
- View/download PDF
30. On Data Augmentation for GAN Training
- Author
-
Viet-Hung Tran, Ngoc-Trung Tran, Ngai-Man Cheung, Trung-Kien Nguyen, and Ngoc-Bao Nguyen
- Subjects
FOS: Computer and information sciences ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,Image and Video Processing (eess.IV) ,Computer Science - Computer Vision and Pattern Recognition ,02 engineering and technology ,Electrical Engineering and Systems Science - Image and Video Processing ,computer.software_genre ,Computer Graphics and Computer-Aided Design ,Data modeling ,Original data ,0202 electrical engineering, electronic engineering, information engineering ,Code (cryptography) ,Task analysis ,FOS: Electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Data mining ,Divergence (statistics) ,computer ,Software ,Generator (mathematics) - Abstract
Recent successes in Generative Adversarial Networks (GAN) have affirmed the importance of using more data in GAN training. Yet it is expensive to collect data in many domains such as medical applications. Data Augmentation (DA) has been applied in these applications. In this work, we first argue that the classical DA approach could mislead the generator to learn the distribution of the augmented data, which could be different from that of the original data. We then propose a principled framework, termed Data Augmentation Optimized for GAN (DAG), to enable the use of augmented data in GAN training to improve the learning of the original distribution. We provide theoretical analysis to show that using our proposed DAG aligns with the original GAN in minimizing the Jensen-Shannon (JS) divergence between the original distribution and model distribution. Importantly, the proposed DAG effectively leverages the augmented data to improve the learning of discriminator and generator. We conduct experiments to apply DAG to different GAN models: unconditional GAN, conditional GAN, self-supervised GAN and CycleGAN using datasets of natural images and medical images. The results show that DAG achieves consistent and considerable improvements across these models. Furthermore, when DAG is used in some GAN models, the system establishes state-of-the-art Frechet Inception Distance (FID) scores. Our code is available., Comment: Accepted in IEEE Transactions on Image Processing
- Published
- 2021
31. InfoMax-GAN: Improved Adversarial Image Generation via Information Maximization and Contrastive Learning
- Author
-
Ngai-Man Cheung, Kwot Sin Lee, and Ngoc-Trung Tran
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Discriminator ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,Computer Science - Computer Vision and Pattern Recognition ,Machine Learning (stat.ML) ,02 engineering and technology ,010501 environmental sciences ,Machine learning ,computer.software_genre ,01 natural sciences ,Machine Learning (cs.LG) ,Statistics - Machine Learning ,0202 electrical engineering, electronic engineering, information engineering ,Code (cryptography) ,Infomax ,0105 earth and related environmental sciences ,Hyperparameter ,Forgetting ,business.industry ,Maximization ,Mutual information ,Range (mathematics) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer - Abstract
While Generative Adversarial Networks (GANs) are fundamental to many generative modelling applications, they suffer from numerous issues. In this work, we propose a principled framework to simultaneously mitigate two fundamental issues in GANs: catastrophic forgetting of the discriminator and mode collapse of the generator. We achieve this by employing for GANs a contrastive learning and mutual information maximization approach, and perform extensive analyses to understand sources of improvements. Our approach significantly stabilizes GAN training and improves GAN performance for image synthesis across five datasets under the same training and evaluation conditions against state-of-the-art works. In particular, compared to the state-of-the-art SSGAN, our approach does not suffer from poorer performance on image domains such as faces, and instead improves performance significantly. Our approach is simple to implement and practical: it involves only one auxiliary objective, has a low computational cost, and performs robustly across a wide range of training settings and datasets without any hyperparameter tuning. For reproducibility, our code is available in Mimicry: https://github.com/kwotsin/mimicry., Comment: Accepted to WACV 2021. An initial version was accepted to NeurIPS 2019 Workshop on Information Theory and Machine Learning
- Published
- 2020
- Full Text
- View/download PDF
32. Dist-GAN: An Improved GAN Using Distance Constraints
- Author
-
Ngai-Man Cheung, Ngoc-Trung Tran, and Tuan-Anh Bui
- Subjects
Distance constraints ,Discriminator ,Computer science ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Sample (statistics) ,02 engineering and technology ,010501 environmental sciences ,01 natural sciences ,Algorithm ,Autoencoder ,MNIST database ,0105 earth and related environmental sciences - Abstract
We introduce effective training algorithms for Generative Adversarial Networks (GAN) to alleviate mode collapse and gradient vanishing. In our system, we constrain the generator by an Autoencoder (AE). We propose a formulation to consider the reconstructed samples from AE as “real” samples for the discriminator. This couples the convergence of the AE with that of the discriminator, effectively slowing down the convergence of discriminator and reducing gradient vanishing. Importantly, we propose two novel distance constraints to improve the generator. First, we propose a latent-data distance constraint to enforce compatibility between the latent sample distances and the corresponding data sample distances. We use this constraint to explicitly prevent the generator from mode collapse. Second, we propose a discriminator-score distance constraint to align the distribution of the generated samples with that of the real samples through the discriminator score. We use this constraint to guide the generator to synthesize samples that resemble the real ones. Our proposed GAN using these distance constraints, namely Dist-GAN, can achieve better results than state-of-the-art methods across benchmark datasets: synthetic, MNIST, MNIST-1K, CelebA, CIFAR-10 and STL-10 datasets. Our code is published here (https://github.com/tntrung/gan) for research.
- Published
- 2018
- Full Text
- View/download PDF
33. Improving GAN with neighbors embedding and gradient matching
- Author
-
Ngai-Man Cheung, Ngoc-Trung Tran, and Tuan-Anh Bui
- Subjects
FOS: Computer and information sciences ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,Scalar (mathematics) ,0202 electrical engineering, electronic engineering, information engineering ,Nonlinear dimensionality reduction ,Computer Science - Computer Vision and Pattern Recognition ,Embedding ,020201 artificial intelligence & image processing ,02 engineering and technology ,General Medicine ,Regularization (mathematics) ,Algorithm - Abstract
We propose two new techniques for training Generative Adversarial Networks (GANs). Our objectives are to alleviate mode collapse in GAN and improve the quality of the generated samples. First, we propose neighbor embedding, a manifold learning-based regularization to explicitly retain local structures of latent samples in the generated samples. This prevents generator from producing nearly identical data samples from different latent samples, and reduces mode collapse. We propose an inverse t-SNE regularizer to achieve this. Second, we propose a new technique, gradient matching, to align the distributions of the generated samples and the real samples. As it is challenging to work with high-dimensional sample distributions, we propose to align these distributions through the scalar discriminator scores. We constrain the difference between the discriminator scores of the real samples and generated ones. We further constrain the difference between the gradients of these discriminator scores. We derive these constraints from Taylor approximations of the discriminator function. We perform experiments to demonstrate that our proposed techniques are computationally simple and easy to be incorporated in existing systems. When Gradient matching and Neighbour embedding are applied together, our GN-GAN achieves outstanding results on 1D/2D synthetic, CIFAR-10 and STL-10 datasets, e.g. FID score of $30.80$ for the STL-10 dataset. Our code is available at: https://github.com/tntrung/gan, Comment: Published as a conference paper at AAAI 2019
- Published
- 2018
- Full Text
- View/download PDF
34. Challenging 3D Head Tracking and Evaluation Using Unconstrained Test Data Set
- Author
-
Maurice Charbit, Fakhreddine Ababsa, Ngoc-Trung Tran, Informatique, Biologie Intégrative et Systèmes Complexes (IBISC), Université d'Évry-Val-d'Essonne (UEVE), Laboratoire Traitement et Communication de l'Information (LTCI), and Télécom ParisTech-Institut Mines-Télécom [Paris] (IMT)-Centre National de la Recherche Scientifique (CNRS)
- Subjects
Landmark ,Biometrics ,Facial motion capture ,Computer science ,business.industry ,Face tracking ,Feature extraction ,3D pose estimation ,Test data sets ,020207 software engineering ,Context (language use) ,02 engineering and technology ,Face (geometry) ,0202 electrical engineering, electronic engineering, information engineering ,Cascaded regression ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business ,Evaluation ,[SPI.SIGNAL]Engineering Sciences [physics]/Signal and Image processing ,Test data - Abstract
International audience; 3D face tracking using one monocular camera is an important topic, since it is useful in many domains such as: video surveillance system, human machine interaction, biometrics, etc. In this paper, we propose a new 3D face tracking which is robust to large head rotations. Underlying cascaded regression approach for 2D landmark detection, we build an extension in context of 3D pose tracking. To better work with out-of-plane issues, we extend the training dataset by including a new set of synthetic images. For evaluation, we propose to use a new recording system to capture automatically face pose ground-truth, and create a new test dataset, named U3PT (Unconstrained 3D Pose Tracking). Theperformance of our method along with the state-of-the-art methods are carried out to analyze advantage as well as limitations need to be improved in the future.
- Published
- 2017
- Full Text
- View/download PDF
35. Samba, un système multi-agents pour la compréhension des dynamiques agraires
- Author
-
Castella, Jean-Christophe, Boissau, Stanislas, Ngoc Trung, Tran, and Dinh Quang, Dang
- Subjects
gestion de l’environnement ,fonctionnement de l’écosystème ,politique de l’environnement ,développement durable ,développement intégré ,plaine inondable ,exploitation des ressources naturelles ,irrigation ,environnement ,milieu deltaïque - Abstract
La réforme agraire qui fait suite à l’indépendance du Viêt-Nam met fin à la propriété individuelle des terres. La terre est déclarée propriété du peuple vietnamien et un système de coopératives est mis en place. Entre la fin des années 70 et le début des années 80, ce système rentre dans une phase de crise caractérisée par une baisse de la production rizicole et une démotivation des coopérateurs (Jésus et Dao Thê Anh, 1998). Deux réformes successives en 1982 (« résolution 100 ») et 1986 (« co...
- Published
- 2017
36. A robust framework for tracking simultaneously rigid and non-rigid face using synthesized data
- Author
-
Fakhreddine Ababsa, Maurice Charbit, Ngoc-Trung Tran, Laboratoire Traitement et Communication de l'Information (LTCI), Télécom ParisTech-Institut Mines-Télécom [Paris] (IMT)-Centre National de la Recherche Scientifique (CNRS), Informatique, Biologie Intégrative et Systèmes Complexes (IBISC), and Université d'Évry-Val-d'Essonne (UEVE)
- Subjects
Computer science ,Facial motion capture ,3D head tracking ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Scale-invariant feature transform ,Tracking (particle physics) ,Artificial Intelligence ,Computer vision ,Face matching ,Pose ,ComputingMethodologies_COMPUTERGRAPHICS ,Synthesized face ,Non-rigid tracking ,Pixel ,business.industry ,3D face tracking ,Animation ,Rigid tracking ,Active appearance model ,Face (geometry) ,Signal Processing ,Computer Vision and Pattern Recognition ,Artificial intelligence ,business ,Geometric modeling ,[SPI.SIGNAL]Engineering Sciences [physics]/Signal and Image processing ,Software - Abstract
Our method can track simultaneously rigid and non-rigid face using synthetic data.It is robust with fast movements, illumination, partly occlusion.It can recover in term of lost tracking without waiting for the frontal face reset.It can keep tracking the face up to profile (even 90?) and complex motions.It is comparable to state-of-the-art methods in terms of pose and landmark tracking. This paper presents a robust framework for simultaneously tracking rigid pose and non-rigid animation of a single face with a monocular camera. Our proposed method consists of two phases: training and tracking. In the training phase, using automatically detected landmarks and the three-dimensional face model Candide-3, we built a cohort of synthetic face examples with a large range of the three axial rotations. The representation of a face's appearance is a set of local patches of landmarks that are characterized by Scale Invariant Feature Transform (SIFT) descriptors. In the tracking phase, we propose an original approach combining geometric and appearance models. The purpose of the geometric model is to provide a SIFT baseline matching between the current frame and an adaptive set of keyframes for rigid parameter estimation. The appearance model uses nearest synthetic examples of the training set to re-estimate rigid and non-rigid parameters. We found a tracking capability up to 90? of vertical axial rotation, and our method is robust even in the presence of fast movements, illumination changes and tracking losses. Numerical results on the rigid and non-rigid parameter sets are reported using several annotated public databases. Compared to other published algorithms, our method provides an excellent compromise between rigid and non-rigid parameter accuracies. The approach has some potential, providing good pose estimation (average error less than 4? on the Boston University Face Tracking dataset) and landmark tracking precision (6.3?pixel error compared to 6.8 of one of state-of-the-art methods on Talking Face video).
- Published
- 2015
- Full Text
- View/download PDF
37. Cascaded Regressions of Learning Features for Face Alignment
- Author
-
Sarra Ben Fredj, Ngoc-Trung Tran, Maurice Charbit, Fakhreddine Ababsa, Laboratoire Traitement et Communication de l'Information (LTCI), Télécom ParisTech-Institut Mines-Télécom [Paris] (IMT)-Centre National de la Recherche Scientifique (CNRS), Informatique, Biologie Intégrative et Systèmes Complexes (IBISC), Université d'Évry-Val-d'Essonne (UEVE), UFE, Observatoire de Paris, and Université Paris sciences et lettres (PSL)-Université Paris sciences et lettres (PSL)
- Subjects
2d images ,Restricted Boltzmann machine ,Training set ,Computer science ,Facial motion capture ,business.industry ,Face tracking ,Learning feature ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Pattern recognition ,Synthetic data ,Regression ,[INFO.INFO-TS]Computer Science [cs]/Signal and Image Processing ,Face (geometry) ,Cascaded regression ,Computer vision ,Artificial intelligence ,State (computer science) ,business ,Face alignment - Abstract
International audience; Face alignment is a fundamental problem in computer vision to localize the landmarks of eyes, nose or mouth in 2D images. In this paper, our method for face alignment integrates three aspects not seen in previous approaches: First, learning local descriptors using Restricted Boltzmann Machine (RBM) to model the local appearance of each facial points independently. Second, proposing the coarse-to-fine regression to localize the landmarks after the estimation of the shape configuration via global regression. Third, and using synthetic data as training data to enable our approach to work better with the profile view, and to forego the need of increasing the number of annotations for training. Our results on challenging datasets compare favorably with state of the art results. The combination with the synthetic data allows our method yielding good results in profile alignment. That highlights the potential of using synthetic data for in-the-wild face alignment.
- Published
- 2015
- Full Text
- View/download PDF
38. U3PT: A New Dataset for Unconstrained 3D Pose Tracking Evaluation
- Author
-
Fakhreddine Ababsa, Ngoc-Trung Tran, Maurice Charbit, Informatique, Biologie Intégrative et Systèmes Complexes (IBISC), Université d'Évry-Val-d'Essonne (UEVE), Laboratoire Traitement et Communication de l'Information (LTCI), and Télécom ParisTech-Institut Mines-Télécom [Paris] (IMT)-Centre National de la Recherche Scientifique (CNRS)
- Subjects
Monocular ,Synthetic data ,Biometrics ,business.industry ,Facial motion capture ,Computer science ,Unconstrained pose tracking ,3D head tracking ,3D pose estimation ,Articulated body pose estimation ,Pose tracking dataset ,[INFO.INFO-TS]Computer Science [cs]/Signal and Image Processing ,Robustness (computer science) ,Computer vision ,Artificial intelligence ,business ,Pose ,3D pose tracking ,Pose estimation - Abstract
International audience; 3D pose tracking using monocular cameras is an important topic, which has been receiving a great attention since last decades. It is useful in many domains such as: Video Surveillance, Human-Computer Interface, Biometrics, etc. The problem gets much challenging if occurring, for example, fast motion, out-of-plane rotation, the illumination changes, expression, or occlusions. In the literature, there are some datasets reported for 3D pose tracking evaluation, however, all of them retains simple background, no-expression, slow motion, frontal rotation, or no-occlusion. It is not enough to test advances of in-the-wild tracking. Indeed, collecting accurate ground-truth of 3D pose is difficult because some special devices or sensors are required. In addition, the magnetic sensors usually used for 3D pose ground-truth, is uncomfortable to wear and move because of their wires. In this paper, we propose a new recording system that allows people move more comfortable. We create a new challenging dataset, named U3PT (Unconstrained 3D Pose Tracking). It could be considered as a benchmark to evaluate and compare the robustness and precision of state-of-the-art methods that aims to work in-the-wild. This paper will also present the performances of two well-known state-of-the-art methods compared to our method on face tracking when applied to this database. We have carried out several experiments and have reported advantages and some limitations to be improved in the future.
- Published
- 2015
- Full Text
- View/download PDF
39. Towards pose-free tracking of non-rigid face using synthetic data
- Author
-
Fakhreddine Ababsa, Ngoc-Trung Tran, Maurice Charbit, Informatique, Biologie Intégrative et Systèmes Complexes (IBISC), Université d'Évry-Val-d'Essonne (UEVE), Laboratoire Traitement et Communication de l'Information (LTCI), Télécom ParisTech-Institut Mines-Télécom [Paris] (IMT)-Centre National de la Recherche Scientifique (CNRS), and Davesne, Frédéric
- Subjects
Facial motion capture ,Computer science ,Scale-invariant feature transform ,Initialization ,[SCCO.COMP]Cognitive science/Computer science ,02 engineering and technology ,Tracking (particle physics) ,Synthetic data ,030218 nuclear medicine & medical imaging ,03 medical and health sciences ,0302 clinical medicine ,0202 electrical engineering, electronic engineering, information engineering ,Computer vision ,Face matching ,Pose ,[SPI.SIGNAL] Engineering Sciences [physics]/Signal and Image processing ,Synthesized face ,Non-rigid tracking ,business.industry ,3D face tracking ,Rigid tracking ,Support vector machine ,Face (geometry) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,[SPI.SIGNAL]Engineering Sciences [physics]/Signal and Image processing ,3D pose tracking - Abstract
International audience; The non-rigid face tracking has been achieved many advances in recent years, but most of empirical experiments are restricted at near-frontal face. This report introduces a robust framework for pose-free tracking of non-rigid face. Our method consists of two phases: training and tracking. In the training phase, a large offline synthesized database is built to train landmark appearance models using linear Support Vector Machine (SVM). In the tracking phase, a two-step approach is proposed: the first step, namely initialization, benefits 2D SIFT matching between the current frame and a set of adaptive keyframes to estimate the rigid parameters. The second step obtains the whole set of parameters (rigid and non-rigid) using a heuristic method via pose-wise SVMs. The combination of these aspects makes our method work robustly up to 90 degree of vertical axial rotation. Moreover, our method appears to be robust even in the presence of fast movements and tracking losses. Comparing to other published algorithms, our method offers a very good compromise of rigid and non-rigid parameter accuracies. This study gives a promising perspective because of the good results in terms of pose estimation (average error is less than 4 o on BUFT dataset) and landmark tracking precision (5.8 pixel error compared to 6.8 of one state-of-the-art method on Talking Face video). These results highlight the potential of using synthetic data to track non-rigid face in unconstrained poses.
- Published
- 2015
- Full Text
- View/download PDF
40. A Non-rigid Face Tracking Method for Wide Rotation Using Synthetic Data
- Author
-
Fakhreddine Ababsa, Ngoc-Trung Tran, Maurice Charbit, Laboratoire Traitement et Communication de l'Information (LTCI), Télécom ParisTech-Institut Mines-Télécom [Paris] (IMT)-Centre National de la Recherche Scientifique (CNRS), Informatique, Biologie Intégrative et Systèmes Complexes (IBISC), and Université d'Évry-Val-d'Essonne (UEVE)
- Subjects
Non-rigid tracking ,Computer science ,business.industry ,Facial motion capture ,Heuristic ,Initialization ,3D face tracking ,Rigid tracking ,Synthetic data ,Out-of-plane tracking ,Support vector machine ,[INFO.INFO-TS]Computer Science [cs]/Signal and Image Processing ,Face (geometry) ,Computer vision ,Artificial intelligence ,Face matching ,business ,Pose ,Rotation (mathematics) ,Synthesized face - Abstract
International audience; This paper propose a new method for wide-rotation non-rigid face tracking that is still a challenging problem in computer vision community. Our method consists of training and tracking phases. In training, we propose to use a large off-line synthetic database to overcome the problem of data collection. The local appearance models are then trained using linear Support Vector Machine (SVM). In tracking, we propose a two-step approach: (i) The first step uses baseline matching for a good initialization. The matching strategy between the current frame and a set of adaptive keyframes is also involved to be recoverable in terms of failed tracking. (ii) The second step estimates the model parameters using a heuristic method via pose-wise SVMs. The combination makes our approach work robustly with wide rotation, up to 90∘ of vertical axis. In addition, our method appears to be robust even in the presence of fast movements thanks to baseline matching. Compared to state-of-the-art methods, our method shows a good compromise of rigid and non-rigid parameter accuracies. This study gives a promising perspective because of the good results in terms of pose estimation (average error is less than 4o on BUFT dataset) and landmark tracking precision (5.8 pixel error compared to 6.8 of one state-of-the-art method on Talking Face video. These results highlight the potential of using synthetic data to track non-rigid face in unconstrained poses
- Published
- 2015
- Full Text
- View/download PDF
41. Learned and designed features for sparse coding in image classification
- Author
-
Dinh-Phong Vo, Dung A. Doan, Ngoc-Trung Tran, and Bac Le
- Subjects
Contextual image classification ,business.industry ,Computer science ,Deep learning ,Cognitive neuroscience of visual object recognition ,Scale-invariant feature transform ,Pattern recognition ,Translation (geometry) ,Machine learning ,computer.software_genre ,Haar-like features ,Artificial intelligence ,Neural coding ,business ,Rotation (mathematics) ,computer - Abstract
There is an amount of designed features (SIFT, SURF, or DAISY) which has been chosen in the standard implementation of some visual recognition and multimedia challenges. The power of these features lie on their invariance designed against rotation, scaling, and translation. Recent trends in deep learning, however, have pointed out that data-driven features learning performs better designed features in some tasks, since they can capture the global (via multi-layers network) or inter-local structures (convolutional network) of images. We argue that combining the two types of features can significantly improve visual object recognition performance. We propose in this paper a framework that uses sparse coding and the fusion of learned and designed features in order to build descriptive codewords. Evaluations on Caltech-101 and 15 Scenes validates our argument, with a better result compared with recent approaches.
- Published
- 2013
- Full Text
- View/download PDF
42. 3D face pose and animation tracking via eigen-decomposition based Bayesian approach
- Author
-
Dijana Petrovska-Delacrétaz, Ngoc-Trung Tran, Gérard Chollet, Fakhreddine Ababsa, Jacques Feldmar, Maurice Charbit, Laboratoire Traitement et Communication de l'Information (LTCI), Télécom ParisTech-Institut Mines-Télécom [Paris] (IMT)-Centre National de la Recherche Scientifique (CNRS), Département Traitement du Signal et des Images (TSI), Télécom ParisTech-Centre National de la Recherche Scientifique (CNRS), Informatique, Biologie Intégrative et Systèmes Complexes (IBISC), Université d'Évry-Val-d'Essonne (UEVE), Services répartis, Architectures, MOdélisation, Validation, Administration des Réseaux (SAMOVAR), Institut Mines-Télécom [Paris] (IMT)-Télécom SudParis (TSP), and Centre National de la Recherche Scientifique (CNRS)
- Subjects
FOS: Computer and information sciences ,Computer science ,business.industry ,Facial motion capture ,Computer Vision and Pattern Recognition (cs.CV) ,Frame (networking) ,Computer Science - Computer Vision and Pattern Recognition ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Scale-invariant feature transform ,Animation ,Active appearance model ,[INFO.INFO-TS]Computer Science [cs]/Signal and Image Processing ,Face (geometry) ,Computer vision ,Artificial intelligence ,[INFO.INFO-HC]Computer Science [cs]/Human-Computer Interaction [cs.HC] ,business ,Likelihood function ,Pose ,[SPI.SIGNAL]Engineering Sciences [physics]/Signal and Image processing - Abstract
International audience; This paper presents a new method to track both the face pose and the face animation with a monocular camera. The approach is based on the 3D face model CANDIDE and on the SIFT (Scale Invariant Feature Transform) descriptors, extracted around a few given landmarks (26 selected vertices of CANDIDE model) with a Bayesian approach. The training phase is performed on a synthetic database generated from the first video frame. At each current frame, the face pose and animation parameters are estimated via a Bayesian approach, with a Gaussian prior and a Gaussian likelihood function whose the mean and the covariance matrix eigenvalues are updated from the previous frame using eigen decomposition. Numerical results on pose estimation and landmark locations are reported using the Boston University Face Tracking (BUFT) database and Talking Face video. They show that our approach, compared to six other published algorithms, provides a very good compromise and presents a promising perspective due to the good results in terms of landmark localization.
- Published
- 2013
- Full Text
- View/download PDF
43. Combining Descriptors Extracted from Feature Maps of Deconvolutional Networks and SIFT Descriptors in Scene Image Classification
- Author
-
Atsuo Yoshitaka, Dung A. Doan, Ngoc-Trung Tran, Bac Le, and Dinh-Phong Vo
- Subjects
Contextual image classification ,business.industry ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Scale-invariant feature transform ,Pattern recognition ,Support vector machine ,ComputingMethodologies_PATTERNRECOGNITION ,Kernel (image processing) ,Histogram ,Computer vision ,Visual Word ,Artificial intelligence ,Neural coding ,business ,Cluster analysis ,Mathematics - Abstract
This paper presents a new method to combine descriptors extracted from feature maps of Deconvolutional Networks and SIFT descriptors by converting them into histograms of local patterns, so the concatenation operation can be applied and ensure to increase the classification rate. We use K-means clustering algorithm to construct codebooks and compute Spatial Histograms to represent the distribution of local patterns in an image. Consequently, we can concatenate these histograms to make a new one that represents more local patterns than the originals. In the classification step, SVM associated with Histogram Intersection Kernel is utilized. In the experiments on Scene-15 Dataset containing 15 categories, the classification rates of our method are around 84% which outperforms Reconfigurable Bag-of-Words (RBoW), Sparse Covariance Patterns (SCP), Spatial Pyramid Matching (SPM), Spatial Pyramid Matching using Sparse Coding (ScSPM) and Visual Word Reweighting (VWR).
- Published
- 2013
- Full Text
- View/download PDF
44. Combined iterative channel estimation and data detection for space-time block codes
- Author
-
Ngoc Trung Tran, Xuan Nam Tran, and The Cuong Dinh
- Subjects
Block code ,Theoretical computer science ,Iterative method ,Computer science ,Data_CODINGANDINFORMATIONTHEORY ,Viterbi algorithm ,Space–time block code ,symbols.namesake ,Convolutional code ,symbols ,Error detection and correction ,Algorithm ,Decoding methods ,Computer Science::Information Theory ,Communication channel - Abstract
This paper considers the problem of channel estimation for space-time block code (STBC) systems with convolutional codes for error correction. We propose an iterative processing scheme which combines channel estimation and data detection to improve system performance and increase transmission efficiency. In the proposed scheme, pilot symbols are used for initial estimation of channel gains and data detection. The information bits detected during the data period are fed back to the channel estimator for re-estimating the channels and improving data detection. The proposed scheme allows to improve the BER performance and decrease the pilot length to 50% at the cost of only 2 processing iterations.
- Published
- 2010
- Full Text
- View/download PDF
45. An effective channel estimation method for transmit diversity systems
- Author
-
Ngoc Trung Tran, Xuan Nam Tran, and The Cuong Dinh
- Subjects
Block code ,business.industry ,Frame (networking) ,Data_CODINGANDINFORMATIONTHEORY ,Interference (wave propagation) ,Symbol (chemistry) ,Transmit diversity ,Orthogonality ,Encoding (memory) ,Telecommunications ,business ,Algorithm ,Computer Science::Information Theory ,Mathematics ,Communication channel - Abstract
This paper presents a simple yet effective channel estimation method for transmit diversity systems. We propose to use the Alamoutipsilas space-time encoding scheme to create pilot symbols and insert into the transmit data symbol streams from the transmit antennas. The orthogonality of the pilot symbols from the two transmit antennas allows to eliminate the interference among transmit antennas. The proposed pilot symbol assisted (PSA) method thus does not require pilot symbol expansion as in the previous methods. We also propose to use a simple temporal average operation to process channel decision statistics to replace temporal filters or interpolators. For a 16 pilot symbols in a 100 data symbol frame transmitted using the Alamoutipsilas space-time block coding, the performance degradation of the estimated symbols is less than 0.2 dB compared with the case of perfect channel estimation.
- Published
- 2008
- Full Text
- View/download PDF
46. Samba, un système multi-agents pour la compréhension des dynamiques agraires
- Author
-
Castella, Jean-Christophe, primary, Boissau, Stanislas, additional, Ngoc Trung, Tran, additional, and Dinh Quang, Dang, additional
- Full Text
- View/download PDF
47. Gestion intégrée des ressources naturelles en zones inondables tropicales
- Author
-
Acreman, Mike, Alhousseini, Soumana, Amani, Abou, Arfi, Robert, Bakayoko, Issa, Bamba, Fantagoma, Bamba, Fatogoma, Barreteau, Olivier, Bélières, Jean-Francois, Bénech, Vincent, Benga, Elisabeth, Bergkamp, Ger, Boissau, Stanislas, Bonté, Philippe, Bonvallot, Jacques, Bouaré, Seydou, Boureïma, Amadou, Bousquet, François, Boutrais, Jean, Braund, Richard, Cadier, Eric, Camara, Seydou, Castella, Jean-Christophe, Chevallier, Pierre, Chohin-Kuper, Anne, Cissé, Navon, Coly, Adrien, Dacosta, Honoré, Dembélé, Lamine, Derniame, Jean-Claude, Dezetter, Alain, Diagana, Cheikh Hamallah, Diakité, Cheik Hamalla, Diarra, Lassine, Diarra, Samuel, Diarra, Wamian, Diawara, Yelli, Diénépo, Kaïmama, Dinh Quang, Dang, Diop, Babacar, Diouf, Mamadou, Doray, Mathieu, Ducrot, Raphaèle, Duvail, Stéphanie, Dzéakou, Patricia, d’Aquino, Patrick, d’Herbès, Jean-Marc, Earl, Julia A., Feuillette, Sarah, Fofana, Lamine, Garcia, Serge, Gerbe, Alain, Gosse-Healy, Bénédicte, Goulven, Patrick Le, Hamerlynck, Olivier, Hassane, Adama, Kane, Alioune, Kassibo, Bréhima, Keita, Nancoman, Kodio, Amadou, Kodio, Amaga, Kositsakulchai, Ekasit, Kuper, Marcel, Laë, Raymond, Lemoalle, Jacques, Liénou, Gaston, Loireau, Maud, Magassa, Hamidou, Mahé, Gil, Maïga, Hamadoun, Maïga, Ousmane, Marie, Jérôme, Marieu, Bertrand, Mariko, Adama, Marlet, Serge, Messaoud, Brahim Ould, Mikolasek, Olivier, Molle, François, Monga, Olivier, Morand, Pierre, Moseley, William G., Mullon, Christian, Naah, Emmanuel, Ngantou, Daniel, Ngoc Trung, Tran, Ngounou Ngatcha, Benjamin, Niaré, Tiéma, Njitchoua, Roger, Nonguierma, André, Noray, Marie-Laure de, N’Diaye, Mamadou Kabirou, Olivry, Jean-Claude, Orange, Didier, Oswald, Marc, Ould Baba, Mohamed Lemine, Page, Christophe Le, Picouet, Cécile, Piron, Marie, Poncet, Yveline, Romagny, Bruno, Royer, Antoine, Samaké, Odiaba, Servat, Eric, Sicard, Bruno, Sidibé, Ibrahim, Sidibé, Souleymane, Sigha Nkamdjou, Luc, Sighomnou, Daniel, Soumaguel, Abdourhamane, Soumaré, Pape Ousmane, Sow, Mariama, Tchotsoua, Michel, Témé, Bino, Togola, Cissouma Diama, Valette, François, Zallé, Dieudonné, and Zaslavsky, Jean
- Subjects
Modèle ,RNK ,irrigation ,NAT011000 ,gestion de l’environnement ,fonctionnement de l’écosystème ,développement intégré ,Irrigation ,Hydrologie ,Geography ,Ressource renouvelable ,environnement ,milieu deltaïque ,Gestion des ressources ,politique de l’environnement ,Gestion des eaux ,Delta ,Modélisation ,développement durable ,Système d'exploitation agricole ,plaine inondable ,P01 - Conservation de la nature et ressources foncières ,exploitation des ressources naturelles ,Plaine d'inondation ,Écosystème - Abstract
De par leur richesse en ressources naturelles renouvelables, les zones inondables tropicales revetent un interet social et economique majeur pour les pays en developpement. Cependant, les fleuves tropicaux sont aujourd'hui de plus en plus amenages pour satisfaire les besoins lies a de nouvelles activites. Les zones jusque-la regulierement inondees par la crue annuelle se reduisent ou alors les rythmes de leur inondation sont profondement modifies. Les impacts de tels changements sont nombreux et portent atteinte a la biodiversite et a la durabilite des systemes d'exploitation. Il s'avere alors necessaire de definir de nouvelles approches de la gestion de l'eau, des espaces et des ressources vivantes, qui tout a la fois preservent les ecosystemes et prennent en consideration les besoins des differents usagers. Tel est l'objectif de cet ouvrage qui pose, dans un premier temps, la problematique societale autour de laquelle cette gestion doit etre repensee, en faisant apparaitre la diversite d'acteurs et d'institutions concernes. Il presente ensuite les acquis les plus recents de la recherche sur le fonctionnement de ces ecosystemes ainsi que sur les pratiques et strategies deployees par les populations qui les exploitent. Enfin est abordee la question des instruments a mettre en place pour assurer l'effectivite d'une gestion durable des zones inondables tropicales : apres avoir fait le point sur les apports de la recherche concernant les outils de traitement et de partage de l'information environnementale, l'ouvrage se termine par un debat sur les conditions de creation et de fonctionnement des institutions de suivi, de concertation et de decision. (Resume d'auteur)
- Published
- 2002
48. Detecting Probable Regions of Humans in Still Images Using Raw Edges
- Author
-
Ngoc, Trung Tran, primary, Dinh, Phong Vo, additional, and Hoai, Bac Le, additional
- Published
- 2009
- Full Text
- View/download PDF
49. Learned and designed features for sparse coding in image classification.
- Author
-
Doan, Dung A., Ngoc-Trung Tran, Dinh-Phong Vo, and Bac Le
- Abstract
There is an amount of designed features (SIFT, SURF, or DAISY) which has been chosen in the standard implementation of some visual recognition and multimedia challenges. The power of these features lie on their invariance designed against rotation, scaling, and translation. Recent trends in deep learning, however, have pointed out that data-driven features learning performs better designed features in some tasks, since they can capture the global (via multi-layers network) or inter-local structures (convolutional network) of images. We argue that combining the two types of features can significantly improve visual object recognition performance. We propose in this paper a framework that uses sparse coding and the fusion of learned and designed features in order to build descriptive codewords. Evaluations on Caltech-101 and 15 Scenes validates our argument, with a better result compared with recent approaches. [ABSTRACT FROM PUBLISHER]
- Published
- 2013
- Full Text
- View/download PDF
50. IRIM - Indexation et Recherche d'Information Multimedia GDR-ISIS
- Author
-
Nicolas Ballas, Benjamin Labbé, Aymen Shabou, Hervé Le Borgne, Philippe Gosselin, Miriam Redi, Bernard Merialdo, Hervé Jégou, Jonathan Delhumeau, Rémi Vieux, Boris Mansencal, Jenny Benois-Pineau, Stéphane Ayache, Abdelkader Haadi, Bahjat Safadi, Franck Thollard, Nadia Derbas, Georges Quénot, Hervé Bredin, Matthieu Cord, Boyang Gao, Chao Zhu, Yuxing Tang, Emmanuel Dellandreav, Charles-Edmond Bichot, Liming Chen, alexandre benoit, Patrick Lambert, Tiberius Strat, Joseph Razik, Sébastion Paris, Hervé Glotin, Ngoc-Trung Tran, Dijana Petrovska-Delacrétaz, Gérard Chollet, Andrei Stoian, Michel Crucianu, Signal, Statistique et Apprentissage (S2A), Laboratoire Traitement et Communication de l'Information (LTCI), Institut Mines-Télécom [Paris] (IMT)-Télécom Paris-Institut Mines-Télécom [Paris] (IMT)-Télécom Paris, Département Traitement du Signal et des Images (TSI), Télécom ParisTech-Centre National de la Recherche Scientifique (CNRS), and Multimédia (MM)
- Subjects
[INFO.INFO-LG]Computer Science [cs]/Machine Learning [cs.LG] ,[INFO.INFO-TI]Computer Science [cs]/Image Processing [eess.IV] ,InformationSystems_INFORMATIONSTORAGEANDRETRIEVAL ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,[INFO.INFO-CV]Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV] - Abstract
International audience; The IRIM group is a consortium of French teams working on Multimedia Indexing and Retrieval. This paper describes its participation to the TRECVID 2012 semantic indexing and instance search tasks. For the semantic indexing task, our approach uses a six-stagesprocessing pipelines for computing scores for the likelihood of a video shot to contain a target concept. These scores are then used for producing a ranked list of images or shots that are the most likely to contain the target concept. The pipeline is composed of the following steps: descriptor extraction, descriptor optimization, classification, fusion of descriptor variants, higher-level fusion, and re-ranking. We evaluated a number of different descriptors and tried dierent fusion strategies. The best IRIM run has a Mean Inferred Average Precision of 0.2378, which ranked us 4th out of 16 participants.For the instance search task, our approach uses two steps. First individual methods of participants are used to compute similrity between an example image of instance and keyframes of a video clip. Then a two-step fusion method is used to combine these individual results and obtain a score for the likelihood of an instance to appear in a video clip. These scores are used to obtain a ranked list of clips the most likely to contain the queried instance. The best IRIM run has a MAP of 0.1192, which ranked us 29th on 79 fully automatic runs.
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.