130 results on '"Daizong Liu"'
Search Results
2. Cross-Task Knowledge Transfer for Semi-supervised Joint 3D Grounding and Captioning.
3. Advancing 3D Object Grounding Beyond a Single 3D Scene.
4. Not All Inputs Are Valid: Towards Open-Set Video Moment Retrieval using Language.
5. Frequency-Aware GAN for Imperceptible Transfer Attack on 3D Point Clouds.
6. Hiding Imperceptible Noise in Curvature-Aware Patches for 3D Point Cloud Attack.
7. FLAT: Flux-Aware Imperceptible Adversarial Attacks on 3D Point Clouds.
8. Towards Robust Temporal Activity Localization Learning with Noisy Labels.
9. Explicitly Perceiving and Preserving the Local Geometric Structures for 3D Point Cloud Attack.
10. Unsupervised Domain Adaptative Temporal Sentence Localization with Mutual Information Maximization.
11. Fewer Steps, Better Performance: Efficient Cross-Modal Clip Trimming for Video Moment Retrieval Using Language.
12. Manifold Constraints for Imperceptible Adversarial Attacks on Point Clouds.
13. Robust Geometry-Dependent Attack for 3D Point Clouds.
14. Hierarchical Local-Global Transformer for Temporal Sentence Grounding.
15. Conditional Video Diffusion Network for Fine-Grained Temporal Sentence Grounding.
16. Transform-Equivariant Consistency Learning for Temporal Sentence Grounding.
17. Point Cloud Attacks in Graph Spectral Domain: When 3D Geometry Meets Graph Signal Processing.
18. Joint Top-Down and Bottom-Up Frameworks for 3D Visual Grounding.
19. A Survey of Attacks on Large Vision-Language Models: Resources, Advances, and Future Trends.
20. A Survey on Text-guided 3D Visual Grounding: Elements, Recent Advances, and Future Directions.
21. 3DHacker: Spectrum-based Decision Boundary Generation for Hard-label 3D Point Cloud Attack.
22. Dual Learning with Dynamic Knowledge Distillation for Partially Relevant Video Retrieval.
23. Annotations Are Not All You Need: A Cross-modal Knowledge Transfer Network for Unsupervised Temporal Sentence Grounding.
24. You Can Ground Earlier than See: An Effective and Efficient Pipeline for Temporal Sentence Grounding in Compressed Videos.
25. Density-Insensitive Unsupervised Domain Adaption on 3D Object Detection.
26. Lite-MKD: A Multi-modal Knowledge Distillation Framework for Lightweight Few-shot Action Recognition.
27. Unified Multi-modal Unsupervised Representation Learning for Skeleton-based Action Understanding.
28. Dense Object Grounding in 3D Scenes.
29. Filling the Information Gap between Video and Query for Language-Driven Moment Retrieval.
30. Matching Words for Out-of-distribution Detection.
31. Jointly Visual- and Semantic-Aware Graph Memory Networks for Temporal Sentence Localization in Videos.
32. Tracking Objects and Activities with Attention for Temporal Sentence Grounding.
33. Distantly-Supervised Named Entity Recognition with Adaptive Teacher Learning and Fine-Grained Student Ensemble.
34. Hypotheses Tree Building for One-Shot Temporal Sentence Localization.
35. From Region to Patch: Attribute-Aware Foreground-Background Contrastive Learning for Fine-Grained Fashion Retrieval.
36. Exploring Optical-Flow-Guided Motion and Detection-Based Appearance for Temporal Sentence Grounding.
37. Multi-Modal Cross-Domain Alignment Network for Video Moment Retrieval.
38. Few-Shot Temporal Sentence Grounding via Memory-Guided Semantic Learning.
39. Imperceptible Transfer Attack and Defense on 3D Point Cloud Classification.
40. A Hybird Alignment Loss for Temporal Moment Localization with Natural Language.
41. Rethinking the Video Sampling and Reasoning Strategies for Temporal Sentence Grounding.
42. Learning to Focus on the Foreground for Temporal Sentence Grounding.
43. Reducing the Vision and Language Bias for Temporal Sentence Grounding.
44. Skimming, Locating, then Perusing: A Human-Like Framework for Natural Language Video Localization.
45. Gaussian Kernel-Based Cross Modal Network for Spatio-Temporal Video Grounding.
46. Exploring the Devil in Graph Spectral Domain for 3D Point Cloud Attacks.
47. Memory-Guided Semantic Learning Network for Temporal Sentence Grounding.
48. Unsupervised Temporal Video Grounding with Deep Semantic Clustering.
49. Exploring Motion and Appearance Information for Temporal Sentence Grounding.
50. Transform-Equivariant Consistency Learning for Temporal Sentence Grounding.
Catalog
Books, media, physical & digital resources
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.