Search

Your search keyword '"Chang, Shih-Fu"' showing total 1,275 results

Search Constraints

Start Over You searched for: Author "Chang, Shih-Fu" Remove constraint Author: "Chang, Shih-Fu"
1,275 results on '"Chang, Shih-Fu"'

Search Results

51. Video in 10 Bits: Few-Bit VideoQA for Efficiency and Privacy

52. Task-Adaptive Negative Envision for Few-Shot Open-Set Recognition

53. Open-Vocabulary Object Detection Using Captions

54. Neuro-Symbolic Representations for Video Captioning: A Case for Leveraging Inductive Biases for Vision and Language

55. Unsupervised Vision-and-Language Pre-training Without Parallel Images and Captions

56. Uncertainty-Aware Few-Shot Image Classification

57. Ref-NMS: Breaking Proposal Bottlenecks in Two-Stage Referring Expression Grounding

58. Analogical Reasoning for Visually Grounded Language Acquisition

59. COVID-19 Literature Knowledge Graph Construction and Drug Repurposing Report Generation

60. Learning Visual Commonsense for Robust Scene Graph Generation

61. Beyond Triplet Loss: Meta Prototypical N-tuple Loss for Person Re-identification

62. Deep Learning Guided Building Reconstruction from Satellite Imagery-derived Point Clouds

63. Cross-media Structured Common Space for Multimedia Event Extraction

64. Unifying Specialist Image Embedding into Universal Image Embedding

65. Training with Streaming Annotation

66. Weakly Supervised Visual Semantic Parsing

67. Bridging Knowledge Graphs to Generate Scene Graphs

68. General Partial Label Learning via Dual Bipartite Graph Autoencoder

70. Flow-Distilled IP Two-Stream Networks for Compressed Video Action Recognition

71. Learning to Learn Words from Visual Scenes

72. Towards Train-Test Consistency for Semi-supervised Temporal Action Localization

73. Context-Gated Convolution

74. Report of 2017 NSF Workshop on Multimedia Challenges, Opportunities and Research Roadmaps

75. Detecting and Simulating Artifacts in GAN Fake Images

76. Variational Context: Exploiting Visual and Textual Context for Grounding Referring Expressions

77. CDSA: Cross-Dimensional Self-Attention for Multivariate, Geo-tagged Time Series Imputation

78. Unsupervised Embedding Learning via Invariant and Spreading Instance Feature

79. Unsupervised Rank-Preserving Hashing for Large-Scale Image Retrieval

80. DMC-Net: Generating Discriminative Motion Cues for Fast Compressed Video Action Recognition

81. Few-Shot End-to-End Object Detection via Constantly Concentrated Encoding Across Heads

82. Counterfactual Critic Multi-Agent Training for Scene Graph Generation

83. Multi-granularity Generator for Temporal Action Proposal

84. Multi-level Multimodal Common Semantic Space for Image-Phrase Grounding

85. Low-shot Learning via Covariance-Preserving Adversarial Augmentation Networks

86. Heated-Up Softmax Embedding

87. Multimodal Social Media Analysis for Gang Violence Prevention

88. AutoLoc: Weakly-supervised Temporal Action Localization

89. Entity-aware Image Caption Generation

90. Online Detection of Action Start in Untrimmed, Streaming Videos

91. Zero-Shot Visual Recognition using Semantics-Preserving Adversarial Embedding Networks

92. Grounding Referring Expressions in Images by Variational Context

93. Multi-Modal Multi-Scale Deep Learning for Large-Scale Image Annotation

94. Skip RNN: Learning to Skip State Updates in Recurrent Neural Networks

95. Learning Spread-out Local Feature Descriptors

96. More cat than cute? Interpretable Prediction of Adjective-Noun Pairs

97. ConvNet Architecture Search for Spatiotemporal Feature Learning

98. PPR-FCN: Weakly Supervised Visual Relation Detection via Parallel Pairwise R-FCN

99. Localizing Actions from Video Labels and Pseudo-Annotations

100. Modeling Multimodal Clues in a Hybrid Deep Learning Framework for Video Classification

Catalog

Books, media, physical & digital resources