94 results on '"Hsin-Min Wang"'
Search Results
2. Magnetic-Field-Assisted Electric-Field-Induced Domain Switching of a Magnetic Single Domain in a Multiferroic/Magnetoelectric Ni Nanochevron/[Pb(Mg1/3Nb2/3)O3]0.68–[PbTiO3]0.32 (PMN–PT) Layered Structure
3. BASPRO: A Balanced Script Producer for Speech Corpus Collection Based on the Genetic Algorithm
4. Speaker-Specific Articulatory Feature Extraction Based on Knowledge Distillation for Speaker Recognition
5. A Study On Incorporating Whisper For Robust Speech Assessment.
6. SpeechCLIP+: Self-Supervised Multi-Task Representation Learning for Speech Via Clip and Speech-Image Data.
7. Multi-Task Pseudo-Label Learning for Non-Intrusive Speech Quality Assessment Model.
8. A Study on Zero-shot Non-intrusive Speech Assessment using Large Language Models.
9. Leveraging Joint Spectral and Spatial Learning with MAMBA for Multichannel Speech Enhancement.
10. Channel-Aware Domain-Adaptive Generative Adversarial Network for Robust Speech Recognition.
11. Robust Audio-Visual Speech Enhancement: Correcting Misassignments in Complex Environments with Advanced Post-Processing.
12. Exploring the Impact of Data Quantity on ASR in Extremely Low-resource Languages.
13. Effective Noise-aware Data Simulation for Domain-adaptive Speech Enhancement Leveraging Dynamic Stochastic Perturbation.
14. The VoiceMOS Challenge 2024: Beyond Speech Quality Prediction.
15. Deep learning for identifying personal and family history of suicidal thoughts and behaviors from EHRs.
16. SVSNet+: Enhancing Speaker Voice Similarity Assessment Models with Representations from Speech Foundation Models.
17. HAAQI-Net: A non-intrusive neural music quality assessment model for hearing aids.
18. Unmasking Illusions: Understanding Human Perception of Audiovisual Deepfakes.
19. Audio-Visual Mandarin Electrolaryngeal Speech Voice Conversion.
20. A Training and Inference Strategy Using Noisy and Enhanced Speech as Target for Speech Enhancement without Clean Speech.
21. Mandarin Electrolaryngeal Speech Voice Conversion using Cross-domain Features.
22. The Voicemos Challenge 2023: Zero-Shot Subjective Speech Quality Prediction for Multiple Domains.
23. LC4SV: A Denoising Framework Learning to Compensate for Unseen Speaker Verification Models.
24. Multi-Target Extractor and Detector for Unknown-Number Speaker Diarization.
25. Generalization Ability Improvement of Speaker Representation and Anti-Interference for Speaker Verification.
26. Deep Learning-Based Non-Intrusive Multi-Objective Speech Assessment Model With Cross-Domain Features.
27. Decomposition and Reorganization of Phonetic Information for Speaker Embedding Learning.
28. MTI-Net: A Multi-Target Speech Intelligibility Prediction Model.
29. NASTAR: Noise Adaptive Speech Enhancement with Target-Conditional Resampling.
30. Disentangling the Impacts of Language and Channel Variability on Speech Separation Networks.
31. Chain-based Discriminative Autoencoders for Speech Recognition.
32. The VoiceMOS Challenge 2022.
33. MBI-Net: A Non-Intrusive Multi-Branched Speech Intelligibility Prediction Model for Hearing Aids.
34. EMGSE: Acoustic/EMG Fusion for Multimodal Speech Enhancement.
35. Partially Fake Audio Detection by Self-Attention-Based Fake Span Discovery.
36. Speech-enhanced and Noise-aware Networks for Robust Speech Recognition.
37. Is Character Trigram Overlapping Ratio Still the Best Similarity Measure for Aligning Sentences in a Paraphrased Corpus?
38. Chinese Movie Dialogue Question Answering Dataset.
39. D4AM: A General Denoising Framework for Downstream Acoustic Models.
40. SVSNet: An End-to-End Speaker Voice Similarity Assessment Model.
41. Improved Lite Audio-Visual Speech Enhancement.
42. Utilizing Whisper to Enhance Multi-Branched Speech Intelligibility Prediction Model for Hearing Aids.
43. Deep Complex U-Net with Conformer for Audio-Visual Speech Enhancement.
44. A Study on Incorporating Whisper for Robust Speech Assessment.
45. Multi-objective Non-intrusive Hearing-aid Speech Assessment Model.
46. Multi-Task Pseudo-Label Learning for Non-Intrusive Speech Quality Assessment Model.
47. AV-Lip-Sync+: Leveraging AV-HuBERT to Exploit Multimodal Inconsistency for Video Deepfake Detection.
48. AVTENet: Audio-Visual Transformer-based Ensemble Network Exploiting Multiple Experts for Video Deepfake Detection.
49. Filter-based Discriminative Autoencoders for Children Speech Recognition.
50. MTI-Net: A Multi-Target Speech Intelligibility Prediction Model.
Catalog
Books, media, physical & digital resources
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.