Search

Your search keyword '"ZHANG Wenwei"' showing total 1,061 results

Search Constraints

Start Over You searched for: Author "ZHANG Wenwei" Remove constraint Author: "ZHANG Wenwei"
1,061 results on '"ZHANG Wenwei"'

Search Results

1. MindSearch: Mimicking Human Minds Elicits Deep AI Searcher

2. CIBench: Evaluating Your LLMs with a Code Interpreter Plugin

3. 4D Contrastive Superflows are Dense 3D Representation Learners

4. ANAH-v2: Scaling Analytical Hallucination Annotation of Large Language Models

5. InternLM-XComposer-2.5: A Versatile Large Vision Language Model Supporting Long-Contextual Input and Output

6. ScanReason: Empowering 3D Visual Grounding with Reasoning Capabilities

7. InternLM-Law: An Open Source Chinese Legal Large Language Model

8. F-LMM: Grounding Frozen Large Multimodal Models

9. ANAH: Analytical Annotation of Hallucinations in Large Language Models

10. AlchemistCoder: Harmonizing and Eliciting Code Capability by Hindsight Tuning on Multi-source Data

11. Benchmarking and Improving Bird's Eye View Perception Robustness in Autonomous Driving

12. An Empirical Study of Training State-of-the-Art LiDAR Segmentation Models

13. MathBench: Evaluating the Theory and Application Proficiency of LLMs with a Hierarchical Mathematics Benchmark

14. The RoboDrive Challenge: Drive Anytime Anywhere in Any Condition

15. Multi-Modal Data-Efficient 3D Scene Understanding for Autonomous Driving

16. InternLM-XComposer2-4KHD: A Pioneering Large Vision-Language Model Handling Resolutions from 336 Pixels to 4K HD

17. InternLM2 Technical Report

18. Calib3D: Calibrating Model Preferences for Reliable 3D Scene Understanding

19. Agent-FLAN: Designing Data and Methods of Effective Agent Tuning for Large Language Models

20. CriticBench: Evaluating Large Language Models as Critic

21. Code Needs Comments: Enhancing Code LLMs with Comment Augmentation

22. InternLM-Math: Open Math Large Language Models Toward Verifiable Reasoning

23. InternLM-XComposer2: Mastering Free-form Text-Image Composition and Comprehension in Vision-Language Large Model

24. Can AI Assistants Know What They Don't Know?

25. EmbodiedScan: A Holistic Multi-Modal 3D Perception Suite Towards Embodied AI

26. T-Eval: Evaluating the Tool Utilization Capability of Large Language Models Step by Step

27. CLIM: Contrastive Language-Image Mosaic for Region Representation

28. Mixed Pseudo Labels for Semi-Supervised Object Detection

33. Fake Alignment: Are LLMs Really Aligned Well?

34. OV-PARTS: Towards Open-Vocabulary Part Segmentation

35. Evaluating Hallucinations in Chinese Large Language Models

36. CLIPSelf: Vision Transformer Distills Itself for Open-Vocabulary Dense Prediction

37. DST-Det: Simple Dynamic Self-Training for Open-Vocabulary Object Detection

38. InternLM-XComposer: A Vision-Language Large Model for Advanced Text-image Comprehension and Composition

39. Object2Scene: Putting Objects in Context for Open-Vocabulary 3D Detection

40. Unified Human-Scene Interaction via Prompted Chain-of-Contacts

41. GPT4RoI: Instruction Tuning Large Language Model on Region-of-Interest

42. Segment Any Point Cloud Sequences by Distilling Vision Foundation Models

43. MultiModal-GPT: A Vision and Language Model for Dialogue with Humans

44. Transformer-Based Visual Segmentation: A Survey

45. RoboBEV: Towards Robust Bird's Eye View Perception under Corruptions

46. Robo3D: Towards Robust and Reliable 3D Perception against Corruptions

47. Position-Guided Point Cloud Panoptic Segmentation Transformer

48. MV-JAR: Masked Voxel Jigsaw and Reconstruction for LiDAR-Based Self-Supervised Pre-Training

49. Tube-Link: A Flexible Cross Tube Framework for Universal Video Segmentation

50. Dense Distinct Query for End-to-End Object Detection

Catalog

Books, media, physical & digital resources