Picture for Xidong Wang

Xidong Wang

Roadmap towards Superhuman Speech Understanding using Large Language Models

Add code
Oct 17, 2024
Figure 1 for Roadmap towards Superhuman Speech Understanding using Large Language Models
Figure 2 for Roadmap towards Superhuman Speech Understanding using Large Language Models
Figure 3 for Roadmap towards Superhuman Speech Understanding using Large Language Models
Figure 4 for Roadmap towards Superhuman Speech Understanding using Large Language Models
Viaarxiv icon

Efficiently Democratizing Medical LLMs for 50 Languages via a Mixture of Language Family Experts

Add code
Oct 14, 2024
Figure 1 for Efficiently Democratizing Medical LLMs for 50 Languages via a Mixture of Language Family Experts
Figure 2 for Efficiently Democratizing Medical LLMs for 50 Languages via a Mixture of Language Family Experts
Figure 3 for Efficiently Democratizing Medical LLMs for 50 Languages via a Mixture of Language Family Experts
Figure 4 for Efficiently Democratizing Medical LLMs for 50 Languages via a Mixture of Language Family Experts
Viaarxiv icon

Less is More: A Simple yet Effective Token Reduction Method for Efficient Multi-modal LLMs

Add code
Sep 17, 2024
Viaarxiv icon

LongLLaVA: Scaling Multi-modal LLMs to 1000 Images Efficiently via Hybrid Architecture

Add code
Sep 04, 2024
Viaarxiv icon

CoD, Towards an Interpretable Medical Agent using Chain of Diagnosis

Add code
Jul 18, 2024
Figure 1 for CoD, Towards an Interpretable Medical Agent using Chain of Diagnosis
Figure 2 for CoD, Towards an Interpretable Medical Agent using Chain of Diagnosis
Figure 3 for CoD, Towards an Interpretable Medical Agent using Chain of Diagnosis
Figure 4 for CoD, Towards an Interpretable Medical Agent using Chain of Diagnosis
Viaarxiv icon

HuatuoGPT-Vision, Towards Injecting Medical Visual Knowledge into Multimodal LLMs at Scale

Add code
Jun 27, 2024
Figure 1 for HuatuoGPT-Vision, Towards Injecting Medical Visual Knowledge into Multimodal LLMs at Scale
Figure 2 for HuatuoGPT-Vision, Towards Injecting Medical Visual Knowledge into Multimodal LLMs at Scale
Figure 3 for HuatuoGPT-Vision, Towards Injecting Medical Visual Knowledge into Multimodal LLMs at Scale
Figure 4 for HuatuoGPT-Vision, Towards Injecting Medical Visual Knowledge into Multimodal LLMs at Scale
Viaarxiv icon

LLMs for Doctors: Leveraging Medical LLMs to Assist Doctors, Not Replace Them

Add code
Jun 26, 2024
Figure 1 for LLMs for Doctors: Leveraging Medical LLMs to Assist Doctors, Not Replace Them
Figure 2 for LLMs for Doctors: Leveraging Medical LLMs to Assist Doctors, Not Replace Them
Figure 3 for LLMs for Doctors: Leveraging Medical LLMs to Assist Doctors, Not Replace Them
Figure 4 for LLMs for Doctors: Leveraging Medical LLMs to Assist Doctors, Not Replace Them
Viaarxiv icon

Apollo: An Lightweight Multilingual Medical LLM towards Democratizing Medical AI to 6B People

Add code
Mar 09, 2024
Figure 1 for Apollo: An Lightweight Multilingual Medical LLM towards Democratizing Medical AI to 6B People
Figure 2 for Apollo: An Lightweight Multilingual Medical LLM towards Democratizing Medical AI to 6B People
Figure 3 for Apollo: An Lightweight Multilingual Medical LLM towards Democratizing Medical AI to 6B People
Figure 4 for Apollo: An Lightweight Multilingual Medical LLM towards Democratizing Medical AI to 6B People
Viaarxiv icon

MLLM-Bench, Evaluating Multi-modal LLMs using GPT-4V

Add code
Nov 23, 2023
Figure 1 for MLLM-Bench, Evaluating Multi-modal LLMs using GPT-4V
Figure 2 for MLLM-Bench, Evaluating Multi-modal LLMs using GPT-4V
Figure 3 for MLLM-Bench, Evaluating Multi-modal LLMs using GPT-4V
Figure 4 for MLLM-Bench, Evaluating Multi-modal LLMs using GPT-4V
Viaarxiv icon

HuatuoGPT-II, One-stage Training for Medical Adaption of LLMs

Add code
Nov 16, 2023
Viaarxiv icon