Picture for Bowen Yang

Bowen Yang

Charles

Can GPT tell us why these images are synthesized? Empowering Multimodal Large Language Models for Forensics

Add code
Apr 16, 2025
Viaarxiv icon

Command A: An Enterprise-Ready Large Language Model

Add code
Apr 01, 2025
Viaarxiv icon

Language Models Can See Better: Visual Contrastive Decoding For LLM Multimodal Reasoning

Add code
Feb 17, 2025
Viaarxiv icon

Rope to Nope and Back Again: A New Hybrid Attention Strategy

Add code
Jan 30, 2025
Figure 1 for Rope to Nope and Back Again: A New Hybrid Attention Strategy
Figure 2 for Rope to Nope and Back Again: A New Hybrid Attention Strategy
Figure 3 for Rope to Nope and Back Again: A New Hybrid Attention Strategy
Figure 4 for Rope to Nope and Back Again: A New Hybrid Attention Strategy
Viaarxiv icon

Mimicking-Bench: A Benchmark for Generalizable Humanoid-Scene Interaction Learning via Human Mimicking

Add code
Dec 23, 2024
Viaarxiv icon

Aya Expanse: Combining Research Breakthroughs for a New Multilingual Frontier

Add code
Dec 05, 2024
Figure 1 for Aya Expanse: Combining Research Breakthroughs for a New Multilingual Frontier
Figure 2 for Aya Expanse: Combining Research Breakthroughs for a New Multilingual Frontier
Figure 3 for Aya Expanse: Combining Research Breakthroughs for a New Multilingual Frontier
Figure 4 for Aya Expanse: Combining Research Breakthroughs for a New Multilingual Frontier
Viaarxiv icon

Real-Time Metric-Semantic Mapping for Autonomous Navigation in Outdoor Environments

Add code
Nov 30, 2024
Viaarxiv icon

ALR$^2$: A Retrieve-then-Reason Framework for Long-context Question Answering

Add code
Oct 04, 2024
Figure 1 for ALR$^2$: A Retrieve-then-Reason Framework for Long-context Question Answering
Figure 2 for ALR$^2$: A Retrieve-then-Reason Framework for Long-context Question Answering
Figure 3 for ALR$^2$: A Retrieve-then-Reason Framework for Long-context Question Answering
Figure 4 for ALR$^2$: A Retrieve-then-Reason Framework for Long-context Question Answering
Viaarxiv icon

Uncertainty Quantification for Bird's Eye View Semantic Segmentation: Methods and Benchmarks

Add code
May 31, 2024
Figure 1 for Uncertainty Quantification for Bird's Eye View Semantic Segmentation: Methods and Benchmarks
Figure 2 for Uncertainty Quantification for Bird's Eye View Semantic Segmentation: Methods and Benchmarks
Figure 3 for Uncertainty Quantification for Bird's Eye View Semantic Segmentation: Methods and Benchmarks
Figure 4 for Uncertainty Quantification for Bird's Eye View Semantic Segmentation: Methods and Benchmarks
Viaarxiv icon

BeautyMap: Binary-Encoded Adaptable Ground Matrix for Dynamic Points Removal in Global Maps

Add code
May 12, 2024
Figure 1 for BeautyMap: Binary-Encoded Adaptable Ground Matrix for Dynamic Points Removal in Global Maps
Figure 2 for BeautyMap: Binary-Encoded Adaptable Ground Matrix for Dynamic Points Removal in Global Maps
Figure 3 for BeautyMap: Binary-Encoded Adaptable Ground Matrix for Dynamic Points Removal in Global Maps
Figure 4 for BeautyMap: Binary-Encoded Adaptable Ground Matrix for Dynamic Points Removal in Global Maps
Viaarxiv icon