Picture for Kyusong Lee

Kyusong Lee

OmChat: A Recipe to Train Multimodal Language Models with Strong Long Context and Video Understanding

Add code
Jul 06, 2024
Figure 1 for OmChat: A Recipe to Train Multimodal Language Models with Strong Long Context and Video Understanding
Figure 2 for OmChat: A Recipe to Train Multimodal Language Models with Strong Long Context and Video Understanding
Figure 3 for OmChat: A Recipe to Train Multimodal Language Models with Strong Long Context and Video Understanding
Figure 4 for OmChat: A Recipe to Train Multimodal Language Models with Strong Long Context and Video Understanding
Viaarxiv icon

OmAgent: A Multi-modal Agent Framework for Complex Video Understanding with Task Divide-and-Conquer

Add code
Jun 25, 2024
Viaarxiv icon

Preserving Knowledge in Large Language Model: A Model-Agnostic Self-Decompression Approach

Add code
Jun 17, 2024
Viaarxiv icon

Real-time Transformer-based Open-Vocabulary Detection with Efficient Fusion Head

Add code
Mar 11, 2024
Viaarxiv icon

How to Evaluate the Generalization of Detection? A Benchmark for Comprehensive Open-Vocabulary Detection

Add code
Aug 25, 2023
Viaarxiv icon

OmDet: Language-Aware Object Detection with Large-scale Vision-Language Multi-dataset Pre-training

Add code
Sep 10, 2022
Figure 1 for OmDet: Language-Aware Object Detection with Large-scale Vision-Language Multi-dataset Pre-training
Figure 2 for OmDet: Language-Aware Object Detection with Large-scale Vision-Language Multi-dataset Pre-training
Figure 3 for OmDet: Language-Aware Object Detection with Large-scale Vision-Language Multi-dataset Pre-training
Figure 4 for OmDet: Language-Aware Object Detection with Large-scale Vision-Language Multi-dataset Pre-training
Viaarxiv icon

VL-CheckList: Evaluating Pre-trained Vision-Language Models with Objects, Attributes and Relations

Add code
Jul 01, 2022
Figure 1 for VL-CheckList: Evaluating Pre-trained Vision-Language Models with Objects, Attributes and Relations
Figure 2 for VL-CheckList: Evaluating Pre-trained Vision-Language Models with Objects, Attributes and Relations
Figure 3 for VL-CheckList: Evaluating Pre-trained Vision-Language Models with Objects, Attributes and Relations
Figure 4 for VL-CheckList: Evaluating Pre-trained Vision-Language Models with Objects, Attributes and Relations
Viaarxiv icon

When is it permissible for artificial intelligence to lie? A trust-based approach

Add code
Mar 14, 2021
Viaarxiv icon

SF-QA: Simple and Fair Evaluation Library for Open-domain Question Answering

Add code
Jan 06, 2021
Figure 1 for SF-QA: Simple and Fair Evaluation Library for Open-domain Question Answering
Figure 2 for SF-QA: Simple and Fair Evaluation Library for Open-domain Question Answering
Figure 3 for SF-QA: Simple and Fair Evaluation Library for Open-domain Question Answering
Figure 4 for SF-QA: Simple and Fair Evaluation Library for Open-domain Question Answering
Viaarxiv icon

VisualSparta: Sparse Transformer Fragment-level Matching for Large-scale Text-to-Image Search

Add code
Jan 01, 2021
Figure 1 for VisualSparta: Sparse Transformer Fragment-level Matching for Large-scale Text-to-Image Search
Figure 2 for VisualSparta: Sparse Transformer Fragment-level Matching for Large-scale Text-to-Image Search
Figure 3 for VisualSparta: Sparse Transformer Fragment-level Matching for Large-scale Text-to-Image Search
Figure 4 for VisualSparta: Sparse Transformer Fragment-level Matching for Large-scale Text-to-Image Search
Viaarxiv icon