Building a Video-and-Language Dataset with Human Actions for Multimodal Logical Inference

Add code
Jun 27, 2021
Figure 1 for Building a Video-and-Language Dataset with Human Actions for Multimodal Logical Inference
Figure 2 for Building a Video-and-Language Dataset with Human Actions for Multimodal Logical Inference
Figure 3 for Building a Video-and-Language Dataset with Human Actions for Multimodal Logical Inference
Figure 4 for Building a Video-and-Language Dataset with Human Actions for Multimodal Logical Inference

Share this with someone who'll enjoy it:

View paper onarxiv icon

Share this with someone who'll enjoy it: