Picture for Yusuke Iwasawa

Yusuke Iwasawa

MKG-Rank: Enhancing Large Language Models with Knowledge Graph for Multilingual Medical Question Answering

Add code
Mar 21, 2025
Viaarxiv icon

MMLU-ProX: A Multilingual Benchmark for Advanced Large Language Model Evaluation

Add code
Mar 13, 2025
Viaarxiv icon

Image Referenced Sketch Colorization Based on Animation Creation Workflow

Add code
Feb 27, 2025
Viaarxiv icon

Beyond In-Distribution Success: Scaling Curves of CoT Granularity for Language Model Generalization

Add code
Feb 25, 2025
Viaarxiv icon

Large Language Models as Theory of Mind Aware Generative Agents with Counterfactual Reflection

Add code
Jan 26, 2025
Figure 1 for Large Language Models as Theory of Mind Aware Generative Agents with Counterfactual Reflection
Figure 2 for Large Language Models as Theory of Mind Aware Generative Agents with Counterfactual Reflection
Figure 3 for Large Language Models as Theory of Mind Aware Generative Agents with Counterfactual Reflection
Figure 4 for Large Language Models as Theory of Mind Aware Generative Agents with Counterfactual Reflection
Viaarxiv icon

Rethinking Evaluation of Sparse Autoencoders through the Representation of Polysemous Words

Add code
Jan 09, 2025
Figure 1 for Rethinking Evaluation of Sparse Autoencoders through the Representation of Polysemous Words
Figure 2 for Rethinking Evaluation of Sparse Autoencoders through the Representation of Polysemous Words
Figure 3 for Rethinking Evaluation of Sparse Autoencoders through the Representation of Polysemous Words
Figure 4 for Rethinking Evaluation of Sparse Autoencoders through the Representation of Polysemous Words
Viaarxiv icon

ADOPT: Modified Adam Can Converge with Any $β_2$ with the Optimal Rate

Add code
Nov 05, 2024
Figure 1 for ADOPT: Modified Adam Can Converge with Any $β_2$ with the Optimal Rate
Figure 2 for ADOPT: Modified Adam Can Converge with Any $β_2$ with the Optimal Rate
Figure 3 for ADOPT: Modified Adam Can Converge with Any $β_2$ with the Optimal Rate
Figure 4 for ADOPT: Modified Adam Can Converge with Any $β_2$ with the Optimal Rate
Viaarxiv icon

Which Programming Language and What Features at Pre-training Stage Affect Downstream Logical Inference Performance?

Add code
Oct 09, 2024
Figure 1 for Which Programming Language and What Features at Pre-training Stage Affect Downstream Logical Inference Performance?
Figure 2 for Which Programming Language and What Features at Pre-training Stage Affect Downstream Logical Inference Performance?
Figure 3 for Which Programming Language and What Features at Pre-training Stage Affect Downstream Logical Inference Performance?
Figure 4 for Which Programming Language and What Features at Pre-training Stage Affect Downstream Logical Inference Performance?
Viaarxiv icon

Answer When Needed, Forget When Not: Language Models Pretend to Forget via In-Context Knowledge Unlearning

Add code
Oct 01, 2024
Figure 1 for Answer When Needed, Forget When Not: Language Models Pretend to Forget via In-Context Knowledge Unlearning
Figure 2 for Answer When Needed, Forget When Not: Language Models Pretend to Forget via In-Context Knowledge Unlearning
Figure 3 for Answer When Needed, Forget When Not: Language Models Pretend to Forget via In-Context Knowledge Unlearning
Figure 4 for Answer When Needed, Forget When Not: Language Models Pretend to Forget via In-Context Knowledge Unlearning
Viaarxiv icon

Language Models Do Hard Arithmetic Tasks Easily and Hardly Do Easy Arithmetic Tasks

Add code
Jun 04, 2024
Figure 1 for Language Models Do Hard Arithmetic Tasks Easily and Hardly Do Easy Arithmetic Tasks
Figure 2 for Language Models Do Hard Arithmetic Tasks Easily and Hardly Do Easy Arithmetic Tasks
Figure 3 for Language Models Do Hard Arithmetic Tasks Easily and Hardly Do Easy Arithmetic Tasks
Figure 4 for Language Models Do Hard Arithmetic Tasks Easily and Hardly Do Easy Arithmetic Tasks
Viaarxiv icon