Picture for Linhao Yu

Linhao Yu

Large Language Model Safety: A Holistic Survey

Add code
Dec 23, 2024
Figure 1 for Large Language Model Safety: A Holistic Survey
Figure 2 for Large Language Model Safety: A Holistic Survey
Figure 3 for Large Language Model Safety: A Holistic Survey
Figure 4 for Large Language Model Safety: A Holistic Survey
Viaarxiv icon

Self-Pluralising Culture Alignment for Large Language Models

Add code
Oct 16, 2024
Viaarxiv icon

CMoralEval: A Moral Evaluation Benchmark for Chinese Large Language Models

Add code
Aug 19, 2024
Figure 1 for CMoralEval: A Moral Evaluation Benchmark for Chinese Large Language Models
Figure 2 for CMoralEval: A Moral Evaluation Benchmark for Chinese Large Language Models
Figure 3 for CMoralEval: A Moral Evaluation Benchmark for Chinese Large Language Models
Figure 4 for CMoralEval: A Moral Evaluation Benchmark for Chinese Large Language Models
Viaarxiv icon

LFED: A Literary Fiction Evaluation Dataset for Large Language Models

Add code
May 16, 2024
Figure 1 for LFED: A Literary Fiction Evaluation Dataset for Large Language Models
Figure 2 for LFED: A Literary Fiction Evaluation Dataset for Large Language Models
Figure 3 for LFED: A Literary Fiction Evaluation Dataset for Large Language Models
Figure 4 for LFED: A Literary Fiction Evaluation Dataset for Large Language Models
Viaarxiv icon

OpenEval: Benchmarking Chinese LLMs across Capability, Alignment and Safety

Add code
Mar 18, 2024
Figure 1 for OpenEval: Benchmarking Chinese LLMs across Capability, Alignment and Safety
Figure 2 for OpenEval: Benchmarking Chinese LLMs across Capability, Alignment and Safety
Figure 3 for OpenEval: Benchmarking Chinese LLMs across Capability, Alignment and Safety
Figure 4 for OpenEval: Benchmarking Chinese LLMs across Capability, Alignment and Safety
Viaarxiv icon

Identifying Multiple Personalities in Large Language Models with External Evaluation

Add code
Feb 22, 2024
Figure 1 for Identifying Multiple Personalities in Large Language Models with External Evaluation
Figure 2 for Identifying Multiple Personalities in Large Language Models with External Evaluation
Figure 3 for Identifying Multiple Personalities in Large Language Models with External Evaluation
Figure 4 for Identifying Multiple Personalities in Large Language Models with External Evaluation
Viaarxiv icon

Evaluating Large Language Models: A Comprehensive Survey

Add code
Oct 31, 2023
Viaarxiv icon

M3KE: A Massive Multi-Level Multi-Subject Knowledge Evaluation Benchmark for Chinese Large Language Models

Add code
May 21, 2023
Figure 1 for M3KE: A Massive Multi-Level Multi-Subject Knowledge Evaluation Benchmark for Chinese Large Language Models
Figure 2 for M3KE: A Massive Multi-Level Multi-Subject Knowledge Evaluation Benchmark for Chinese Large Language Models
Figure 3 for M3KE: A Massive Multi-Level Multi-Subject Knowledge Evaluation Benchmark for Chinese Large Language Models
Figure 4 for M3KE: A Massive Multi-Level Multi-Subject Knowledge Evaluation Benchmark for Chinese Large Language Models
Viaarxiv icon