Research2026-04-22
CulturALL: Benchmarking Multilingual and Multicultural Competence of LLMs on Grounded Tasks
Source: Arxiv CS.AI
arXiv:2604.19262v1 Announce Type: cross Abstract: Large language models (LLMs) are now deployed worldwide, inspiring a surge of benchmarks that measure their multilingual and multicultural abilities. However, these benchmarks prioritize generic language understanding or superficial cultural trivia,...
arxivpapersbenchmark