-
SEA-LION: Southeast Asian Languages in One Network
Paper • 2504.05747 • Published -
Do Large Language Models Speak All Languages Equally? A Comparative Study in Low-Resource Settings
Paper • 2408.02237 • Published -
A Three-Pronged Approach to Cross-Lingual Adaptation with Multilingual LLMs
Paper • 2406.17377 • Published -
Democratizing LLMs for Low-Resource Languages by Leveraging their English Dominant Abilities with Linguistically-Diverse Prompts
Paper • 2306.11372 • Published
Collections
Discover the best community collections!
Collections including paper arxiv:2404.05829
-
PLaMo-100B: A Ground-Up Language Model Designed for Japanese Proficiency
Paper • 2410.07563 • Published • 2 -
LLM-jp: A Cross-organizational Project for the Research and Development of Fully Open Japanese LLMs
Paper • 2407.03963 • Published • 19 -
Tagengo: A Multilingual Chat Dataset
Paper • 2405.12612 • Published • 3 -
Continual Pre-Training for Cross-Lingual LLM Adaptation: Enhancing Japanese Language Capabilities
Paper • 2404.17790 • Published • 5
-
Rho-1: Not All Tokens Are What You Need
Paper • 2404.07965 • Published • 93 -
VASA-1: Lifelike Audio-Driven Talking Faces Generated in Real Time
Paper • 2404.10667 • Published • 23 -
Instruction-tuned Language Models are Better Knowledge Learners
Paper • 2402.12847 • Published • 26 -
DoRA: Weight-Decomposed Low-Rank Adaptation
Paper • 2402.09353 • Published • 30
-
SambaLingo: Teaching Large Language Models New Languages
Paper • 2404.05829 • Published • 13 -
sambanovasystems/SambaLingo-Arabic-Chat
Text Generation • 7B • Updated • 80 • 64 -
sambanovasystems/SambaLingo-Arabic-Base
Text Generation • 7B • Updated • 42 • 37 -
sambanovasystems/SambaLingo-Arabic-Base-70B
Text Generation • 69B • Updated • 29 • 1
-
CulturaX: A Cleaned, Enormous, and Multilingual Dataset for Large Language Models in 167 Languages
Paper • 2309.09400 • Published • 85 -
Tuning LLMs with Contrastive Alignment Instructions for Machine Translation in Unseen, Low-resource Languages
Paper • 2401.05811 • Published • 8 -
Is Preference Alignment Always the Best Option to Enhance LLM-Based Translation? An Empirical Analysis
Paper • 2409.20059 • Published • 17 -
Are Character-level Translations Worth the Wait? Comparing Character- and Subword-level Models for Machine Translation
Paper • 2302.14220 • Published
-
SEA-LION: Southeast Asian Languages in One Network
Paper • 2504.05747 • Published -
Do Large Language Models Speak All Languages Equally? A Comparative Study in Low-Resource Settings
Paper • 2408.02237 • Published -
A Three-Pronged Approach to Cross-Lingual Adaptation with Multilingual LLMs
Paper • 2406.17377 • Published -
Democratizing LLMs for Low-Resource Languages by Leveraging their English Dominant Abilities with Linguistically-Diverse Prompts
Paper • 2306.11372 • Published
-
Rho-1: Not All Tokens Are What You Need
Paper • 2404.07965 • Published • 93 -
VASA-1: Lifelike Audio-Driven Talking Faces Generated in Real Time
Paper • 2404.10667 • Published • 23 -
Instruction-tuned Language Models are Better Knowledge Learners
Paper • 2402.12847 • Published • 26 -
DoRA: Weight-Decomposed Low-Rank Adaptation
Paper • 2402.09353 • Published • 30
-
SambaLingo: Teaching Large Language Models New Languages
Paper • 2404.05829 • Published • 13 -
sambanovasystems/SambaLingo-Arabic-Chat
Text Generation • 7B • Updated • 80 • 64 -
sambanovasystems/SambaLingo-Arabic-Base
Text Generation • 7B • Updated • 42 • 37 -
sambanovasystems/SambaLingo-Arabic-Base-70B
Text Generation • 69B • Updated • 29 • 1
-
PLaMo-100B: A Ground-Up Language Model Designed for Japanese Proficiency
Paper • 2410.07563 • Published • 2 -
LLM-jp: A Cross-organizational Project for the Research and Development of Fully Open Japanese LLMs
Paper • 2407.03963 • Published • 19 -
Tagengo: A Multilingual Chat Dataset
Paper • 2405.12612 • Published • 3 -
Continual Pre-Training for Cross-Lingual LLM Adaptation: Enhancing Japanese Language Capabilities
Paper • 2404.17790 • Published • 5
-
CulturaX: A Cleaned, Enormous, and Multilingual Dataset for Large Language Models in 167 Languages
Paper • 2309.09400 • Published • 85 -
Tuning LLMs with Contrastive Alignment Instructions for Machine Translation in Unseen, Low-resource Languages
Paper • 2401.05811 • Published • 8 -
Is Preference Alignment Always the Best Option to Enhance LLM-Based Translation? An Empirical Analysis
Paper • 2409.20059 • Published • 17 -
Are Character-level Translations Worth the Wait? Comparing Character- and Subword-level Models for Machine Translation
Paper • 2302.14220 • Published