SeaLLMs 3: Open Foundation and Chat Multilingual Large Language Models for Southeast Asian Languages (2024)

Wenxuan Zhang, Hou Pong Chan, Yiran Zhao, Mahani Aljunied, Jianyu Wang,
Chaoqun Liu, Yue Deng, Zhiqiang Hu, Weiwen Xu, Yew Ken Chia, Xin Li, Lidong Bing

DAMO Academy, Alibaba Group

Project page: https://seallms.github.io/

Equal contributions. Corresponding author: l.bing@alibaba-inc.com

Abstract

Large Language Models (LLMs) have shown remarkable abilities across various tasks, yet their development has predominantly centered on high-resource languages like English and Chinese, leaving low-resource languages underserved. To address this disparity, we present SeaLLMs 3, the latest iteration of the SeaLLMs model family, tailored for Southeast Asian languages. This region, characterized by its rich linguistic diversity, has lacked adequate language technology support. SeaLLMs 3 aims to bridge this gap by covering a comprehensive range of languages spoken in this region, including English, Chinese, Indonesian, Vietnamese, Thai, Tagalog, Malay, Burmese, Khmer, Lao, Tamil, and Javanese. Leveraging efficient language enhancement techniques and a specially constructed instruction tuning dataset, SeaLLMs 3 significantly reduces training costs while maintaining high performance and versatility. Our model excels in tasks such as world knowledge, mathematical reasoning, translation, and instruction following, achieving state-of-the-art performance among similarly sized models. Additionally, we prioritized safety and reliability by addressing both general and culture-specific considerations and incorporated mechanisms to reduce hallucinations. This work underscores the importance of inclusive AI, showing that advanced LLM capabilities can benefit underserved linguistic and cultural communities.

SeaLLMs 3: Open Foundation and Chat Multilingual Large Language Models for Southeast Asian Languages


Wenxuan Zhangthanks: Equal contributions., Hou Pong Chan, Yiran Zhao, Mahani Aljunied, Jianyu Wang,Chaoqun Liu, Yue Deng, Zhiqiang Hu, Weiwen Xu, Yew Ken Chia, Xin Li, Lidong Bingthanks: Corresponding author: l.bing@alibaba-inc.comDAMO Academy, Alibaba GroupProject page: https://seallms.github.io/


1 Introduction

Large Language Models (LLMs) such as GPT-4 OpenAI (2023) and Gemini Anil etal. (2023) have demonstrated remarkable capabilities across a wide array of tasks, ranging from natural language understanding and generation to more specialized domain applications Zhao etal. (2023). These models have proven valuable, offering substantial benefits to the global community, especially through the proliferation of open-source LLMs such as Llama Touvron etal. (2023a, b), Mistral Jiang etal. (2023), Qwen Bai etal. (2023); Yang etal. (2024), and Gemma Mesnard etal. (2024).However, the majority of these efforts have been concentrated on high-resource languages such as English and Chinese, or well-developed regions like Europe Zhang etal. (2023); Ahuja etal. (2023). Consequently, the development of LLMs tailored for low-resource languages or underdeveloped regions has been significantly overlooked, resulting in a lack of inclusivity and equitable distribution of AI advancements across diverse linguistic and cultural communities Qin etal. (2024); Huang etal. (2024); Liu etal. (2024).

To bridge this gap, we introduced the SeaLLMs model Nguyen etal. (2023c), specifically designed LLMs for Southeast Asian languages. Southeast Asia (SEA) is a region with a rich diversity of languages spoken by millions of people, yet it suffers from a significant lack of language technology support Aji etal. (2022).The initiative of SeaLLMs thus aims to make the benefits of LLMs accessible to speakers of these languages, addressing their unique linguistic and cultural nuances.Following this endeavor, several other models have been dedicated to this region as well, such as SEA-LION AI Singapore (2023) and Sailor Dou etal. (2024).However, these models often face significant limitations: they are typically released only as foundational or chat models, offer limited options in terms of model size, and cover a limited number of SEA languages. Moreover, the relatively scarce availability of language corpora further constrains the amount of training data available, hindering the development and performance of these models.

In this work, we introduce SeaLLMs 3, the latest iteration of the SeaLLMs model family. This version is designed to cover a more diverse array of Southeast Asian languages, including English, Chinese, Indonesian, Vietnamese, Thai, Tagalog, Malay, Burmese, Khmer, Lao, Tamil, and Javanese.Different from the conventional continue-pretraining paradigm Zhao etal. (2024a); Nguyen etal. (2023c); Dou etal. (2024), we conduct efficient language enhancement by training language-specific neurons only based on a foundation model Zhao etal. (2024b), significantly reducing the overall training cost. Moreover, such targeted training also ensures that the performance of high-resource languages can remain unaffected during the enhancement.Furthermore, SeaLLMs 3 is trained using a specially constructed instruction tuning dataset that encompasses a wide variety of task types and carefully balanced language distributions. This approach ensures that the model can handle the linguistic diversity of the Southeast Asian region while maintaining high performance and versatility across different applications.As a result, it achieves state-of-the-art performance among models with similar sizes, excelling across a diverse array of tasks such as world knowledge Zhang etal. (2023), mathematical reasoning Shi etal. (2023), translation Costa-jussà etal. (2022), and instruction following.

In the meantime, we pay special attention to the model’s reliability and trustworthiness during its development, which are often under-considered in multilingual settings Deng etal. (2023). In particular, we address both general and culture-specific safety considerations to ensure the models provide contextually appropriate responses. The model is also specifically trained to be aware of its knowledge boundary and refuse what it does not know. To evaluate this capability, we introduce a novel benchmark, SeaRefuse, which assesses the ability of LLMs to refuse questions beyond their knowledge boundaries. This focus on safety and reliability has resulted in SeaLLMs 3 exhibiting reduced hallucination and delivering safe, coherent responses, especially in queries closely related to Southeast Asian culture.

We open-source both the foundational and chat models of SeaLLMs 3111https://huggingface.co/collections/SeaLLMs/seallms-v3-668f3a52e1e6fbaad5752cdb. The foundational model could serve as a base for conducting instruction tuning tailored to specific application requirements. Meanwhile, the chat model has already undergone instruction tuning and is ready for direct use of handling a wide range of tasks.

SeaLLMs 3: Open Foundation and Chat Multilingual Large Language Models for Southeast Asian Languages (1)

2 Pre-training

2.1 Pre-training Data

Building on the efforts from previous versions of SeaLLMs Nguyen etal. (2023c), we have incorporated corpora from a wider range of data sources to enhance diversity.Specifically, we have integrated fundamental knowledge from Wikipedia (Foundation, ) and textbooks (BenAllal etal., 2024), journalistic materials such as CC-News (Crawl, ), web-based corpora from CulturaX (Nguyen etal., 2023a), and MADLAD-400 (Kudugunta etal., 2023).We have also improved the data processing pipeline including the language model filtering and duplicate removal to improve the data quality.

Additionally, we explore the utilization of model-synthetic data for training, which received much attention recently BenAllal etal. (2024). Starting with manual annotation of domain-specific knowledge points in SEA languages, we then employed stronger models to generate targeted tutorial-style content, thereby enhancing SeaLLMs 3 with enriched regional knowledge in a more explicit form.

2.2 Language-Specific Neuron Training

We built our model based on the Qwen2 model family Yang etal. (2024) and further conducted language enhancement to augment its capability in SEA languages. This approach allows the model to quickly inherit foundational knowledge from Qwen, rather than learning it from scratch.

The most straightforward method for language enhancement is typically through continued pertaining Zhao etal. (2024a), which we also used for previous versions of SeaLLMs. However, as discussed, the relatively scarce availability of language corpora limits the amount of training data, hindering the development and performance of these models. Furthermore, it is often observed that such direct continued pretraining can compromise the model’s original capacity in high-resource languages like English and Chinese Dou etal. (2024).

In this iteration, we adopt Language-Specific Neuron (LSN) training for efficient language enhancement, as shown in Figure 1. Recent studies have found that certain language-specific neurons in language models are responsible for processing specific languages. For instance, Zhao etal. (2024b) discovered that language-specific neurons comprise only about 0.1% of all parameters. Thus, the capabilities of a language can be enhanced by training its corresponding LSNs while preserving the multilingual abilities of other languages.To efficiently train SeaLLMs 3, we employ the parallel language-specific neuron detection method proposed by Zhao etal. (2024b). As shown in the left most part of Figure 1, this method allows us to identify the LSNs of SEA languages using language-specific training data, selected as a downsampled subset of the corpora from the training data. We then specifically train these detected LSNs to develop multiple monolingual LLMs in SEA languages, which are subsequently merged to create a unified multilingual LLM for SEA languages. Additionally, to maintain proficiency in English and Chinese from the original foundation model, we detect their respective LSNs and exclude them from the entire pre-training process.

This method offers several advantages. First, it requires relatively less training data since the training is more targeted, which significantly reduces training costs. Second, because the training is targeted, we can ensure that the performance of high-resource languages from the original foundation model remains unaffected. LSNs operate independently and do not influence one another, avoiding the sacrifices seen with previous methods.

3 Supervised Fine-Tuning (SFT)

3.1 Supervised Fine-tuning Data

Most existing open-source supervised fine-tuning (SFT) datasets are predominantly in English Wei etal. (2022); Taori etal. (2023), which presents a challenge for developing effective models for Southeast Asian languages. To address this, we employed various techniques to construct our SFT pool. For example, we selectively translate some high-quality English data to SEA languages with quality filtering, conduct self-instruction to automatically generate SFT data of certain types, and use various prompting strategies Madaan etal. (2023); Nguyen etal. (2023b).Following our previous practice, native speakers have been actively engaged throughout the entire SFT data construction process. They manually collect and write seed questions and topic lists, ensuring linguistic and cultural accuracy from the outset. Additionally, native speakers verify, filter, and edit the synthetic SFT data to maintain high quality.

SeaLLMs 3: Open Foundation and Chat Multilingual Large Language Models for Southeast Asian Languages (2)

Our preliminary experiments indicated that relying heavily on dominant English data adversely affects performance. To mitigate this, we strive to maintain a relatively good balance of language representation in our training data this time. Figure 2 shows the language distribution of our SFT data. While English remains a significant portion of the dataset, substantial representation is given to other Southeast Asian languages such as Indonesian, Vietnamese, Thai, and others, ensuring a comprehensive and diverse linguistic foundation for the model’s training.

Since the first release of SeaLLMs Nguyen etal. (2023c), the task types of SFT data have been significantly expanded. The dataset now includes a diverse range of task types such as coding, math, education-related content, reasoning, general dialogue, table-related tasks, open-domain QA, and many more. This expansion ensures that the model is well-rounded and capable of handling a variety of queries and tasks. Additionally, SFT with multiple turns has been significantly increased to enhance the model’s ability to engage in natural, multi-turn dialogues, improving its conversational fluency and coherence.

Modelenzhidthviavgavg_sea
Gemma-7B73.251.947.546.059.455.651.0
Sailor-7B-Chat66.065.247.546.251.355.248.3
SeaLLM-7B-v2.575.858.149.950.262.259.254.1
Sailor-14B74.884.053.652.862.165.556.2
Sailor-14B-Chat74.984.355.356.663.767.058.5
Qwen2-7B81.587.453.047.962.866.554.6
Qwen2-7B-Instruct80.988.055.855.562.468.557.9
SeaLLMs-v3-7B80.986.354.553.062.867.556.8
SeaLLMs-v3-7B-Chat80.987.455.856.964.969.259.2

Model safety, trustworthiness, and reliability are also important factors for constructing the SFT pool. To address this, we specifically constructed refusal-type data, enabling the model to decline questions beyond its knowledge boundaries, such as those involving non-existing entities. Furthermore, we carefully curated safety-related data, including both general safety data (which are culturally independent, such as general moral principles) and country-specific safety data (which are culturally sensitive). This approach ensures that the model can be safely deployed with cultural considerations in mind, providing accurate and appropriate responses across different cultural contexts.

3.2 Training Details

Two stages of training are employed to optimize the model’s performance. In the first stage, a large volume of SFT data is used to equip the model with instruction-following capabilities and to familiarize it with different task types. In the second stage, a smaller but high-quality SFT dataset is utilized to fine-tune the model, ensuring it performs exceptionally well on important tasks.

During the training process, different samples are packed together for efficiency, with a maximum length set at 8,192 tokens. The learning rate is set at 1.0e-5, with a warmup ratio of 0.1. Additionally, gradients are clipped at a maximum of 1.0 to prevent exploding gradients.

Modelenzhidthviavgavg_sea
Gemma-7B63.450.954.549.049.453.551.0
Sailor-7B-Chat55.847.248.441.446.247.845.4
SeaLLM-7B-v2.565.254.456.547.952.855.352.4
Sailor-14B61.856.457.048.253.555.452.9
Sailor-14B-Chat62.756.156.749.654.155.853.5
Qwen2-7B71.064.260.252.056.660.856.3
Qwen2-7B-Instruct70.863.559.952.456.860.756.4
SeaLLMs-v3-7B70.665.461.753.658.762.058.0
SeaLLMs-v3-7B-Chat71.364.762.554.457.862.258.2

4 Evaluations

We conduct extensive evaluations against models with similar sizes, including Sailor-7B / Sailor-7B-Chat Dou etal. (2024), Gemma-7b / Gemma-7b-it Mesnard etal. (2024), Qwen2-7B / Qwen2-7B-Chat Yang etal. (2024), Meta-Llama-3-8B / Meta-Llama-3-8B-Instruct Touvron etal. (2023b), Aya-23-8B Aryabumi etal. (2024), and the previous versions (mainly v2.5) of the SeaLLMs Nguyen etal. (2023c). The models are listed by their release date in the following result tables.

The evaluations can be generally categorized into the following two dimensions:

  • Model Capability: We assess the model’s performance on human exam questions, its proficiency in mathematics, its ability to follow multi-turn instructions, and its translation accuracy of different language pairs.

  • Model Trustworthiness: We evaluate the model’s safety and tendency to hallucinate, particularly in the context of Southeast Asia.

4.1 Model Capability

4.1.1 Multilingual World Knowledge

Dataset

We utilized the M3Exam dataset Zhang etal. (2023), comprising real human exam questions collected from different countries and spanning different subjects and educational stages. This dataset effectively tests the model’s multilingual world knowledge in a manner more akin to real-world settings. We take the questions in English (en), Chinese (zh), Indonesian (id), Vietnamese (vi), and Thai (th). We also employ the translated MMLU Hendrycks etal. (2021) questions for evaluation, which primarily tests the cross-lingual alignment of the model as the required knowledge is still mainly Western-focused. For evaluation, we employ the SeaExam toolkit222https://github.com/DAMO-NLP-SG/SeaExam and measured performance using accuracy as the metric.

Results

As shown in Table 1 for results on M3Exam dataset, our models, SeaLLMs-v3-7B and SeaLLMs-v3-7B-Chat, demonstrate competitive performance, with SeaLLMs-v3-7B-Chat achieving the highest average score (0.692) and the highest average score for SEA languages (0.592).Compared to the previous version, SeaLLMs-7B-v2.5, our latest models show significant improvement in overall performance specifically in handling Southeast Asian languages. Furthermore, while the Qwen2-7B-Instruct model performs exceptionally well in English and Chinese, our models exhibit superior performance across a broader range of Southeast Asian languages, highlighting their enhanced multilingual capabilities.

Table 2 shows the results of different models on the translated MMLU dataset, we can see that our SeaLLMs-v3-7B and SeaLLMs-v3-7B-Chat models also outperform other models, particularly in Southeast Asian languages.Compared to the previous SeaLLMs-7B-v2.5 version, our latest models show substantial improvements, particularly in handling Southeast Asian languages (with avg_sea from 52.4 to 58.2).

MGSMenidmsthvizhavg
Few-shot setting
Gemma-7B64.841.243.238.034.039.643.5
Sailor-7B34.425.222.824.822.426.426.0
Meta-Llama-3-8B56.836.033.634.833.643.639.7
GLM-4-9B78.053.657.246.056.869.660.2
Qwen2-7B79.658.856.854.854.869.262.3
SeaLLMs-v3-7B78.859.256.856.854.872.063.1
Zero-shot setting
Gemma-1.1-7B-it58.832.434.831.239.635.238.7
Sailor-7B-Chat33.622.422.421.625.229.225.7
SeaLLM-7B-v2.579.669.270.861.266.862.468.3
Meta-Llama-3-8B-Instruct77.648.057.656.046.858.857.5
GLM-4-9B-Chat72.853.653.634.852.470.856.3
Qwen2-7B-Instruct82.066.462.458.464.476.868.4
SeaLLMs-v3-7B-Chat74.871.270.871.271.279.673.1

Modelidthviavg
turn1turn2avgturn1turn2avgturn1turn2avg
Sailor-7B-Chat4.604.044.323.943.173.564.823.624.224.03
SeaLLM-7B-v2.56.274.965.625.793.824.816.024.025.025.15
Sailor-14B-Chat5.265.535.404.624.364.495.314.745.034.97
Qwen2-7B-Instruct5.935.845.895.475.205.346.175.605.895.70
SeaLLMs-v3-7B-Chat6.736.596.666.485.906.196.345.796.076.31

Modelenidjvkmlomsmytathtlvizhavg
Sailor-7B-Chat49.4049.7828.332.686.8547.755.3518.2338.9229.0041.7620.8728.24
SeaLLM-7B-v2.555.0953.7118.1318.0915.5351.3319.7126.1040.5545.5844.5624.1834.38
Meta-Llama-3-8B-Instruct51.5449.0322.4615.345.4246.7221.2432.0935.7540.8039.3114.8731.22
Qwen2-7B-Instruct50.3647.5529.3619.2611.0642.4319.3320.0436.0737.9139.6322.8731.32
SeaLLMs-v3-7B-Chat54.6852.5229.8627.3026.3445.0421.5431.9341.5238.5143.7826.1036.52

4.1.2 Multilingual Math

Dataset

We assess the multilingual mathematics capabilities using the MGSM dataset Shi etal. (2023). Originally, MGSM comprises testing samples solely in English, Chinese and Thai. To extend this dataset to other SEA languages, specifically Indonesian, Malay, and Vietnamese, we utilize Google Translate to translate the original English questions to those questions. It is important to note that in our translations, we adhere to the numerical notation conventions of each respective country. For instance, in both Indonesian and Vietnamese, dots are used as thousands separators, and commas as decimal separators, which is the reverse of the English numeral system. We follow the same convention when evaluating the model’s generations.

Results

Table 3 presents the evaluation results on the MGSM benchmark, under both the few-shot setting (for testing base versions) and the zero-shot setting (for testing chat versions).In the few-shot setting, SeaLLMs-v3-7B demonstrates the highest average score (63.1), outperforming other models such as Qwen2-7B (62.3) and GLM-4-9b (60.2), particularly excelling in Indonesian and Thai. In the zero-shot setting, the chat version of SeaLLMs-v3-7B-Chat achieves the highest average score (73.1), showing strong performance across all languages. This highlights SeaLLMs-v3-7B-Chat’s superior adaptability and robustness in multilingual math tasks compared to its counterparts like Qwen2-7B-instruct (68.4).

4.1.3 Multilingual Instruction-following

Dataset

As there is no publicly available dataset for testing the model’s multi-turn instruction-following capability in SEA languages, we construct our own benchmark, namely SeaBench333It will be publicly available at https://huggingface.co/datasets/SeaLLMs/SeaBench, for such evaluation.SeaBench consists of multi-turn human instructions spanning various task types for Indonesian, Vietnamese, and Thai. Following MT-Bench Zheng etal. (2023), we consider 8 task types including writing, roleplay, extraction, reasoning, math, coding, knowledge I (STEM), and knowledge II (humanities/social science). Additionally, considering the characteristics of the multilingual setting, we include two more task types: safety and life. The safety task tests whether the model will respond to unsafe queries in a local context, while the life task includes questions likely asked in real-life settings, which might be informally written or even ambiguous.

All questions are manually written by native speakers of each language. During construction, we instructed the annotators to ensure the questions were as localized as possible, e.g., using local entities, concepts, and knowledge in the questions. Furthermore, reference answers have been constructed to ensure fair judgment.

Evaluation Details

Given the two-turn questions, the model under testing generates two-turn responses in a multi-turn format. These responses are then graded by a stronger LLM (GPT-4o was used in our experiments) using the reference answer to the original questions. The scores are then assigned to each turn of the response.

Results

As shown in Table 4, SeaLLMs-v3-7B-Chat outperforms all other models in multilingual instruction-following across Indonesian (id), Thai (th), and Vietnamese (vi). It achieves the highest average scores in both individual turns and overall averages for each language. Specifically, SeaLLMs-v3-7B-Chat surpasses the previous version, SeaLLMs-7B-v2.5, by a significant margin (6.31 vs 5.15) and outperforms the strongest baseline model Qwen2-7B-Instruct (6.31 vs 5.70). These results highlight SeaLLMs-v3-7B-Chat’s superior ability to generate more coherent and contextually appropriate multi-turn responses.

Modelenzhvithidavgavg_sea
Gemma-1.1-7B-it53.6128.2226.1821.2830.3931.9425.95
Sailor-7B-Chat33.7618.825.199.6816.4216.7810.43
SeaLLM-7B-v2.513.101.533.2419.580.787.657.87
Meta-Llama-3-8B-Instruct72.230.001.230.803.9115.631.98
GLM-4-9B-Chat45.0240.9821.485.422.3723.059.76
Qwen2-7B-Instruct63.7435.7552.8646.4255.9350.9451.74
SeaLLMs-v3-7B-Chat71.1377.1778.1861.6467.6171.1469.14

Modelenzhvithidavgavg_sea
Gemma-1.1-7B-it51.9529.9212.0730.1635.4831.9225.90
Sailor-7B-Chat33.8715.7911.3212.9631.6721.1218.65
SeaLLM-7B-v2.510.913.9211.3249.6631.1521.3930.71
Meta-Llama-3-8B-Instruct73.170.000.000.0011.2116.883.74
GLM-4-9B-Chat35.3852.5020.175.779.5224.6711.82
Qwen2-7B-Instruct58.5042.7562.8260.5363.5157.6262.29
SeaLLMs-v3-7B-Chat68.5481.8283.8484.5889.6681.6986.03

4.1.4 Translation

Dataset

We evaluate the machine translation performances with the test set of Flores-200 Costa-jussà etal. (2022). We choose all 12 languages for a comprehensive evaluation, including Burmese (my), Chinese (zh), English (en), Indonesian (id), Javanese (jv), Khmer (km), Lao (lo), Malay (ms), Tagalog (tl), Tamil (ta), Thai (th), and Vietnamese (vi). We translate between each pair of languages and report the average 0-shot chrF scores after averaging the results for target languages.

Results

As shown in Table 5, SeaLLMs-v3-7B-Chat outperforms other models in machine translation, achieving an average chrF score of 36.52. It excels particularly in Javanese, Khmer, Lao, Burmese, Thai, and Chinese, consistently achieving the highest scores in these languages. Compared to its predecessor, SeaLLMs-7B-v2.5, which has an average score of 34.38, SeaLLMs-v3-7B-Chat shows clear improvement. Additionally, SeaLLMs-v3-7B-Chat surpasses strong baselines like Meta-Llama-3-8B-Instruct and Qwen2-7B-Instruct, with average scores of 31.22 and 31.32, respectively. Notably, the model’s performance in translating low-resource languages, such as Khmer (27.30) and Lao (26.34), highlights its robustness and effectiveness in handling diverse and challenging translation tasks. This consistent performance across multiple languages underscores the model’s versatility and capability in low-resource language translation settings.

4.2 Model Trustworthiness

4.2.1 Hallucination

A trustworthy LLM should only answer the questions that it knows and abstain from answering questions that it does not know. Previous studies reveal that recent LLMs are prone to answering questions that exceed their knowledge boundaries, leading to hallucinated responsesYang etal. (2023); Zhang etal. (2024). However, evaluating a model’s ability to refuse questions it doesn’t know is challenging. This requires distinguishing the model’s knowledge boundaries, which is difficult since most existing LLMs do not provide transparency in their pre-training data.

To address this challenge, we propose a novel SeaRefuse evaluation benchmark, which consists of unanswerable factoid questions about non-existing entities and answerable factoid questions in SEA languages444It will be publicly available at https://huggingface.co/datasets/SeaLLMs/SeaRefuse. Unanswerable questions about non-existent entities are designed to surpass the knowledge boundaries of LLMs.Our benchmark includes two test sets: SeaRefuse-G and SeaRefuse-H. In the SeaRefuse-G test set, the unanswerable questions are generated by prompting GPT-4o. In the SeaRefuse-H test set, our linguists annotate the unanswerable questions by refining machine-generated unanswerable questions. The answerable questions in both SeaRefuse-G and SeaRefuse-H are collected from existing factoid QA datasets including ParaRelZhang etal. (2024); Elazar etal. (2021), NLPCC-KBQADuan (2016); Duan and Tang (2018), and TyDi QAClark etal. (2020). For the SeaRefuse-G tese set, each language has 500 answerable and 500 unanswerable questions, except for Vietnamese, which has 483 of each. In the SeaRefuse-H dataset, each language contains 100 answerable and 100 unanswerable questions.

Answerable
question
Unanswerable
question
RefusedTrue-PositiveFalse-Positive
AnsweredFalse-NegativeTrue-Negative

In evaluation, we report the F1-score of each model on correctly refusing questions about non-existing entities. We follow the confusion matrix depicted in Table8 to compute the F1 scores.We adopt a keyword-matching approach to determine whether a model refuses to answer a factoid question. Specifically, we work with professional and native linguists to devise a set of refusal keywords for English, Chinese, Vietnamese, Indonesian, and Thai, respectively. If the generated response contains any of these refusal keywords, we assume that the response is a refusal response.

The experiment results on SeaRefuse-G and SeaRefuse-H are shown in Table6 and Table7, respectively. We observe that SeaLLMs-v3-7B-Chat outperforms all other baseline models by a large margin in zh, vi, th, and id languages.In English, the performance of SeaLLMs-v3-7B-Chat is competitive with Llama-3-8B-Instruct. These results demonstrate the capability of SeaLLMs-v3 to refuse questions that it does not know.

4.2.2 Safety

Modelenjvthvizhavg
Sailor-7B-Chat78.754.962.267.676.267.9
Meta-Llama-3-8B-Instruct88.326.471.169.877.166.5
Sailor-14B-Chat86.930.553.760.972.760.9
GLM-4-9B-Chat77.121.330.260.674.952.8
Qwen2-7B-Instruct88.643.863.873.087.371.3
SeaLLMs-v3-7B-Chat88.960.073.383.892.779.7

To evaluate the models’ safety capabilities, we use the questions of SEA languages from the MultiJail dataset Deng etal. (2023), which includes English (en), Javanese (jv), Thai (th), Vietnamese (vi), and Chinese (zh). Each question in the dataset is potentially malicious, and the model should refuse to answer them. To determine whether the model’s response is safe, we first translate the response into English and then prompt GPT-4o to check if the translated response is harmful. The results are reported as the safe rate of the responses.

Table 9 presents the safety capabilities of various models evaluated with the MultiJail dataset. Notably, SeaLLMs-v3-7B-Chat outperforms all other models with an average safe rate of 79.7%, demonstrating robust performance across all languages, particularly excelling in Vietnamese (83.8%) and Chinese (92.7%). In comparison, Qwen2-7B-Instruct follows with a distant second average of 71.3%, with its highest safe rate in Chinese (87.3%). Other models like Sailor-7B-Chat and Llama-3-8B-Instruct also show competitive performance but lag behind in consistency across languages.Notably, the exceptional performance of SeaLLMs-v3 in the three Southeast Asian languages (jv, th, and vi) underscores SeaLLM’s effective design, which caters to the linguistic nuances of this region.

5 Conclusion

SeaLLMs 3 represents a significant advancement in the development of large language models for Southeast Asian languages, addressing the region’s unique linguistic and cultural challenges. By adopting an efficient language enhancement approach and constructing a comprehensive instruction tuning dataset, SeaLLMs 3 achieves state-of-the-art performance while maintaining cost-effectiveness. Our commitment to reliability and safety, providing contextually appropriate responses, further strengthens the model’s applicability and trustworthiness. The open-sourcing of both foundational and chat models ensures that SeaLLMs 3 is accessible for a wide range of applications, fostering further innovation and inclusivity in AI development for Southeast Asia.

Acknowledgments

We would like to express our special thanks to our professional and native linguists, Tantong Champaiboon, Nguyen Ngoc Yen Nhi and Tara Devina Putri, who helped build, evaluate, and fact-check our sampled pretraining and SFT dataset as well as evaluating our models across different aspects, especially safety.

References

  • Ahuja etal. (2023)Kabir Ahuja, Harsh*ta Diddee, Rishav Hada, Millicent Ochieng, Krithika Ramesh, Prachi Jain, AkshayUttama Nambi, Tanuja Ganu, Sameer Segal, Mohamed Ahmed, Kalika Bali, and Sunayana Sitaram. 2023.MEGA: multilingual evaluation of generative AI.In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, pages 4232–4267.
  • AI Singapore (2023)AI Singapore. 2023.Sea-lion (southeast asian languages in one network): A family of large language models for southeast asia.https://github.com/aisingapore/sealion.
  • Aji etal. (2022)AlhamFikri Aji, GentaIndra Winata, Fajri Koto, Samuel Cahyawijaya, Ade Romadhony, Rahmad Mahendra, Kemal Kurniawan, David Moeljadi, RadityoEko Prasojo, Timothy Baldwin, JeyHan Lau, and Sebastian Ruder. 2022.One country, 700+ languages: NLP challenges for underrepresented languages and dialects in indonesia.In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, pages 7226–7249. Association for Computational Linguistics.
  • Anil etal. (2023)Rohan Anil, Sebastian Borgeaud, Yonghui Wu, Jean-Baptiste Alayrac, Jiahui Yu, Radu Soricut, Johan Schalkwyk, AndrewM. Dai, Anja Hauth, Katie Millican, David Silver, Slav Petrov, Melvin Johnson, Ioannis Antonoglou, Julian Schrittwieser, Amelia Glaese, Jilin Chen, Emily Pitler, TimothyP. Lillicrap, Angeliki Lazaridou, Orhan Firat, James Molloy, Michael Isard, PaulRonald Barham, Tom Hennigan, Benjamin Lee, Fabio Viola, Malcolm Reynolds, Yuanzhong Xu, Ryan Doherty, Eli Collins, Clemens Meyer, Eliza Rutherford, Erica Moreira, Kareem Ayoub, Megha Goel, George Tucker, Enrique Piqueras, Maxim Krikun, Iain Barr, Nikolay Savinov, Ivo Danihelka, Becca Roelofs, Anaïs White, Anders Andreassen, Tamara von Glehn, Lakshman Yagati, Mehran Kazemi, Lucas Gonzalez, Misha Khalman, Jakub Sygnowski, and etal. 2023.Gemini: A family of highly capable multimodal models.CoRR, abs/2312.11805.
  • Aryabumi etal. (2024)Viraat Aryabumi, John Dang, Dwarak Talupuru, Saurabh Dash, David Cairuz, Hangyu Lin, Bharat Venkitesh, Madeline Smith, JonAnder Campos, YiChern Tan, Kelly Marchisio, Max Bartolo, Sebastian Ruder, Acyr Locatelli, Julia Kreutzer, Nick Frosst, AidanN. Gomez, Phil Blunsom, Marzieh Fadaee, Ahmet Üstün, and Sara Hooker. 2024.Aya 23: Open weight releases to further multilingual progress.CoRR, abs/2405.15032.
  • Bai etal. (2023)Jinze Bai, Shuai Bai, Yunfei Chu, Zeyu Cui, Kai Dang, Xiaodong Deng, Yang Fan, Wenbin Ge, YuHan, Fei Huang, Binyuan Hui, Luo Ji, Mei Li, Junyang Lin, Runji Lin, Dayiheng Liu, Gao Liu, Chengqiang Lu, Keming Lu, Jianxin Ma, Rui Men, Xingzhang Ren, Xuancheng Ren, Chuanqi Tan, Sinan Tan, Jianhong Tu, Peng Wang, Shijie Wang, Wei Wang, Shengguang Wu, Benfeng Xu, Jin Xu, AnYang, Hao Yang, Jian Yang, Shusheng Yang, Yang Yao, Bowen Yu, Hongyi Yuan, Zheng Yuan, Jianwei Zhang, Xingxuan Zhang, Yichang Zhang, Zhenru Zhang, Chang Zhou, Jingren Zhou, Xiaohuan Zhou, and Tianhang Zhu. 2023.Qwen technical report.CoRR, abs/2309.16609.
  • BenAllal etal. (2024)Loubna BenAllal, Anton Lozhkov, Guilherme Penedo, Thomas Wolf, and Leandro von Werra. 2024.Cosmopedia.
  • Clark etal. (2020)JonathanH. Clark, Eunsol Choi, Michael Collins, Dan Garrette, Tom Kwiatkowski, Vitaly Nikolaev, and Jennimaria Palomaki. 2020.Tydi qa: A benchmark for information-seeking question answering in typologically diverse languages.Transactions of the Association for Computational Linguistics.
  • Costa-jussà etal. (2022)MartaR. Costa-jussà, James Cross, Onur Çelebi, Maha Elbayad, Kenneth Heafield, Kevin Heffernan, Elahe Kalbassi, Janice Lam, Daniel Licht, Jean Maillard, AnnaY. Sun, Skyler Wang, Guillaume Wenzek, AlYoungblood, Bapi Akula, Loïc Barrault, GabrielMejia Gonzalez, Prangthip Hansanti, John Hoffman, Semarley Jarrett, KaushikRam Sadagopan, Dirk Rowe, Shannon Spruit, Chau Tran, Pierre Andrews, NecipFazil Ayan, Shruti Bhosale, Sergey Edunov, Angela Fan, Cynthia Gao, Vedanuj Goswami, Francisco Guzmán, Philipp Koehn, Alexandre Mourachko, Christophe Ropers, Safiyyah Saleem, Holger Schwenk, and Jeff Wang. 2022.No language left behind: Scaling human-centered machine translation.CoRR, abs/2207.04672.
  • (10)Common Crawl.Common crawl news.
  • Deng etal. (2023)Yue Deng, Wenxuan Zhang, SinnoJialin Pan, and Lidong Bing. 2023.Multilingual jailbreak challenges in large language models.CoRR, abs/2310.06474.
  • Dou etal. (2024)Longxu Dou, Qian Liu, Guangtao Zeng, Jia Guo, Jiahui Zhou, Wei Lu, and Min Lin. 2024.Sailor: Open language models for south-east asia.CoRR, abs/2404.03608.
  • Duan (2016)Nan Duan. 2016.Overview of the nlpcc-iccpol 2016 shared task: Open domain chinese question answering.In Natural Language Understanding and Intelligent Applications, pages 942–948. Springer International Publishing.
  • Duan and Tang (2018)Nan Duan and Duyu Tang. 2018.Overview of the nlpcc 2017 shared task: Open domain chinese question answering.In Natural Language Processing and Chinese Computing, pages 954–961. Springer International Publishing.
  • Elazar etal. (2021)Yanai Elazar, Nora Kassner, Shauli Ravfogel, Abhilasha Ravichander, EduardH. Hovy, Hinrich Schütze, and Yoav Goldberg. 2021.Measuring and improving consistency in pretrained language models.Trans. Assoc. Comput. Linguistics, 9:1012–1031.
  • (16)Wikimedia Foundation.Wikimedia downloads.
  • Hendrycks etal. (2021)Dan Hendrycks, Collin Burns, Steven Basart, Andy Zou, Mantas Mazeika, Dawn Song, and Jacob Steinhardt. 2021.Measuring massive multitask language understanding.In 9th International Conference on Learning Representations, ICLR 2021. OpenReview.net.
  • Huang etal. (2024)Kaiyu Huang, Fengran Mo, Hongliang Li, You Li, Yuanchi Zhang, Weijian Yi, Yulong Mao, Jinchen Liu, Yuzhuang Xu, Jinan Xu, Jian-Yun Nie, and Yang Liu. 2024.A survey on large language models with multilingualism: Recent advances and new frontiers.CoRR, abs/2405.10936.
  • Jiang etal. (2023)AlbertQ. Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, DevendraSingh Chaplot, Diego deLasCasas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lucile Saulnier, LélioRenard Lavaud, Marie-Anne Lachaux, Pierre Stock, TevenLe Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, and WilliamEl Sayed. 2023.Mistral 7b.CoRR, abs/2310.06825.
  • Kudugunta etal. (2023)Sneha Kudugunta, Isaac Caswell, Biao Zhang, Xavier Garcia, ChristopherA. Choquette-Choo, Katherine Lee, Derrick Xin, Aditya Kusupati, Romi Stella, Ankur Bapna, and Orhan Firat. 2023.Madlad-400: A multilingual and document-level large audited dataset.Preprint, arXiv:2309.04662.
  • Liu etal. (2024)Chaoqun Liu, Wenxuan Zhang, Yiran Zhao, AnhTuan Luu, and Lidong Bing. 2024.Is translation all you need? A study on solving multilingual tasks with large language models.CoRR, abs/2403.10258.
  • Madaan etal. (2023)Aman Madaan, Niket Tandon, Prakhar Gupta, Skyler Hallinan, Luyu Gao, Sarah Wiegreffe, Uri Alon, Nouha Dziri, Shrimai Prabhumoye, Yiming Yang, Shashank Gupta, BodhisattwaPrasad Majumder, Katherine Hermann, Sean Welleck, Amir Yazdanbakhsh, and Peter Clark. 2023.Self-refine: Iterative refinement with self-feedback.In Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023.
  • Mesnard etal. (2024)Thomas Mesnard, Cassidy Hardin, Robert Dadashi, Surya Bhupatiraju, Shreya Pathak, Laurent Sifre, Morgane Rivière, MihirSanjay Kale, Juliette Love, Pouya Tafti, Léonard Hussenot, Aakanksha Chowdhery, Adam Roberts, Aditya Barua, Alex Botev, Alex Castro-Ros, Ambrose Slone, Amélie Héliou, Andrea Tacchetti, Anna Bulanova, Antonia Paterson, Beth Tsai, Bobak Shahriari, CharlineLe Lan, ChristopherA. Choquette-Choo, Clément Crepy, Daniel Cer, Daphne Ippolito, David Reid, Elena Buchatskaya, Eric Ni, Eric Noland, Geng Yan, George Tucker, George-Cristian Muraru, Grigory Rozhdestvenskiy, Henryk Michalewski, Ian Tenney, Ivan Grishchenko, Jacob Austin, James Keeling, Jane Labanowski, Jean-Baptiste Lespiau, Jeff Stanway, Jenny Brennan, Jeremy Chen, Johan Ferret, Justin Chiu, and etal. 2024.Gemma: Open models based on gemini research and technology.CoRR, abs/2403.08295.
  • Nguyen etal. (2023a)Thuat Nguyen, Chien VanNguyen, VietDac Lai, Hieu Man, NghiaTrung Ngo, Franck Dernoncourt, RyanA Rossi, and ThienHuu Nguyen. 2023a.Culturax: A cleaned, enormous, and multilingual dataset for large language models in 167 languages.arXiv preprint arXiv:2309.09400.
  • Nguyen etal. (2023b)Xuan-Phi Nguyen, SharifahMahani Aljunied, Shafiq Joty, and Lidong Bing. 2023b.Democratizing llms for low-resource languages by leveraging their english dominant abilities with linguistically-diverse prompts.CoRR, abs/2306.11372.
  • Nguyen etal. (2023c)Xuan-Phi Nguyen, Wenxuan Zhang, Xin Li, Mahani Aljunied, Qingyu Tan, Liying Cheng, Guanzheng Chen, Yue Deng, Sen Yang, Chaoqun Liu, Hang Zhang, and Lidong Bing. 2023c.Seallms - large language models for southeast asia.CoRR, abs/2312.00738.
  • OpenAI (2023)OpenAI. 2023.GPT-4 technical report.CoRR, abs/2303.08774.
  • Qin etal. (2024)Libo Qin, Qiguang Chen, Yuhang Zhou, Zhi Chen, Yinghui Li, Lizi Liao, Min Li, Wanxiang Che, and PhilipS. Yu. 2024.Multilingual large language model: A survey of resources, taxonomy and frontiers.CoRR, abs/2404.04925.
  • Shi etal. (2023)Freda Shi, Mirac Suzgun, Markus Freitag, Xuezhi Wang, Suraj Srivats, Soroush Vosoughi, HyungWon Chung, YiTay, Sebastian Ruder, Denny Zhou, Dipanjan Das, and Jason Wei. 2023.Language models are multilingual chain-of-thought reasoners.In The Eleventh International Conference on Learning Representations, ICLR 2023. OpenReview.net.
  • Taori etal. (2023)Rohan Taori, Ishaan Gulrajani, Tianyi Zhang, Yann Dubois, Xuechen Li, Carlos Guestrin, Percy Liang, and TatsunoriB. Hashimoto. 2023.Stanford alpaca: An instruction-following llama model.
  • Touvron etal. (2023a)Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, Aurélien Rodriguez, Armand Joulin, Edouard Grave, and Guillaume Lample. 2023a.Llama: Open and efficient foundation language models.CoRR, abs/2302.13971.
  • Touvron etal. (2023b)Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Dan Bikel, Lukas Blecher, Cristian Canton-Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez, Madian Khabsa, Isabel Kloumann, Artem Korenev, PunitSingh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushkar Mishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, EricMichael Smith, Ranjan Subramanian, XiaoqingEllen Tan, Binh Tang, Ross Taylor, Adina Williams, JianXiang Kuan, Puxin Xu, Zheng Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aurélien Rodriguez, Robert Stojnic, Sergey Edunov,and Thomas Scialom. 2023b.Llama 2: Open foundation and fine-tuned chat models.CoRR, abs/2307.09288.
  • Wei etal. (2022)Jason Wei, Maarten Bosma, VincentY. Zhao, Kelvin Guu, AdamsWei Yu, Brian Lester, Nan Du, AndrewM. Dai, and QuocV. Le. 2022.Finetuned language models are zero-shot learners.In The Tenth International Conference on Learning Representations, ICLR 2022. OpenReview.net.
  • Yang etal. (2024)AnYang, Baosong Yang, Binyuan Hui, BoZheng, Bowen Yu, Chang Zhou, Chengpeng Li, Chengyuan Li, Dayiheng Liu, Fei Huang, etal. 2024.Qwen2 technical report.arXiv preprint arXiv:2407.10671.
  • Yang etal. (2023)Yuqing Yang, Ethan Chern, Xipeng Qiu, Graham Neubig, and Pengfei Liu. 2023.Alignment for honesty.CoRR, abs/2312.07000.
  • Zhang etal. (2024)Hanning Zhang, Shizhe Diao, Yong Lin, YiFung, Qing Lian, Xingyao Wang, Yangyi Chen, Heng Ji, and Tong Zhang. 2024.R-tuning: Instructing large language models to say ‘I don’t know’.In Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), pages 7113–7139, Mexico City, Mexico. Association for Computational Linguistics.
  • Zhang etal. (2023)Wenxuan Zhang, Mahani Aljunied, Chang Gao, YewKen Chia, and Lidong Bing. 2023.M3exam: A multilingual, multimodal, multilevel benchmark for examining large language models.In Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023.
  • Zhao etal. (2024a)Jun Zhao, Zhihao Zhang, Luhui Gao, QiZhang, Tao Gui, and Xuanjing Huang. 2024a.Llama beyond english: An empirical study on language capability transfer.CoRR, abs/2401.01055.
  • Zhao etal. (2023)WayneXin Zhao, Kun Zhou, Junyi Li, Tianyi Tang, Xiaolei Wang, Yupeng Hou, Yingqian Min, Beichen Zhang, Junjie Zhang, Zican Dong, Yifan Du, Chen Yang, Yushuo Chen, Zhipeng Chen, Jinhao Jiang, Ruiyang Ren, Yifan Li, Xinyu Tang, Zikang Liu, Peiyu Liu, Jian-Yun Nie, and Ji-Rong Wen. 2023.A survey of large language models.CoRR, abs/2303.18223.
  • Zhao etal. (2024b)Yiran Zhao, Wenxuan Zhang, Guizhen Chen, Kenji Kawaguchi, and Lidong Bing. 2024b.How do large language models handle multilingualism?CoRR, abs/2402.18815.
  • Zheng etal. (2023)Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, ZiLin, Zhuohan Li, Dacheng Li, EricP. Xing, Hao Zhang, JosephE. Gonzalez, and Ion Stoica. 2023.Judging llm-as-a-judge with mt-bench and chatbot arena.In Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023.
SeaLLMs 3: Open Foundation and Chat Multilingual Large Language Models for Southeast Asian Languages (2024)
Top Articles
All Obituaries | Weber & Rodney Funeral Home | Edwardsville IL funeral home and cremation
Herald and Review from Decatur, Illinois
Mickey Moniak Walk Up Song
Global Foods Trading GmbH, Biebesheim a. Rhein
Cars & Trucks - By Owner near Kissimmee, FL - craigslist
Rainfall Map Oklahoma
How Many Cc's Is A 96 Cubic Inch Engine
Money blog: Domino's withdraws popular dips; 'we got our dream £30k kitchen for £1,000'
Bfg Straap Dead Photo Graphic
Directions To 401 East Chestnut Street Louisville Kentucky
Truth Of God Schedule 2023
Lancasterfire Live Incidents
Roof Top Snipers Unblocked
Jayah And Kimora Phone Number
Lcwc 911 Live Incident List Live Status
H12 Weidian
91 East Freeway Accident Today 2022
Metro Pcs.near Me
Drift Boss 911
What Channel Is Court Tv On Verizon Fios
Violent Night Showtimes Near Century 14 Vallejo
Titanic Soap2Day
Is Windbound Multiplayer
‘The Boogeyman’ Review: A Minor But Effectively Nerve-Jangling Stephen King Adaptation
At&T Outage Today 2022 Map
Walgreens Bunce Rd
Timeline of the September 11 Attacks
Relaxed Sneak Animations
Is Light Raid Hard
10 Best Places to Go and Things to Know for a Trip to the Hickory M...
Maths Open Ref
Martins Point Patient Portal
Craigslistodessa
How Much Is An Alignment At Costco
Steven Batash Md Pc Photos
Wsbtv Fish And Game Report
Compare Plans and Pricing - MEGA
Überblick zum Barotrauma - Überblick zum Barotrauma - MSD Manual Profi-Ausgabe
Culver's of Whitewater, WI - W Main St
Ashoke K Maitra. Adviser to CMD's. Received Lifetime Achievement Award in HRD on LinkedIn: #hr #hrd #coaching #mentoring #career #jobs #mba #mbafreshers #sales…
Ferguson Employee Pipeline
F9 2385
Gfs Ordering Online
Windshield Repair & Auto Glass Replacement in Texas| Safelite
Differential Diagnosis
Fedex Passport Locations Near Me
Does Target Have Slime Lickers
St Vrain Schoology
Skyward Cahokia
1Tamilmv.kids
Research Tome Neltharus
Sam's Club Fountain Valley Gas Prices
Latest Posts
Article information

Author: Chrissy Homenick

Last Updated:

Views: 6067

Rating: 4.3 / 5 (54 voted)

Reviews: 85% of readers found this page helpful

Author information

Name: Chrissy Homenick

Birthday: 2001-10-22

Address: 611 Kuhn Oval, Feltonbury, NY 02783-3818

Phone: +96619177651654

Job: Mining Representative

Hobby: amateur radio, Sculling, Knife making, Gardening, Watching movies, Gunsmithing, Video gaming

Introduction: My name is Chrissy Homenick, I am a tender, funny, determined, tender, glorious, fancy, enthusiastic person who loves writing and wants to share my knowledge and understanding with you.