Update README.md
Browse files
README.md
CHANGED
|
@@ -36,7 +36,7 @@ The Tamil-Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an improved ins
|
|
| 36 |
Tamil LLM: A Breakthrough in Tamil Language Understanding In the realm of language models, the fine-tuned Tamil Mistral model represents a significant advancement. Unlike its English counterpart, the Tamil Mistral model is specifically tailored to comprehend and generate text in the Tamil language. This innovation addresses a critical gap, as the English Mistral model fails to effectively engage with Tamil, a language rich in culture and heritage. Through extensive fine-tuning with a base Tamil Mistral model, this iteration has been meticulously enhanced to grasp the nuances and intricacies of the Tamil language. As a result, we are delighted to present a revolutionary model that enables seamless interaction through text. Welcome to the future of conversational Tamil language processing with our instructive model.
|
| 37 |
|
| 38 |
# Dataset
|
| 39 |
-
|
| 40 |
|
| 41 |
# Training time
|
| 42 |
18 hrs to train on NVIDIA RTX A6000 48GB with batch size of 30
|
|
|
|
| 36 |
Tamil LLM: A Breakthrough in Tamil Language Understanding In the realm of language models, the fine-tuned Tamil Mistral model represents a significant advancement. Unlike its English counterpart, the Tamil Mistral model is specifically tailored to comprehend and generate text in the Tamil language. This innovation addresses a critical gap, as the English Mistral model fails to effectively engage with Tamil, a language rich in culture and heritage. Through extensive fine-tuning with a base Tamil Mistral model, this iteration has been meticulously enhanced to grasp the nuances and intricacies of the Tamil language. As a result, we are delighted to present a revolutionary model that enables seamless interaction through text. Welcome to the future of conversational Tamil language processing with our instructive model.
|
| 37 |
|
| 38 |
# Dataset
|
| 39 |
+
Tamil open instruct dataset (400k) instruction google translated
|
| 40 |
|
| 41 |
# Training time
|
| 42 |
18 hrs to train on NVIDIA RTX A6000 48GB with batch size of 30
|