The very first and largest usable Burmese Language GPT in Myanmar

MyanmarGPT is the very first and largest usable Burmese language GPT in Myanmar with strong community contributions. It was created by me, Min Si Thu.
These two models are trained by using private property datasets, manually cleaned by Min Si Thu.
There are two versions of MyanmarGPT at the moment, 2023 December.
Extended, released in 2024, January 28.
Released in 2024, February 23.
MyanmarGPT is 128 million parameters Burmese Language Model. It is very lightweight and easy to use on all devices.
MyanmarGPT-Big is a 1.42 billion parameters Multi-Language Model. It is an enterprise-level LLM for Burmese Language mainly and other languages. Currently supports 61 Languages.
Fine-tuned on MyanmarGPT, question answering model for the Burmese language. With the knowledge of "A Brief History of the World"
How to use - Tutorial on Building MyanmarGPT-Chat on local machine
Fine-tuned on MyanmarGPT-Chat, question answering model for the Burmese language. Faster, lightweight and multiplatform available model.
Install hugging face transformer
pip install transformers# Using Pipeline
from transformers import pipeline
pipe = pipeline("text-generation", model="jojo-ai-mst/MyanmarGPT")
outputs = pipe("အီတလီ",do_sample=False)
print(outputs)# Using AutoTokenizer and CausalLM
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("jojo-ai-mst/MyanmarGPT")
model = AutoModelForCausalLM.from_pretrained("jojo-ai-mst/MyanmarGPT")
input_ids = tokenizer.encode("ချစ်သား", return_tensors='pt')
output = model.generate(input_ids, max_length=50)
print(tokenizer.decode(output[0], skip_special_tokens=True))# Using Pipeline
from transformers import pipeline
pipe = pipeline("text-generation", model="jojo-ai-mst/MyanmarGPT-Big")
outputs = pipe("အီတလီ",do_sample=False)
print(outputs)# Using AutoTokenizer and CausalLM
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("jojo-ai-mst/MyanmarGPT-Big")
model = AutoModelForCausalLM.from_pretrained("jojo-ai-mst/MyanmarGPT-Big")
input_ids = tokenizer.encode("ချစ်သား", return_tensors='pt')
output = model.generate(input_ids, max_length=50)
print(tokenizer.decode(output[0], skip_special_tokens=True))[<iframe width="898" height="505" src="https://www.youtube.com/embed/RujWqJwmrLM" title="Chat GPT (AI) ကို မြန်မာလို သုံးစွဲနိုင်တော့မလား - DVB Youth Voice" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" allowfullscreen></iframe>](https://youtu.be/RujWqJwmrLM)