Page 4 - LARGE LANUGAGE MODEL AND SMALL LANUGAGE MODEL
P. 4
4. Transformer-based models
These models use a neural network architecture that can understand
long-range dependencies in text data, making them useful for a
wide range of language tasks, including generating text, translating
languages, and answering questions.
5. MT-NLG (Megatron-Turing Natural Language Generation)
This advanced language model is based on transformer architecture
and can perform a wide range of natural language tasks, including
natural language inferences and reading comprehension.
6. XLNet
This pre-trained LLM was developed by researchers from Carnegie
Mellon University and Google and has been shown to outperform
BERT on a wide range of language tasks, including text classification
and question-answering.
Other types of LLMs include Cohere and Orca.