The explosive growth of language models and their applications have led ...
Pretrained multilingual large language models have typically used heuris...
We study the design decisions of publicly available instruction tuning
m...
Large language models (LLMs) have demonstrated impressive capabilities i...
Scaling language models improves performance but comes with significant
...
We evaluate the reasoning abilities of large language models in multilin...
There have been a lot of interest in the scaling properties of Transform...
Large pretrained Transformer language models have been shown to exhibit
...
Large language models have been shown to achieve remarkable performance
...
Recent neural network-based language models have benefited greatly from
...
Recent developments in machine translation and multilingual text generat...
There remain many open questions pertaining to the scaling behaviour of
...
State-of-the-art models in natural language processing rely on separate ...
Transformers are state of the art models in NLP that map a given input
s...
The research community has proposed copious modifications to the Transfo...
In many applications of machine learning, certain categories of examples...
We re-evaluate the standard practice of sharing weights between input an...
State-of-the-art multilingual models depend on vocabularies that cover a...
The quality of machine translation systems has dramatically improved ove...
Transformer-based models have achieved stateof-the-art results in many t...