Review
Version 1
This version is not peer-reviewed
XtremeLLMs: Towards Extremely Large Language Models
Version 1
: Received: 18 August 2024 / Approved: 20 August 2024 / Online: 21 August 2024 (03:49:01 CEST)
How to cite: Mienye, I. D.; Swart, T. G.; Obaido, G. XtremeLLMs: Towards Extremely Large Language Models. Preprints 2024, 2024081483. https://doi.org/10.20944/preprints202408.1483.v1 Mienye, I. D.; Swart, T. G.; Obaido, G. XtremeLLMs: Towards Extremely Large Language Models. Preprints 2024, 2024081483. https://doi.org/10.20944/preprints202408.1483.v1
Abstract
The continuous expansion of Large Language Models (LLMs) has significantly transformed the fields of artificial intelligence (AI) and natural language processing (NLP). This paper reviews the rapidly evolving domain of language models and introduces the concept of Extremely Large Language Models (XtremeLLMs), a new category defined for models exceeding one trillion parameters. These models are monumental in scale and engineered to enhance performance across a diverse range of language tasks. This study aims to establish a comprehensive framework that explores the significant opportunities and complex challenges presented by such extensive scaling and emphasises the implications for future advancements in the field.
Keywords
Artificial intelligence; deep learning; LLM; machine learning; XtremeLLM
Subject
Computer Science and Mathematics, Artificial Intelligence and Machine Learning
Copyright: This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Comments (0)
We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.
Leave a public commentSend a private comment to the author(s)
* All users must log in before leaving a comment