Preprint Review Version 1 This version is not peer-reviewed

XtremeLLMs: Towards Extremely Large Language Models

Version 1 : Received: 18 August 2024 / Approved: 20 August 2024 / Online: 21 August 2024 (03:49:01 CEST)

How to cite: Mienye, I. D.; Swart, T. G.; Obaido, G. XtremeLLMs: Towards Extremely Large Language Models. Preprints 2024, 2024081483. https://doi.org/10.20944/preprints202408.1483.v1 Mienye, I. D.; Swart, T. G.; Obaido, G. XtremeLLMs: Towards Extremely Large Language Models. Preprints 2024, 2024081483. https://doi.org/10.20944/preprints202408.1483.v1

Abstract

The continuous expansion of Large Language Models (LLMs) has significantly transformed the fields of artificial intelligence (AI) and natural language processing (NLP). This paper reviews the rapidly evolving domain of language models and introduces the concept of Extremely Large Language Models (XtremeLLMs), a new category defined for models exceeding one trillion parameters. These models are monumental in scale and engineered to enhance performance across a diverse range of language tasks. This study aims to establish a comprehensive framework that explores the significant opportunities and complex challenges presented by such extensive scaling and emphasises the implications for future advancements in the field.

Keywords

Artificial intelligence; deep learning; LLM; machine learning; XtremeLLM

Subject

Computer Science and Mathematics, Artificial Intelligence and Machine Learning

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.