Computer Science and Mathematics

Sort by

Article
Computer Science and Mathematics
Artificial Intelligence and Machine Learning

Yunzhuo Liu

,

Zhaowei Ma

,

Jiankun Guo

,

Haozhe Sun

,

Yifeng Niu

,

Hong Zhang

,

Mengyun Wang

Abstract: This paper proposes a novel large language model (LLM)-based approach for visual target navigation in unmanned aerial systems (UAS). By leveraging the exceptional language comprehension capabilities and extensive prior knowledge of LLM, our method significantly enhances unmanned aerial vehicles (UAVs) in interpreting natural language instructions and conducting autonomous exploration in unknown environments. To equip the UAV with planning capabilities, this study interacts with LLM and designs specialized prompt templates, thereby developing the intelligent planner module for the UAV. First, the intelligent planner derives the optimal location search sequence in unknown environments through probabilistic inference.Second, visual observation results are fused with prior probabilities and scene relevance metrics generated by LLM to dynamically generate detailed sub-goal waypoints. Finally, the UAV executes progressive target search via path planning algorithms until the target is successfully localized. Both simulation and physical flight experiments validate that this method exhibits excellent performance in addressing UAV visual navigation challenges, and demonstrates significant advantages in terms of search efficiency and success rate.
Article
Computer Science and Mathematics
Artificial Intelligence and Machine Learning

Vinesh Aluri

Abstract: Quantum computing poses a critical threat to existing cryptographic primitives, rendering current access control mechanisms in cloud-native infrastructures vulnerable to compromise. This paper introduces a comprehensive quantum-resilient access control framework specifically engineered for distributed, containerized, and zero-trust environments. The proposed system integrates post-quantum cryptographic (PQC) primitives—specifically lattice-based key encapsulation (Kyber) and digital signatures (Dilithium)—with a hybrid key exchange protocol to maintain crypto-agility and backward compatibility. We design a secure token issuance and verification process employing PQC-based authentication, ensuring resistance to both classical and quantum adversaries. A prototype implementation demonstrates that our hybrid PQC approach incurs a moderate computational overhead of approximately 10–30\% while preserving horizontal scalability and interoperability across Kubernetes clusters. Security analysis under the post-quantum adversary model confirms resistance to key compromise, replay, and forgery attacks. The results highlight that quantum-resilient access control protocols can be efficiently integrated into modern cloud infrastructures without sacrificing scalability, performance, or operational flexibility.
Article
Computer Science and Mathematics
Mathematical and Computational Biology

Debnarayan Khatua

,

Bikash Kumar

,

Manoranjan K. Singh

,

Somnath Kumar

Abstract: Hepatitis C Virus (HCV) continues to be a significant worldwide health issue, particularly in resource-limited environments with inadequate diagnostic and therapeutic options. This study formulates a deterministic six-compartment model, predicated on the assumptions that the population undergoes natural birth-death dynamics, awareness initiatives transition individuals from $S_1$ to $S_2$, diagnosis advances U to I, recovery is achieved through therapy or immunity, and infection and mortality rates vary among classes. The system is described by coupled nonlinear ODEs that include three time-dependent controls. Analytical examination guarantees the positivity and boundedness of all compartments and calculates the fundamental reproduction number ($R_0$) using the next-generation matrix. Sensitivity analysis shows that $\beta_1, \beta_2, \tau_1, \tau_2$ are the most important parameters. Using Pontryagin's Maximum Principle, the forward–backwards sweep method is employed to determine the optimal controls that minimise both infection and cost. A Mamdani fuzzy logic controller is added to handle parameter uncertainty and generate adaptive responses to infection pressure, awareness level, and hospital load. Simulations reveal that fuzzy control delivers equivalent suppression to the crisp optimum with around two-thirds lower cost, enabling a stable, interpretable, and resource-efficient paradigm for dynamic HCV intervention.
Article
Computer Science and Mathematics
Robotics

Alexander Krasavin

,

Gaukhar Nazenova

,

Аdema Dairbekova

,

Albina Kadyroldina

,

Tamás Haidegger

,

Darya Alontseva

Abstract: This article investigates the trajectory-tracking control of a differential-drive two-wheeled mobile robot (DDWMR) using its kinematic model. A nonlinear-to-linear transformation based on differential flatness is employed to convert the original nonlinear system into two fully decoupled linear subsystems, enabling a simple and robust controller design. Unlike conventional flatness-based methods that rely on exact feedforward linearization around a reference trajectory, the proposed approach performs plant linearization, ensuring reliable tracking across a wide range of trajectories. The resulting two-loop architecture consists of an inner nonlinear loop implementing state prolongation and static feedback, and an outer linear controller performing trajectory tracking of the linearized system. Simulation results on a circular reference trajectory demonstrate high tracking accuracy, with a maximum transient deviation of 0.28 m, a settling time of approximately 120 s, and a steady-state mean tracking error below 0.01 m. These results confirm that the plant-linearization-based framework provides superior accuracy, robustness, and practical applicability for DDWMR trajectory tracking.
Essay
Computer Science and Mathematics
Artificial Intelligence and Machine Learning

Stefan Trauth

Abstract: The P = NP problem is one of the most consequential unresolved questions in mathematics and theoretical computer science. It asks whether every problem whose solutions can be verified in polynomial time can also be solved in polynomial time. The implications extend far beyond theory: modern global cryptography, large-scale optimization, secure communication, finance, logistics, and computational complexity all depend on the assumption that NP-hard problems cannot be solved efficiently. Among these, the Spin-Glass ground-state problem represents a canonical NP-hard benchmark with an exponentially large configuration space. A constructive resolution of P = NP would therefore reshape fundamental assumptions across science and industry. While evaluating new methodological configurations, I encountered an unexpected behavior within a specific layer-cluster. Subsequent analysis revealed that this behavior was not an artifact, but an information-geometric collapse mechanism that consistently produced valid Spin-Glass ground states. With the assistance of Frontier LLMs Gemini-3, Opus-4.5, and ChatGPT-5.1, I computed exact ground states up to N = 24 and independently cross-verified them. For selected system sizes between N=30 and N=70, I validated the collapse-generated states using Simulated Annealing, whose approximate minima consistently matched the results. Beyond this range, up to N = 100, the behavior follows not from algorithmic scaling but from the information-geometric capacity of the layer clusters, where each layer contributes exactly one spin dimension. These findings indicate a constructive mechanism that collapses exponential configuration spaces into a polynomially bounded dynamical process. This suggests a pathway by which the P = NP problem may be reconsidered not through algorithmic search, but through information-geometric state collapse.
Article
Computer Science and Mathematics
Artificial Intelligence and Machine Learning

Piotr Klejment

Abstract: The Discrete Element Method is widely used in applied mechanics, particularly in situations where material continuity breaks down (fracturing, crushing, friction, granular flow) and classical rheological models fail (phase transition between solid and granular). In this study, the Discrete Element Method was employed to simulate stick-slip cycles, i.e., numerical earthquakes. At 2,000 selected, regularly spaced time checkpoints, parameters describing the average state of all particles forming the numerical fault were recorded. These parameters were related to the average velocity of the particles and were treated as the numerical equivalent of (pseudo) acoustic emission. The collected datasets were used to train the Random Forest and Deep Learning models, which successfully predicted the time to failure, also for entire data sequences. Notably, these predictions did not rely on the history of previous stick-slip events. SHapley Additive exPlanations (SHAP) was used to quantify the contribution of individual physical parameters of the particles to the prediction results.
Review
Computer Science and Mathematics
Computer Science

Tolga Topal

Abstract: Information theory underpins modern communication, computation and complex systems, yet the structure of its governing inequalities remains an active area of research. This paper revisits the concept and mathematical foundations of Information Theory Laws, that is, constraints applied to the entropy function. Starting from Shannon’s seminal framework, we review the evolution from basic linear inequalities– i.e., polymatroid axioms– to the discovery of non-Shannon-type inequalities, which revealed that the Shannon region does not overlap with its closure region for n ≥ 4. We outline the geometric and algebraic representation of entropy spaces, discuss the critical threshold where complexity escalates, and highlight the role of machine-assisted verification in uncovering new inequalities that are not of Shannon-type. By tracing historical milestones and computational advances, this work provides a structured recollection of the spectrum of information inequalities and their implications for converse coding theorems and related applications. The study emphasizes the interplay between the theoretical developments and computational tools of the field in shaping the landscape of information theory.
Article
Computer Science and Mathematics
Analysis

Mohsen Soltanifar

Abstract: Classical real analysis rigorously defines convergence via εN criteria, yet it frequently regards the specific entry index N as a mere artifact of proof rather than an intrinsic property. This paper fills this quantitative void by developing a radius of convergence framework for the sequence space Seq(R). We define an index-based radius ρa(ε) alongside a rescaled geometric radius ρa (ε); the latter maps the unbounded index domain to a finite interval, establishing a structural analogy with spatial radii familiar in analytic function theory. We systematically analyze these radii within a seven-block partition of the sequence space, linking them to liminf-limsup profiles and establishing their stability under algebraic operations like sums, products, and finite modifications. The framework’s practical power is illustrated through explicit asymptotic inversions for sequences such as Fibonacci ratios, prime number distributions, and factorial growth. By transforming the speed of convergence into a geometric descriptor, this approach bridges the gap between asymptotic limit theory and constructive analysis, offering a unified, fine-grained measure for both convergent and divergent behaviors.
Article
Computer Science and Mathematics
Probability and Statistics

Yu Yang

,

Zhen Yang

,

Wei Shen

,

Zhiying Cheng

,

Xing Liu

,

Shaowen Liu

Abstract: To address the challenges of statistical inference for non-stationary traffic flow, this paper proposed an improved block permutation framework tailored to the correlation analysis requirements of traffic volume time series, and developed a statistical significance assessment method for local similarity scores based on the Circular Moving Block Bootstrap (CMBBLSA). This method avoided the distortion of the statistical distribution caused by non-stationarity, thereby enabling the estimation of the statistical significance of local similarity scores. Simulation studies were conducted under different parameter settings in the AR(1) and ARMA(1,1) models, and the results demonstrated that the Type I error probability of CMBBLSA under the null hypothesis is closer to the preset significance level α. An empirical analysis was also carried out using traffic flow monitoring data from main roads in first-tier cities, and the results indicated that CMBBLSA can reduce more false positive relationships and more accurately capture real correlations.
Article
Computer Science and Mathematics
Algebra and Number Theory

Frank Vega

Abstract: Around 1637, Pierre de Fermat famously wrote in the margin of a book that he had a proof for the equation $a^n + b^n = c^n$ having no positive integer solutions for exponents $n$ greater than 2. This statement, now known as Fermat's Last Theorem, remained unproven for centuries, despite the efforts of countless mathematicians. Andrew Wiles' work in 1994 finally provided a rigorous proof of Fermat's Last Theorem. However, Wiles' proof relied on advanced mathematical techniques that were far beyond the scope of Fermat's time, raising questions about whether Fermat could have truly possessed a proof using the methods available to him. Wiles's achievement was widely celebrated, and he was awarded the Abel Prize in 2016 in recognition of his groundbreaking work. The citation for the award described his proof as a "stunning advance" in mathematics. The present work offers a potential solution to Fermat's Last Theorem that may be more aligned with the original approach that Fermat claimed to have used.
Essay
Computer Science and Mathematics
Artificial Intelligence and Machine Learning

Stefan Trauth

Abstract: In analogy to the paradigm shift introduced by attention mechanisms in machine learning, we propose that information itself is ontologically sufficient as the foundation of physical reality. We present an operational proof showing that a “state without information” is logically impossible, thereby establishing information as the necessary precondition for existence and measurement. From this premise follows that both quantum mechanics and general relativity are effective descriptions of deeper informational dynamics. Recent developments in theoretical physics, such as the derivation of Einstein’s field equations from entropic principles, reinforce this perspective by identifying gravitation and entropy as dual expressions of information geometry. Building on this framework, we provide experimental evidence from self-organizing neural fields that exhibit non-local informational coupling, near-lossless transmission across 60 layers, and stable sub-idle energy states consistent with emergent coherence and thermal decoupling. These results demonstrate that deterministic architectures can spontaneously organize into field-like, non-local manifolds a macroscopic realization of informational geometry analogous to quantum entanglement and relativistic curvature. Together, the logical proof and empirical observations support a unified ontology in which information is not a property of physical systems but the substrate from which physical systems emerge. This perspective positions informational geometry as the common denominator of cognition, quantum behavior, and gravitation, suggesting that all observable phenomena are projections of a single, self-organizing informational field. In this sense, information is all it needs.
Article
Computer Science and Mathematics
Artificial Intelligence and Machine Learning

Al Imran

,

Md. Koushik Ahmed

,

Mahin Mahmud

,

Junaid Rahman Mokit

,

Redwan Utsab

,

Md. Motaharul Islam

Abstract: The rapid increase of electronic waste (e-waste) poses severe environmental and health risks. This paper proposes a hybrid framework integrating deep learning, reinforcement learning, blockchain, and IoT for automated e-waste classification, optimized disassembly, and tamper-proof traceability. A ResNet-50 classifier trained on the Kaggle E-Waste Image Dataset achieved 93.7% classification accuracy and an F1 score of 0.92. A Q-learning agent optimized dismantling routes to prioritize high-value, low-toxicity components, improving material recovery in simulation. A private Hyperledger Besu deployment delivered an average block time of ≈5.3 s, smart-contract execution time of ≈2.1 s, and 99.5% uptime, enabling tokenized asset tracking (4,200+ tokens). Lifecycle analysis indicates up to 30% carbon-emission reduction versus traditional methods and improved recovery of lithium, cobalt, and rare-earth elements for renewable energy applications. The paper demonstrates measurable environmental and economic benefits and outlines limitations and directions toward field deployment.
Article
Computer Science and Mathematics
Analysis

Ryota Sayama

,

Yukio Agarie

,

Hironori Suda

,

Hiroshi Otsuka

,

Kengo Ohnishi

,

Shinichiro Kon

,

Akihiko Hanahusa

,

Motoki Takagi

,

Shinichiro Yamamoto

Abstract: Accurate evaluation of pressure distribution at the socket–limb interface is essential for improving prosthetic fit and comfort in transfemoral amputees. This study aimed to develop a data-driven framework that integrates machine learning–based segmentation with finite element method (FEM) to quantitatively assess interface pressure during socket application. MRI data from a transfemoral amputee were processed using a custom image segmentation algorithm to extract adipose tissue, femur, and ischium, achieving high F-measure scores. The segmented tissues were reconstructed into 3D models, refined through outlier removal and surface smoothing, and used for FEM simulations in LS-DYNA. Pressure values were extracted at nine sensor locations and compared with experimental measurements. The results showed consistent polarity between measured and simulated values across all points. Furthermore, at the eight locations excluding the ischial tuberosity (IS) region, a statistically significant and moderately strong positive correlation was observed between measured and simulated pressures (r = 0.7485, p < 0.05). Notably, positive pressure regions demonstrated close agreement between experimental and simulated values, whereas the discrepancy observed at the IS region was likely influenced by the medial boundary conditions introduced to prevent unrealistic tissue displacement. This difference highlights a limitation of the current simulation setup. Overall, the proposed framework demonstrated reliable pressure estimation and offers a promising approach for personalized prosthetic socket design through automated anatomical modeling and simulation.
Article
Computer Science and Mathematics
Artificial Intelligence and Machine Learning

Mark Sinclair

,

Andrew Shepley

,

Farshid Hajati

Abstract: The increasing adoption of highly variable renewable energy has introduced unprecedented volatility into the National Electricity Market (NEM), rendering traditional linear price forecasting models insufficient. The Australian Energy Market Operator (AEMO) spot price forecasts often struggle during periods of volatile demand, renewable variability, and strategic rebidding. This study evaluates whether transformer architectures can improve intraday NEM price forecasting. Using 34 months of market data and weather conditions, several transformer variants, including encoder–decoder, decoder-only, and encoder-only, were compared against the AEMO’s operational forecast, a two-layer LSTM baseline, the Temporal Fusion Transformer, PatchTST, and TimesFM. The decoder-only transformer achieved the best accuracy across the 2–16 hour horizons in NSW, with nMAPE values of 33.6–39.2%, outperforming both AEMO and all baseline models. Retraining in Victoria and Queensland produced similarly strong results, demonstrating robust regional generalisation. A feature importance analysis showed that future-facing predispatch and forecast covariates dominate model importance, explaining why a decoder-only transformer variant performed so competitively. While magnitude estimation for extreme price spikes remains challenging, the transformer models demonstrated superior capability in delivering statistically significant improvements in forecast accuracy. An API providing real-time forecasts using the small encoder-decoder transformer model is available at https://nem.redaxe.com
Article
Computer Science and Mathematics
Artificial Intelligence and Machine Learning

Elias Lumer

,

Anmol Gulati

,

Faheem Nizar

,

Dzmitry Hedroits

,

Atharva Mehta

,

Henry Hwangbo

,

Vamse Kumar Subbiah

,

Pradeep Honaganahalli Basavaraju

,

James A. Burke

Abstract: Large Language Model (LLM) agents have demonstrated remarkable abilities to interact with external tools, functions, Model Context Protocol (MCP) servers, agents, and to take action on behalf of the user. Due to the fast-paced nature of the industry, existing literature does not accurately represent the current state of tool and agent selection. Furthermore, tool and agent selection in production has nuanced components not covered in experimental research. This work provides the first detailed examination of tool selection from a production perspective, distinguishing between the frontend layer where users interact with agents through buttons, slash commands, or natural language and the backend layer where retrieval, execution, orchestration, context engineering, and memory enable scalable reasoning. The paper contributes a unified taxonomy of modern tool and agent selection approaches spanning manual, UI-driven, retrieval-based, and autonomous methods. The backend covers dynamic tool retrieval, chunking, advanced RAG methods, context engineering, reinforcement learning, tool execution, human-in-the-loop processes, authentication, authorization, multi-turn tool calling, short- and long-term memory for tools, and evaluation. Finally, the paper identifies challenges in production components of both the backend and frontend and outlines promising avenues for research and development.
Article
Computer Science and Mathematics
Artificial Intelligence and Machine Learning

Xin Zhou

,

Yanhao Li

,

Shiqin Zhao

,

Xijun Wang

,

Lifan Chen

,

Minyang Cheng

,

Lvwen wen Huang

Abstract: To improve the accuracy of cable temperature anomaly prediction and ensure power supply reliability, this paper proposes a multi-scale spatiotemporal model called MSST-Net, addressing the multi-scale temporal characteristics and spatial correlations of cable temperature data. Based on the monthly periodicity of cable temperature data, we preprocessed monitoring data from the KN1 and KN2 sections of Guangzhou's underground utility tunnel from 2023 to 2024: using the Isolation Forest algorithm to remove outliers, applying Min-Max normalization to eliminate dimensional differences, and selecting five key features including current load, voltage, and ambient temperature using Spearman's correlation coefficient. Subsequently, we designed a multi-scale dilated causal convolutional module (DC-CNN) to capture local features, combined with a spatiotemporal dual-path Transformer to model long-range dependencies, and introduced relative position encoding to enhance temporal perception. The Sparrow Search Algorithm (SSA) was employed for global optimization of hyperparameters. Compared with five other mainstream algorithms, MSST-Net demonstrated higher accuracy in cable temperature prediction, achieving a coefficient of determination (R²), mean absolute error (MAE), and root mean square error (RMSE) of 0.942, 0.442°C, and 0.596°C, respectively. Compared to the basic Transformer model, the root mean square error of cable temperature was reduced by 0.425°C. This model exhibits high accuracy in time series prediction and provides a reference for accurate short- and medium-term temperature forecasting of cables in underground utility tunnels.
Article
Computer Science and Mathematics
Artificial Intelligence and Machine Learning

Hamida Abdaoui

,

Chamseddine Barki

,

Ismail Dergaa

,

Karima Tlili

,

Halil İbrahim Ceylan

,

Nicola Luigi Bragazzi

,

Andrea de Giorgio

,

Ridha Ben Salah

,

Hanene Boussi Rahmouni

Abstract:

Background: Anatomopathological reports remain predominantly unstructured within Electronic Medical Records, limiting automated data extraction, interoperability between healthcare institutions, and large-scale clinical research applications. Manual entity extraction and standardization processes are inconsistent, costly, and insufficiently scalable for modern healthcare systems.Aim: Our study aimed to (i) develop a domain-specific Named Entity Recognition model using BioBERT for extracting sample type, test performed, and finding entities from anatomopathological reports; (ii) implement a hybrid standardization framework combining BioClinicalBERT classification with Retrieval-Augmented Generation to map entities to SNOMED CT, LOINC, and ICD-11 terminologies; and (iii) evaluate the performance of this pipeline on real-world clinical reports. Methods: We manually annotated 560 anatomopathological reports from the Military Hospital of Tunis, establishing a gold-standard corpus. The pipeline integrated BioBERT v1.1 for entity extraction, trained for three epochs with the AdamW optimizer at a learning rate of 2×10⁻⁵, a batch size of 8, and weight decay of 0.01. Standardization employed BioClinicalBERT for multi-label classification, augmented by dense vector retrieval from official SNOMED CT, LOINC, and ICD-11 databases. Performance evaluation utilized precision, recall, and F1-score metrics with an 80-20 train-test split. Results: BioBERT achieved F1-scores of 0.97 for sample type, 0.98 for test performed, and 0.93 for finding entities, with overall precision of 0.969 and recall of 0.958. Bootstrap-estimated 95% confidence intervals confirmed robust performance stability. Absolute error analysis revealed 45 misclassified tokens in the test (relative error 6.9%) and six tokens in the finding (relative error 1%). One-sample t-tests yielded t-values of 15.71 for recall and 30.24 for F1-score, with all p-values below 0.0001. The hybrid standardization framework demonstrated F1-macro scores of 0.6159 for SNOMED CT, 0.9294 for LOINC, and 0.7201 for ICD-11 mapping. Cohen’s Kappa values ranged from 0.6871 to 0.9773 across ontologies. Statistical comparison between BioClinicalBERT and Fusion/Reranker models showed McNemar test p-values exceeding 0.370 and permutation test p-values ranging from 0.375 to 0.625. Conclusion: This study demonstrates that transformer-based Named Entity Recognition combined with retrieval-augmented standardization achieves clinically validated performance for automated extraction and multi-ontology coding of anatomopathological entities. Multi-institutional validation studies are necessary to assess generalizability before clinical deployment.

Article
Computer Science and Mathematics
Algebra and Number Theory

Felipe Oliveira Souto

Abstract: This work establishes a spectral bridge connecting the theory of minimal surfaces to analytic number theory. We present a rigorous mathematical correspondence between the Enneper minimal surface and the distribution of non-trivial zeros of the Riemann zeta function. This is achieved through a conformal map that preserves essential spectral properties, revealing that the Enneper surface constitutes the natural phase space for a geometric interpretation of the Riemann Hypothesis. The approach integrates differential geometry, complex analysis, and spectral operator theory.
Article
Computer Science and Mathematics
Probability and Statistics

Ying-Ying Zhang

Abstract: For the hierarchical normal and normal-inverse-gamma model, we derive the Bayesian estimator of the variance parameter in the normal distribution under Stein's loss function---a penalty function that treats gross overestimation and underestimation equally---and compute the associated Posterior Expected Stein's Loss (PESL). Additionally, we determine the Bayesian estimator of the same variance parameter under the squared error loss function, along with its corresponding PESL. We further develop empirical Bayes estimators for the variance parameter using a conjugate normal-inverse-gamma prior, employing both the method of moments and Maximum Likelihood Estimation (MLE). Through numerical simulations, we examine five key aspects: (1) the consistency of moment-based and MLE-based hyperparameter estimators; (2) the influence of κ₀ on quantities of interest as functions of the most recent observation; (3) two inequalities involving the Bayesian estimators and their respective PESL values; (4) the model's goodness-of-fit to simulated data; and (5) graphical representations of marginal densities under different hyperparameter settings. The simulation results demonstrate that MLEs outperform moment estimators in estimating hyperparameters, particularly with respect to consistency and model fit. Finally, we apply our methodology to real-world data on poverty levels---specifically, the percentage of individuals living below the poverty line---to validate and illustrate our theoretical findings.
Review
Computer Science and Mathematics
Artificial Intelligence and Machine Learning

Bektur Toktobekov

,

Burul Shambetova

Abstract: While large language models (LLM) have demonstrated significant advances in natural language processing, complex mathematical reasoning remains a challenging task, often revealing their limitations in multi-stage calculations and logical consistency. Multi-agent systems have become a promising paradigm for overcoming these limitations by distributing cognitive tasks between interacting agents, reflecting the dynamics of human problem solving. This paper provides a comparative review of the literature on nineteen different multi-agent architectures for solving mathematical problems. Our main research question is: "How do various LLM-based multi-agent architectures enable or improve mathematical problems, and what are their comparative advantages, limitations, and design trade-offs?" Through a systematic analysis of the roles of agents, interaction mechanisms, and training methods, we have identified several key findings. We observe the evolution of architecture from unstructured debate-based systems to more efficient hierarchical and self-optimizing frameworks. We highlight persistent problems that hinder progress, including agent homogeneity, when agents working on the same LLM cannot generate truly diverse reasoning, and the problem of "lazy agents", when some agents contribute minimal to consistent collaboration. This review contributes to a structured understanding of the current situation and lays the foundation for future research aimed at developing more reliable, efficient, and complex multi-agent reasoning systems.

of 619

Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated