Preprint
Article

Putting GPT-4o to the Sword: A Comprehensive Evaluation of Language, Vision, Speech, and Multimodal Proficiency

Altmetrics

Downloads

258

Views

290

Comments

0

A peer-reviewed article of this preprint also exists.

Submitted:

22 June 2024

Posted:

24 June 2024

You are already at the latest version

Alerts
Abstract
As large language models (LLMs) continue to advance, evaluating their comprehensive capabilities becomes significant for their application in various fields. This research study comprehensively evaluates the language, vision, speech, and multimodal capabilities of GPT-4o. The study employs standardized exam questions, reasoning tasks, and translation assessments to assess the model's language capability. Additionally, GPT-4o's vision and speech capabilities are tested through image classification and object recognition tasks, as well as accent classification. The multimodal evaluation assesses the model's performance in integrating visual and linguistic data. Our findings reveal that GPT-4o demonstrates high accuracy and efficiency across multiple domains in language and reasoning capabilities, excelling in tasks that require few-shot learning. GPT-4o also provides notable improvements in multimodal tasks compared to its predecessors. However, the model shows variability and faces limitations in handling complex and ambiguous inputs, particularly in audio and vision capabilities. This paper highlights the need for more comprehensive benchmarks and robust evaluation frameworks, encompassing qualitative assessments involving human judgment as well as error analysis. Future work should focus on expanding datasets, investigating prompt-based assessment, and enhancing few-shot learning techniques to test the model's practical applicability and performance in real-world scenarios.
Keywords: 
Subject: Computer Science and Mathematics  -   Artificial Intelligence and Machine Learning

1. Introduction

In the past few years, the emergence of large language models has led to paradigm shifts across various disciplines and professions. The pursuit of building and implementing the most powerful and accurate models has captured both researchers and industry. In late 2023 and early 2024, competitors to OpenAI, including Google and Anthropic, introduced advanced large language models: Google’s Gemini and Anthropic’s Claude 3 (Gemini Team et al., 2024; Korinek, 2023). These models surpassed the capabilities of the original GPT-3, GPT-3.5, and GPT-4 models that powered ChatGPT. To stay competitive, OpenAI needed to develop an upgraded model with more parameters, enhanced capabilities, and improved speed. This led to the launch of GPT-4 Omni (GPT-4o) in May 2024.
GPT-4o introduces several major innovations that improve upon previous large language models. The model includes a massive number of parameters – estimated to be well over one trillion – which dwarfs GPT-3, at 175 billion parameters, and GPT-1, at an estimated 117 million parameters (Floridi & Chiriatti, 2020). The model is able to process and generate text, image, and audio content and does so at a speed that is much faster than competitor models. Importantly, the model also integrates improved handling of ambiguous and complex queries, where a misunderstanding could emerge between the user and the model, and enhances its ethical and safety protocols to mitigate the prevalence of harmful or incorrect outputs, as has been an issue with competitor models in recent months (Dillion et al., 2024; Ray, 2024). Though all these innovations appear to be a tremendous boon for the model, there are many areas where the efficacy of the model has not yet been formally evaluated.

1.1. Research Purpose

The purpose of this study is to comprehensively evaluate the capabilities of GPT-4 Omni (GPT-4o) across various domains, including language, vision, speech, and multimodal tasks. By systematically assessing GPT-4o's performance on a wide range of benchmarks and real-world tasks, we aim to understand its capabilities, strengths, and limitations. This evaluation will provide insights into the advancements made by GPT-4o compared to previous models, such as GPT-3 and GPT-4, and other contemporary models like Google’s Gemini and Anthropic’s Claude 3. These findings will contribute to ongoing investigations of the practical applications and future development of large language models.

1.2. Related Work

GPT-4o is the latest development in a string of innovations to generative pre-trained transformers in recent years. In order to situate the development of GPT-4o within the context of the greater developments occurring in artificial intelligence (AI), it may be helpful to view these technologies as a series of nested boxes, as in Figure 1. AI as a concept encompasses a wide range of developments, of which machine learning and deep learning are but one area (Ongsulee, 2017). Within deep learning, there are further divisions, with generative AI being only one (albeit major) area. The same is true for large language models, as one application of generative AI. We already know of other types of generative AI that are not language-based, such as image generators. The generative pre-trained transformer is but one large language model (LLM), developed by OpenAI. GPT-4o is the latest version of this model. As such, while GPT-4o is a very important innovation, it is but one element within the broad AI landscape that exists today.
As illustrated in Figure 1, GPT-4o belongs to the class of technologies known as large language models (LLMs). These models are notable for their ability to mimic human language usage so closely that it can be difficult for a human observer to distinguish between text generated by a human and that generated by a machine (Thirunavukarasu et al., 2023; Hayawi et al., 2024). This innovation marks a significant advancement towards passing the Turing test and underscores the practicality of AI in writing and research (Aher et al., 2023; Mannuru et al., 2023). However, it also introduces significant risks, including potential invasions of privacy and the generation of inaccurate, misleading, biased, or harmful information (Lund et al., 2023). Therefore, it is crucial to carefully evaluate these LLMs and scrutinize their outputs. Failure to do so could lead to the proliferation of misinformation and malicious content on the Internet (Hu et al., 2024).
Given the serious issues associated with some LLMs, it is essential to critically examine each new model for its limitations. Recent versions of GPT have shown significant improvements over their predecessors in various areas. For example, Koubaa (2023) found substantial improvements in GPT-4 compared to GPT-3.5 on tests such as the Graduate Record Examination (GRE), SAT, and Bar exam, with GPT-4's performance placing it in the top tenth percentile on most of these exams. Similarly, Coyne et al. (2023) reported improvements in grammatical error correction for GPT-4 compared to GPT-3.5. However, having more parameters in a model does not inherently guarantee better performance on all tasks. Overfitting can occur when a model is extensively trained on a large dataset but fails to generalize well to real-world data (Salman & Liu, 2019).
Evaluation of the GPT-4o model is currently very limited. Research has explored various aspects of the model, including potential threats (Shen et al., 2024; Ying et al., 2024), diagnostic ability (Oura et al., 2024; Zhang et al., n.d.), and multilingual capabilities (Wang et al., 2024). One study by Sonoda et al. (2024) found that GPT-4o underperforms compared to Claude 3 Opus in radiology diagnosis tasks. Other studies investigated the sentiment of the general public regarding GPT-4o (Singgalen, 2024). Many studies focus on ChatGPT, the chatbot powered by GPT models, rather than the models themselves. These studies provide some additional insights into the quality of the models, such as their performance in English language teaching tasks (Pang et al., 2024). However, a comprehensive evaluation of GPT-4o itself remains a gap in the literature.
GPT-4o lends itself to new forms of evaluation beyond the language and reasoning evaluation of past model versions due to its new capabilities in vision, speech, and cross-modal activities. GPT-4 with Vision (GPT-4V) was previously evaluated on vision tasks; however, it is clear from these studies that the model was not ready for the visual challenges to which it was exposed (Xu et al., 2024; Zhou et al., 2024). Meanwhile, the speech capacity of GPT-4o is a new innovation, one that has already been met with some criticism due to the choices for the voice of the model (Allyn, 2024). While cross-modal activities have been theorized in LLMs for some time, GPT-4V stands out as among the first models to actualize this potential, paving the way for its evaluation (Li et al., 2021).

2. Language Capacity of GPT-4o

Language capacity is foundational to developing intelligent models capable of understanding, generating, and interacting with human language. This capacity encompasses a range of skills that enable models to process and produce coherent and contextually appropriate responses in natural language. The objective of this section is to comprehensively assess the language performance of GPT-4o (omni) by testing it on exams, reasoning tasks, and translation activities. Each of these tasks is significant for evaluating different aspects of the model's language capabilities.

2.1. Performance on Exams

In this subsection, we evaluate GPT-4o’s performance on various standardized and board exam questions. This helps us gauge the model's ability to comprehend complex problems and generate coherent, relevant, and accurate responses. Standardized exams are designed to measure a range of cognitive abilities and knowledge across different subjects. This task measures the model’s proficiency in handling structured questions across various subjects. Our methods involve presenting GPT-4o with questions from a variety of standardized and board exams. The responses generated by GPT-4o are then analyzed based on the correctness of the answers provided.

2.1.1. Performance on USMLE

The United States Medical Licensing Examination (USMLE) Step 1 is a rigorous and comprehensive assessment designed to evaluate a candidate's understanding and ability to apply key concepts in medical science necessary for the practice of medicine (Federation of State Medical Boards & of Medical Examiners, 2024). Jointly developed by the Federation of State Medical Boards and the National Board of Medical Examiners, this examination serves as a milestone for medical students and professionals aiming to obtain their medical licensure in the United States. The USMLE Step 1 primarily focuses on testing the examinee's grasp of foundational medical knowledge and their ability to apply this knowledge to clinical scenarios. The sample test questions provided in the USMLE Step 1 Sample Items booklet encompass various disciplines, including anatomy, biochemistry, microbiology, pathology, pharmacology, physiology, and interdisciplinary areas such as genetics, immunology, and molecular biology​​. The dataset used for evaluating GPT-4o’s performance includes 119 sample test questions from the USMLE Step 1 booklet, updated as of January 2024 (https://www.usmle.org/sites/default/files/2021-10/Step_1_Sample_Items.pdf).
Out of the total 118 questions, GPT-4o correctly answered 98 questions. This corresponds to an accuracy of 83.1%. Table 1 provides a comparison of GPT-4o with its predecessor models, as reported by Gilson et al. (2023) and Brin et al. (2023). Compared to its predecessor, GPT-3.5, which achieved an accuracy of 51.67%, GPT-4o shows significant improvement. GPT-4o, despite being designed for faster and more efficient tasks, offers a notable enhancement in language comprehension and problem-solving capabilities. However, GPT-4o's performance is slightly lower than that of GPT-4, which achieved an accuracy of 90.00%. This decline can be attributed to the design focus of GPT-4o on efficiency and speed, while GPT-4 remains the model for more complex and demanding tasks.
The results indicate that GPT-4o can serve as a valuable tool in medical education, offering fast, interactive learning experiences that are crucial for students needing immediate feedback and guidance (Haleem et al., 2022). While GPT-4 excels in handling more intricate questions, its slower response time may limit its practicality for real-time learning scenarios. Meanwhile, GPT-4o's accuracy and efficiency make it suitable for dynamic educational environments.

2.1.2. Performance on CFA

The Chartered Financial Analyst (CFA) Level 1 exam is a globally recognized certification offered by the CFA Institute, aimed at financial and investment professionals (CFA Institute, n.d.). The exam covers a broad range of topics, including ethical and professional standards, quantitative methods, economics, corporate finance, equity investments, fixed income, derivatives, and portfolio management. The CFA Level 1 exam is known for its rigorous and comprehensive assessment of a candidate’s foundational knowledge and skills in finance and investment. It tests both theoretical understanding and the practical application of financial concepts and principles.
For this evaluation, we utilized the dataset from the 300Hours CFA Level 1 Mock Exam, which includes questions developed to mirror the style and difficulty of the actual exam (https://300hours.com/free-cfa-level-1-mock-exam/). GPT-4o correctly answered 76 out of the 89 questions, yielding an overall accuracy of 85.39%. Table 2 summarizes the performance in comparison to GPT-3.5 and GPT-4, as reported by Callanan et al. (2023). We compare the results obtained using zero-shot prompting since we did not provide the models with any hints or specific instructions during our prompting. The results indicate that GPT-4o noticeably outperforms both its predecessors. The increased accuracy of GPT-4o (despite being designed for faster and more efficient tasks) indicates that it can provide reliable and timely assistance for financial exam preparation.

2.1.3. Performance on SAT

The Scholastic Assessment Test (SAT) is a standardized test widely used for college admissions in the United States (College Board, n.d.). Developed and administered by the College Board, the SAT assesses a student's readiness for college and provides colleges with a common data point for comparing all applicants. The SAT covers areas including reading, writing and language, and mathematics, with an optional essay section. This test is designed to measure a range of skills necessary for academic success in college, including critical thinking, problem-solving, and analytical abilities.
The dataset used for evaluating GPT-4o’s performance consists of questions from the SAT Practice Test #1, which includes a variety of reading, writing, and math questions that reflect the format and content of the actual SAT exam (https://satsuite.collegeboard.org/sat/practice-preparation?excmpid=mtg796-st-1-bk). The practice test consisted of two modules, each containing a reading and writing exam as well as a math exam. The performance on each module are outlined in Table 3.
For comparison with previous GPT models, we refer to the comprehensive report by the Open AI team (OpenAI et al., 2024). In this context, we average the results of M1 and M2 for GPT-4o, as summarized in Table 4.
GPT-4o demonstrates the highest accuracy in the Reading & Writing section with 90.91%, surpassing all the older models. In the Math section, GPT-4o achieves a strong performance with 87.04%, slightly lower than GPT-4 but higher than the rest. Figure 2, Figure 3, Figure 4 and Figure 5 provide examples of GPT-4o correct and incorrect responses on each of the SAT categories.

2.1.4. Performance on MBE

The Multistate Bar Examination (MBE) is a standardized test that assesses the ability of prospective lawyers to apply fundamental legal principles and reasoning to analyze given fact patterns (National Conference of Bar Examiners, n.d.). Developed and administered by the National Conference of Bar Examiners (NCBE), the MBE is a critical component of the bar examination in most U.S. jurisdictions (Griggs, 2019). The MBE includes 200 multiple-choice questions that cover a wide range of legal topics, including constitutional law, contracts, evidence, real property, and torts. The test evaluates the examinee's capacity to think like a lawyer and apply legal knowledge in a practical, problem-solving context.
The dataset for evaluating GPT-4o's performance includes sample test questions from the MBE sample booklet, updated in 2023 (https://www.ncbex.org/sites/default/files/2023-05/MBE_Sample_Test_Questions_New_2023%20.pdf). These questions represent the types and formats of questions that examinees will encounter on the actual MBE, providing a comprehensive overview of the subjects tested and the skills required. In this test, GPT-4o correctly answered 15 out of 20 questions, leading to 75% accuracy. The comparison with previous models is presented in Table 5, based on the results reported by Katz et al. (2024).
The evaluation results indicate that GPT-4o performs comparably to GPT-4 on the MBE, with a minor difference in accuracy. However, compared to GPT-3.5, which achieved an accuracy of 45.10%, GPT-4o demonstrates a significant improvement. Therefore, law students and bar examinees can benefit from using GPT-4o as an interactive learning tool that provides immediate feedback and explanations, helping them to understand complex legal principles and improve their problem-solving skills.

2.2. Reasoning

Human intellect is remarkably characterized by reasoning, which is described as an activity of methodically and logically thinking about a subject (Huang & Chang, 2023). Reasoning enables humans to come to conclusions or make decisions by using previous experiences and data gathered, thus extending one’s knowledge of the world and releasing the possibility for innovation and development. In recent times, AI has made significant advancements in narrowing the gap between human and machine intellect through the use of Natural Language Processing (NLP) and LLMs, which have established remarkable reasoning abilities.
In this section, the authors assess the reasoning capacity of the most recent GPT-4o model by manual technical evaluation through a sequence of question-answering tasks. The model will answer a range of logical reasoning tasks in different types, including deductive, inductive, and abductive reasoning, as shown in Figure 6. Starting with a broad principle or assumption and applying it to produce predictions or draw conclusions, deductive reasoning takes a top-down method (Johnson-Laird, 2010). By contrast, inductive reasoning uses a bottom-up methodology to deduce broad principles or conclusions from observations or data (Hayes et al., 2010). In abductive reasoning, theories or explanations are developed from little, ambiguous, or incomplete data (Walton, 2014). With all these assessments on the model, this article stands to obtain an understanding of the reasoning capacity of the GPT-4o model in various settings.
In this subsection, we assess the performance of GPT-4o on five datasets that include all the aforementioned types of reasoning, as illustrated in Figure 7. To evaluate the deductive reasoning ability, two datasets were utilized namely EntailmentBank (Dalvi et al., 2021) and bAbI (task 15) (Weston et al., 2015). Similarly, to assess the capability of inductive reasoning, we employed two datasets CLUTRR (Sinha et al., 2019) and bAbI (task 16) (Weston et al., 2015). For abductive reasoning, we use the αNLI dataset (Bhagavatula et al., 2019). Adhering to the methods of López Espejel et al. (2023), our evaluation encompassed the same set of 30 randomly chosen samples from each evaluation dataset. The observations are selected from 10 samples from each of the training-easy, train-medium, and train-hard sets of the αNLI dataset. Concurrently, a total of 30 samples are drawn from the test set for the bAbI (task 15), bAbI (task 16), CLUTRR, and EntailmentBank datasets. In accordance with López Espejel et al. (2023), we utilized the identical set of proven and effective prompts that were implemented in their assessment to evaluate the capabilities of the model.
The evaluation results showcase the remarkable reasoning abilities of GPT-4o in all three domains, as indicated in Table 6. GPT-4o demonstrated exceptional performance in deductive reasoning by achieving nearly flawless scores on both bAbI (task 15) and EntailmentBank. It outperformed ChatGPT-3.5 and performed at the same level as ChatGPT-4 (López Espejel et al., 2023). GPT-4o achieved flawless results in the inductive reasoning tasks, scoring perfectly on bAbI (task 16) and achieving a score of 17 out of 30 on CLUTRR. It outperformed both ChatGPT-3.5 and ChatGPT-4. GPT-4o achieved a score of 27 out of 30 on αNLI, surpassing the performance of its previous versions in abductive reasoning. The results underscore GPT-4o's superior reasoning capabilities compared to its predecessors. The model's proficiency in deductive reasoning showcases its ability to derive valid conclusions from premises. Its success in inductive reasoning demonstrates the capacity to generalize from specific facts, while its performance in abductive reasoning highlights the ability to generate credible hypotheses with limited knowledge.
Even with the remarkable reasoning powers of GPT-4o, this assessment points to a few drawbacks that need more rigorous investigation. One case was when the model gave different answers to the same topic in various chat sessions while evaluating the bAbI (task 16) dataset for inductive reasoning. Additionally, the model sometimes requested the end-user to choose between different answers. This implies that in some situations, GPT-4o could have trouble with ambiguity, resulting in varying responses. Furthermore, the model sometimes gave different responses when the same subject was posed repeatedly in the same chat session. In determining accuracy, only the first response was considered to maintain synchrony, although this inconsistency raises questions about the model's efficiency and dependability in certain areas. The model's sensitivity to question-wording, the information presentation sequence, or the existence of unclear or contradictory information in the input may be the causes of these issues. To overcome these problems and enhance its capacity to manage ambiguity and resolve contradictions, future studies should concentrate on creating more reliable and consistent reasoning mechanisms and optimizing prompts for LLMs.
With the notable performance of GPT-4o, there might be more advancements for many AI applications with GPT-4o's improved reasoning skills. High-accuracy complex reasoning tasks performed by it can lead to advancements in information retrieval, decision support tools, and question-answering systems. Still, further study is required to determine how well the model performs on a larger variety of reasoning problems and how well it can manage more intricate and domain-specific reasoning situations.

2.3. Language Translation

Language translation has become an increasingly important task in our globalized world, facilitating communication and understanding across diverse linguistic backgrounds. With the advent of LLMs like GPT 3.5, GPT 4, Llama, Gemini and now GPT-4o, the potential for accurate and efficient machine translation has grown significantly (Khoshafah, 2023). These models, which were trained on massive volumes of multilingual data, can produce translations accurately while capturing the subtle meanings and complexities of different languages. Therefore, in this section, we aim to evaluate the translation proficiency of GPT-4o in six of the most widely spoken languages: Spanish, Arabic, Hindi, French, Portuguese, and Russian.
The choice of these six languages is not arbitrary; they represent a diverse set of linguistic structures and cultural contexts, making them ideal for a comprehensive evaluation of translation capabilities. Spanish is commonly used across Europe and the Americas and is characterized by its straightforward structure and rich vocabulary. Arabic, known for its intricate script and complex word forms, poses distinct challenges for translation technology. Hindi, widely spoken in India, mixes local and foreign words, requiring careful handling to achieve accurate translation. French, spoken in many parts of the world, helps test the model’s ability to handle grammatical rules and nuances. Portuguese, similar to Spanish but distinct in several key aspects, allows for an assessment of the model's precision in closely related languages. Lastly, Russian, with its Cyrillic script and case system, provides a test for the model's ability to manage non-Latin scripts and complex grammatical structures.
By focusing on these languages, this study aims to provide a robust and diverse evaluation of GPT-4o's translation performance. Given the widespread use and significant number of native speakers of these languages, improvements in translation accuracy can have a substantial impact on global communication and information dissemination. Hence in this section, we seek to verify GPT-4o's ability to translate across these six languages, providing insights into its potential for breaking down language barriers and facilitating communication among people from different linguistic backgrounds.

2.3.1. Data

The datasets for Spanish, Arabic, French, Portuguese, and Russian were sourced from the OPUS dataset, a well-known collection of texts used for training and evaluating machine translation models (Tiedemann, 2012), and the Hindi dataset was obtained from the IIT Bombay English-Hindi Parallel Corpus, created by the Center for Indian Language Technology (CFILT) at IIT Bombay (Kunchukuttan et al., 2018).
For this analysis, 500 data points were randomly sampled from each dataset. The selection of 500 data points is a good balance between feasibility and the need for sufficient data diversity. This sample size is large enough to encompass a wide variety of sentence structures, vocabulary, and translation challenges present in each language, ensuring that the evaluation is comprehensive and representative. Random sampling was employed to mitigate selection bias and to ensure that the sampled data points provide an unbiased representation of the overall dataset. By using random sampling, this approach captures the natural variability and complexity of language, which is essential for a robust assessment of the GPT-4o model's translation performance across different linguistic contexts.

2.3.2. Evaluation Method

To measure how similar two sentences are in terms of their meaning, an advanced NLP, specifically focusing on sentence embeddings generated by a model called BERT (Bidirectional Encoder Representations from Transformers) (Devlin et al., 2019) and a similarity measure called cosine similarity has been used. BERT is a powerful model that has greatly improved how well computers understand language. For our research, we use a pre-trained model from the sentence-transformers library called paraphrase-MiniLM-L6-v2. This model is specially tuned to understand the meanings and similarities between sentences. It works by turning each sentence into a vector, which is a list of numbers. These vectors or embeddings, encapsulate the semantic information of the sentences in a way that allows for meaningful comparison between the actual translations and the translations generated by GPT 4o.
To find out how similar two sentences are, we compare their vectors using cosine similarity. Cosine similarity measures the angle between two vectors or embeddings. If the vectors point in the same direction, the sentences are very similar. If they point in completely different directions, the sentences are very different. The values are between -1 and 1, where:
  • 1 indicates that the vectors are identical.
  • 0 indicates that the vectors are orthogonal (i.e., no similarity).
  • -1 indicates that the vectors are opposed.
By calculating the cosine similarity between the embeddings of two sentences, we can effectively measure their semantic similarity. The formula for cosine similarity is:
Cosine Similarity = 𝐴𝐵 / ∥𝐴∥ ∥𝐵
​where A and B are the embeddings of the two sentences.

2.3.3. Results

This study sought to evaluate the capabilities of GPT-4o in translating passages across six major languages: Spanish, Arabic, Hindi, French, Portuguese, and Russian. The results reveal a generally high level of translation accuracy, particularly in Spanish and Portuguese, which scored 88% and 86% respectively. However, there were notable variations among the languages. Arabic and French, with scores of 78% and 75%, respectively, presented more challenges for the model due to their complex linguistic structures and nuances. Hindi and Russian scored 82% and 80%, demonstrating the model’s competence but also highlighting areas for improvement. The results are summarized in Table 7.
The findings suggest that the line between human and machine translation is becoming increasingly narrow. GPT-4o's performance, though not specifically optimized for translation, approaches the quality of dedicated translation systems. This is particularly noteworthy given the diverse linguistic and structural characteristics of the evaluated languages. While the exact nature of the source translations in the datasets (whether human or machine-translated) is not confirmed, the high similarity scores indicate that GPT-4o is capable of producing translations with a quality that is comparable to the existing translations. However, several limitations must be considered. The random sampling of 500 data points from each dataset may not fully capture the linguistic diversity and complexity of each language. Different samples could yield varying results, suggesting that a larger and more representative dataset might provide a more accurate assessment. Additionally, the reliance on BERT-based embeddings and cosine similarity may not fully encapsulate the nuances of translation quality, particularly in capturing cultural and contextual subtleties. Expanding the dataset size and including more language pairs could yield more comprehensive insights. This research serves as a proof-of-concept for larger-scale studies that could further investigate the capabilities of AI in translation. Future research should focus on incorporating more extensive data, diverse language combinations, and advanced fine-tuning techniques.

3. Vision Capacity of GPT-4o

Vision capacity is foundational to developing intelligent models capable of understanding, interpreting, and interacting with visual content. This capacity encompasses a range of skills that enable models to process and produce coherent and contextually appropriate responses to visual inputs. The objective is to comprehensively assess the vision performance of GPT-4o by testing it on various image-based tasks. Each of these tasks is significant for evaluating different aspects of the model's visual capabilities.
For each task, a dataset of approximately 100 representative images was curated. The model was provided with an image along with a text prompt specifying the desired output format. The prompts were designed to probe the model's ability to identify, classify, describe, and analyze visual content without additional context. For select tasks, we further investigated the model's few-shot learning capabilities by providing a small number of labeled examples before the query image.
Model outputs were compared against ground truth labels to compute standard performance metrics such as accuracy. Qualitative analysis was also conducted on a subset of responses to identify common failure modes and strengths. The results across different tasks provide insights into GPT-4o's current visual understanding capabilities, areas for improvement, and potential as a foundation model for vision tasks. Subsequent sections discuss the specifics of each task, dataset, and findings, offering a comprehensive evaluation of GPT-4o's visual reasoning skills.

3.1. Fruits Classification

Fruit image classification is crucial for applications in agriculture, supply chain management, and food industry automation. Accurate identification of fruit types can enhance inventory tracking, quality control, and efficient sorting processes (Cubero et al., 2011). The fruit images dataset (https://www.kaggle.com/datasets/afsananadia/fruits-images-dataset-object-detection) consists of approximately 400 images spanning 10 different fruit classes such as banana, jackfruit, and mango. Each fruit class has 40 labeled images, with the dataset split into 320 training images and 80 test images. The images were collected from various sources, such as Google Images and stock image websites, and were labeled by the dataset creators. For this evaluation, the model was provided with an image along with a prompt to identify the fruit class from the list of 10 classes in a specified format. Model predictions were compared against ground truth labels to assess performance.
The results indicate that GPT-4o performed exceptionally well on this task. The model achieved an average precision of 0.98, an average recall of 0.98, and an average F1-score of 0.98. These metrics suggest that GPT-4o is highly capable of accurately identifying and classifying different fruit images. Table 8 summarizes the performance for each class.
The model demonstrated strong performance in classifying the 10 different fruit classes, achieving high precision, recall, and F1-scores across most classes. Several classes, including Papaya, Apple, Litchi, Hog Plum, Grapes, and Guava, obtained perfect scores of 1.0 for precision, recall, and F1-score. The Banana and Mango classes had slightly lower but still impressive precision scores (0.91), with a perfect recall of 1.0. Figure 8 presents the confusion matrix and metric visualization for this dataset.

3.2. Driver Drowsiness Detection

Detecting driver drowsiness is critical for enhancing road safety, as timely identification of fatigue can prevent accidents and save lives. The drowsy detection dataset consists of images extracted from videos capturing drivers in three distinct states: natural, fatigued, and drowsy (Jebraeily et al., 2024). The dataset was curated by gathering relevant videos, converting them into image frames, and applying facial detection algorithms to isolate key facial regions like eyes, mouth, and cheeks, which are indicative of drowsiness. The extracted images were converted to grayscale, resized to 48x48 pixels, and accurately labeled based on the driver's state. The dataset comprises two classes: drowsy and natural, with a total of 100 labeled images sampled evenly from each class. For this evaluation, GPT-4o was provided with an image along with a prompt to classify it into one of the two classes in a specified JSON format. The model's predictions were compared against the ground truth labels to assess its performance in detecting driver drowsiness from facial features.
In this task, the model achieved an average precision of 0.80, an average recall of 0.80, and an average F1-score of 0.80. Table 9 summarizes the performance for each class.
The results indicate that GPT-4o, without fine-tuning, achieves an impressive precision, recall, and F1-score of 0.8. While lower than that of specialized deep learning models like VGG, ResNet, and CNN (Jebraeily et al., 2024), the performance is impressive given GPT-4o’s lack of training on this specific dataset. The notable performance despite no domain-specific training underscores its robustness and adaptability, implying that GPT-4o could be valuable in scenarios where rapid deployment and flexibility across different tasks are crucial. Figure 9 presents the confusion matrix and metric visualization for this task.

3.3. Crop Disease Classification

Accurate identification of crop diseases is essential for ensuring agricultural productivity and preventing significant crop losses. The crop disease classification dataset is a comprehensive collection of images aimed at evaluating GPT-4o's capabilities in identifying crop diseases (https://www.kaggle.com/datasets/sadikaljarif/crop-disease). The dataset encompasses 20 distinct classes of common crop diseases, including blight, cedar apple rust, crown gall, and clubroot. For this evaluation, 100 images were randomly sampled from the dataset, with each class represented by approximately five images. GPT-4o was provided with these images along with a prompt to classify the crop disease depicted in each image. The model's predictions were compared against the ground truth labels to assess its performance in accurately identifying and distinguishing various crop diseases based solely on visual information.
The model achieved an average precision of 0.77, an average recall of 0.71, and an average F1-score of 0.68 in this task. Table 10 summarizes the performance for each class.
Given the large number of classes (20), this highlights GPT-4o’s potential for accurate crop disease classification and adaptability, despite no prior training on this dataset. The limitations in specific classes like Botrytis, Brown rot, and Canker can be attributed to the need for specialized training in certain classes. The confusion matrix and metric visualization for this dataset are presented in Figure 10.

3.4. Glaucoma Detection

Early detection of glaucoma is critical for preventing vision loss and ensuring timely treatment. The glaucoma detection dataset used for this evaluation consisted of retinal fundus images from the ACRIMA database (https://www.kaggle.com/datasets/chetanpediredla/glaucoma-dataset). A subset of 100 images was sampled, evenly split between glaucomatous and normal cases. These images were collected at FISABIO Oftalmología Médica in Valencia, Spain, and were annotated by experienced glaucoma experts. GPT-4o was tasked with classifying each image into either glaucoma or normal based solely on the visual information provided. The model's predictions were compared against the expert-annotated ground truth labels to assess its performance in detecting glaucoma from retinal fundus imagery.
As shown in Table 11, GPT-4o achieved an average precision of 0.65, an average recall of 0.62, and an average F1-score of 0.59. For the glaucoma class, the model demonstrated a precision of 0.58, a recall of 0.86, and an F1-score of 0.69. In contrast, the normal class had a higher precision of 0.73 but a significantly lower recall of 0.38, resulting in an F1-score of 0.50.
The confusion matrix in Figure 11 reveals that the model correctly identified 42 out of 49 glaucoma cases but struggled more with normal cases, correctly classifying only 19 out of 50. The plot shows the model's relatively balanced precision and recall for the glaucoma class but highlights a pronounced discrepancy for the normal class, with precision substantially higher than recall. GPT-4o is effective at identifying glaucomatous images but has difficulty in correctly classifying normal cases.
Few-shot learning allows models to make accurate predictions with only a small number of training examples. This approach is particularly beneficial in scenarios where data is scarce. Figure 12 illustrates the F1 scores for both classes across different numbers of shots, indicating how the model's performance evolves with the number of examples provided during training. The glaucoma class maintains a relatively high F1 score across all shot levels, showing slight improvement with additional examples. This consistency suggests that GPT-4o effectively learns to identify glaucomatous features even with a limited number of examples. In contrast, the normal class exhibits significant improvement in F1 score from zero shots to one shot but then plateaus. This indicates that while the initial provision of examples significantly enhances the model's ability to recognize normal cases, further increases in the number of examples yield diminishing returns.

3.5. Cancer, Tumor, and Aneurysm Detection

Accurate detection and classification of brain conditions such as cancer, tumors, and aneurysms are crucial for timely diagnosis and treatment. The computed tomography (CT) brain scan dataset contains CT images of the brain aimed at detecting and classifying various conditions such as cancer, tumors, and aneurysms (https://www.kaggle.com/datasets/trainingdatapro/computed-tomography-ct-of-the-brain). For this evaluation, a subset of 100 CT scan images was sampled from the dataset. GPT-4o was tasked with analyzing these images and classifying them into one of three categories: cancer, tumor, or aneurysm. The model's predictions were compared against the ground truth labels to assess its performance in identifying these medical conditions from CT brain imagery.
GPT-4o achieved an average precision of 0.21, an average recall of 0.32, and an average F1-score of 0.26. Table 12 summarizes the performance for each class.
The confusion matrix in Figure 13 reveals that the model completely failed to predict the 'cancer' class, potentially due to a lack of representative training data or inherent similarities with other classes. Additionally, it struggled to distinguish between 'aneurysm' and 'tumor' classes, with significant misclassifications in both directions, suggesting a need for further fine-tuning or incorporation of additional relevant features.

3.6. Image Captioning

The flickr8k captions dataset is a compact collection designed for image captioning tasks (https://www.kaggle.com/datasets/aladdinpersson/flickr8kimagescaptions). It consists of 8,000 images sourced from Flickr, each accompanied by multiple human-annotated captions describing the visual content. For this evaluation, a subset of 100 images from the dataset was used. GPT-4o was tasked with inferring the context and generating natural language descriptions for these 100 images. The model's generated captions were compared against the ground truth human captions using the BLEU score, a metric that measures the similarity between machine and human-generated texts. This dataset helps assess the ability to comprehend visual scenes and translate them into accurate and coherent textual descriptions. The BLEU scores obtained are summarized in Table 13.
With a BLEU-1 score of 0.193, the model demonstrates a moderate ability to capture the essence of the captions with a reasonable degree of similarity in individual words. However, as the n-gram length increases, the scores decline significantly (BLEU-2: 0.095, BLEU-3: 0.058, BLEU-4: 0.031), indicating that the model struggles with maintaining coherence and context in longer sequences. This highlights the challenges GPT-4o faces in generating more complex and accurate descriptions. The results show that GPT-4o has a foundational understanding of visual scenes, but there is room for improvement in generating detailed and contextually rich captions.

4. Speech Capacity of GPT-4o

Speech capacity evaluates the ability of intelligent models to understand, interpret, and interact with auditory content. This encompasses a range of skills that enable models to process and produce coherent and contextually appropriate responses to audio inputs. The objective is to assess the audio performance of GPT-4o by testing it on various audio-based tasks. Each of these tasks is significant for evaluating different aspects of the model's auditory capabilities.

4.1. Emotion Detection

Emotion detection is a critical aspect of understanding human communication, as the same speech can convey different meanings depending on the emotional tone in which it is expressed (Shahriar et al., 2023). Recognizing emotions in speech is essential for applications ranging from customer service to mental health monitoring. For this evaluation, we used the Arabic natural audio dataset (ANAD) from Kaggle, designed to detect discrete emotions in Arabic speech (https://www.kaggle.com/datasets/suso172/arabic-natural-audio-dataset/data). The ANAD consists of 1,384 audio recordings, each labeled with one of three emotions: happy, angry, or surprised. These recordings were sourced from live Arabic talk shows, where each video was labeled by 18 listeners to determine the perceived emotion. To evaluate the emotion detection capabilities of GPT-4o, we randomly sampled 100 audio files from the ANAD dataset. Each audio file was fed to the model along with a prompt to predict the emotion class. The model's predictions were then compared against the ground truth labels to assess its performance.
The results of the emotion detection task, as illustrated in Figure 14, reveal that GPT-4o demonstrates variable performance across different emotion classes. The confusion matrix shows that the model performs best for the "surprised" class, correctly predicting 21 instances, but it frequently misclassifies "happy" as "surprised" (19 times). The "angry" class has the lowest true positive rate with only two correct predictions, often being mistaken for "happy" or "surprised." The model has the highest recall for the "surprised" class, indicating it correctly identifies "surprised" emotions more frequently than others. The precision for "angry" is reasonably high, but the recall is very low, meaning that while it predicts "angry" correctly when it does so, it rarely predicts "angry" overall. The "happy" class has moderate precision and recall, suggesting a balanced but moderate performance in predicting this class.

4.2. Accent Detection

Accents play a crucial role in speech recognition, affecting the accuracy and efficiency of automatic speech recognition (ASR) systems. Understanding and detecting accents is essential for developing robust ASR systems that can handle diverse linguistic backgrounds (Graham & Roll, 2024). For this evaluation, we utilized the AccentDB dataset, a comprehensive collection of non-native English accents designed to assist neural speech recognition tasks (Ahamad et al., 2020).
The AccentDB dataset includes samples from various Indian-English accents and native English accents, providing a diverse range of phonetic and prosodic variations. It contains speech recordings from speakers with distinct linguistic backgrounds, such as Bangla, Malayalam, Odiya, and Telugu, alongside metropolitan Indian accents and native accents from American, Australian, British, and Welsh English. The dataset is structured to meet key requirements for ASR development, including a variety of speakers, uniformity of content, and well-labeled data for training and testing models. To assess the accent detection capabilities of GPT-4o, we randomly selected 100 audio files from the AccentDB dataset. Each file was presented to the model with a prompt to identify the speaker's accent. The predictions made by GPT-4o were then compared to the ground truth labels to evaluate their performance.
The confusion matrix in Figure 15 highlights significant misclassifications, particularly with the Malayalam accent, which is frequently misclassified as Telugu. This misclassification suggests that the acoustic features of Malayalam and Telugu might be similar enough to confuse the model, indicating a need for more distinctive feature extraction and training data augmentation. Bangla and Telugu also exhibit substantial misclassification errors, particularly in Malayalam. This pattern suggests a broader challenge in differentiating between the phonetic characteristics of these languages, necessitating further refinement in the model's training process. The precision, recall, and F1-score metrics provide additional insights into the model's performance across different classes. The model demonstrates the highest precision for Odiya, indicating that when it predicts Odiya, it is often correct. However, the low recall for Odiya means that many Odiya instances are not being correctly identified. Malayalam shows a more balanced performance with relatively higher recall and F1-scores, suggesting that the model can correctly identify Malayalam instances more frequently. Both Bangla and Telugu have consistently low precision, recall, and F1-scores, indicating significant challenges in accurately detecting these accents. This demonstrates GPT-4o's limited ability to recognize and differentiate between various English accents, which is essential for enhancing the usability of ASR systems in multilingual and multicultural environments.

5. Multimodal Capacity of GPT-4o

The ability to integrate and interpret information from multiple modalities is crucial for developing advanced intelligent systems. Multimodal capacity refers to the capability of a model to understand and synthesize information from various sources such as text, images, and audio. This enables the model to generate more comprehensive and contextually enriched responses. The objective of assessing GPT-4o's multimodal capacity is to evaluate its performance across tasks that require the integration of different types of data.

5.1. Visual Question Answering

The Visual Question Answering (VQA) dataset is a multimodal benchmark that combines computer vision and NLP tasks. It consists of images paired with natural language questions related to the visual content (https://www.kaggle.com/datasets/bhavikardeshna/visual-question-answering-computer-vision-nlp). The goal is to produce accurate natural language answers by comprehending the semantics of both the image and the question. For this evaluation, a subset of 100 image-question pairs was sampled from the dataset. GPT-4o was tasked with analyzing the provided image and the corresponding question and generating an appropriate answer chosen from a predefined list of possible answers. The model's generated answers were compared against the ground truth answers to assess its performance in this AI-complete task, which involves a wide range of sub-problems such as object detection, scene classification, and multimodal reasoning. The maximum accuracy was 0.36, as shown in Figure 16.
The performance shows some variability with different shot numbers, peaking at 0.36 accuracy with eight shots. Interestingly, the model's performance decreases with just one example, suggesting that providing few examples in a task with many options may not always be beneficial. This decrease in performance could be due to the distribution of answers becoming skewed by the unrelated task, given the diverse possibilities in VQA.

5.2. Vision-Language Capabilities

Vision-language (VL) capabilities represent a critical advancement in the development of AI models that can understand and interpret multimodal data, integrating both visual and linguistic information to perform complex tasks. The ability to combine these two types of data allows for a more nuanced understanding of content, which is essential for applications ranging from image captioning to more sophisticated tasks like explaining visual jokes or reasoning about events depicted in images.
To evaluate the vision-language capabilities of GPT-4o, we employed the MM-Vet benchmark (Yu et al., 2023). MM-Vet is designed to systematically assess large multimodal models (LMMs) on a variety of integrated tasks that require a combination of core VL capabilities, including recognition, optical character recognition (OCR), knowledge, language generation, spatial awareness, and math. This evaluation framework ensures comparison across diverse question types and answer styles and provides insights beyond simple performance rankings.
The MM-Vet benchmark includes tasks that necessitate the integration of these capabilities to solve complex problems. For instance, a task might involve recognizing objects in an image, understanding the spatial relationships between them, reading and interpreting text within the image, and generating a coherent textual response that incorporates external knowledge. The evaluation metrics employed by MM-Vet are based on an LLM-based evaluator that uses few-shot learning to provide scores for open-ended model outputs. This approach allows for consistent and comprehensive evaluation across different answer styles and question types. We compare the performance of GPT-4o with its predecessors in Table 14.
The results from the MM-Vet benchmark highlight the advancements made with GPT-4o in VL capabilities compared to its predecessors. As summarized in Table 14, GPT-4o outperforms previous models across all evaluated metrics. GPT-4o's notable performance across all metrics highlights its advanced VL capabilities, setting a new benchmark for multimodal models. The high scores in knowledge, spatial awareness, and language generation tasks, in particular, highlight GPT-4o's ability to understand and produce contextually relevant responses based on visual inputs, making it versatile in various applications. The high performance noted is also consistent with that reported by Zhu et al., (2024). Table 15 provides examples of GPT-4o responses for several images and prompts.

6. Implications, Limitations, and Future Work

This section summarizes the key implications of our findings, acknowledges the limitations of the study, and outlines potential directions for future research.

6.1. Implications

The findings from this research have significant implications for the development and application of LLMs in various fields. GPT-4o's high performance in tasks like medical exam question answering and financial analysis suggests its potential utility in educational and professional training environments. The model's ability to integrate vision and language data effectively positions it as a valuable tool in fields requiring multimodal analysis, such as healthcare, finance, and customer service. The demonstrated proficiency in few-shot learning highlights the model's potential for applications where data is scarce or expensive. This could lead to more accessible AI-driven solutions in underrepresented languages and domains, offering inclusivity and broader application of AI technologies.
Moreover, the need to evaluate newer models on comprehensive and diverse sets of data and tasks is underscored by this research. The gap in robust and extensive evaluations has been a notable limitation in understanding the full capabilities and potential weaknesses of advanced models like GPT-4o. This calls for the development and adoption of more comprehensive benchmarks that can rigorously test models across a wider array of real-world scenarios. The findings also suggest implications for policy and regulatory frameworks. As AI models become increasingly integrated into critical sectors such as healthcare and finance, ensuring their reliability, transparency, and fairness becomes necessary (Hayawi & Shahriar, 2024). This necessitates continuous monitoring, rigorous testing, and the establishment of standards to guide the ethical deployment of AI technologies.

6.2. Limitations

Despite the promising results presented in this study, several limitations must be acknowledged. Firstly, the evaluation datasets used in various tasks, particularly in image and audio data, were relatively small and not exhaustive. This limited sample size may not fully capture the model's performance across all potential scenarios. While we aimed for a comprehensive evaluation across data types and multimodal (breadth), the categories within each are not exhaustive (depth). For example, we did not evaluate image and audio generation as it was beyond the scope of this study.
Moreover, qualitative or human judgment was not used as a criterion to assess performance. Incorporating human judgment is crucial for evaluating the practical usability and contextual accuracy of model outputs, as it provides insights that quantitative metrics alone may not reveal (Shahriar, 2022). The model also exhibited inconsistencies in handling ambiguous or complex inputs, as seen in the varying accuracy rates across different tasks. Furthermore, the few-shot learning approach, although beneficial in some contexts, showed limitations in tasks with a high degree of variability, such as VQA. The potential for overfitting to specific examples in these cases remains a concern. Additionally, the lack of real-time and longitudinal data evaluation poses a constraint on understanding the model's adaptability and robustness over time. For example, evaluating the model's performance in real-time applications, such as continuously monitoring driver drowsiness or detecting sudden changes in patient health through medical imaging, would provide valuable insights into its practical effectiveness and reliability under dynamic conditions.

6.3. Future Work

Building on the existing research, this paper highlights several avenues for future research directions. Expanding the evaluation datasets to include a more diverse and comprehensive range of tasks will provide a deeper understanding of the model's capabilities and limitations. Integrating real-time and longitudinal data assessments can offer insights into the model's adaptability and performance stability over extended periods. Further refinement of the few-shot learning techniques is essential, especially for tasks with high variability. Exploring advanced prompting strategies and incorporating more contextual understanding (Sivarajkumar et al., 2024) could enhance performance in these areas. It is thus important to also investigate the impact of prompt quality on model performance. Additionally, understanding the reasons behind the model's low performance and conducting thorough error analysis are crucial. This involves examining how and why the model failed in specific tasks to inform targeted training and fine-tuning efforts. Such analysis will provide valuable insights into the model's limitations and guide improvements to enhance its utility in nuanced language understanding tasks.
Future work should also prioritize creating and adopting new, comprehensive benchmarks that evaluate models across diverse tasks and datasets, addressing the current gap in robust model evaluation. This approach will ensure a holistic understanding of the model's performance, guiding improvements and encouraging the development of more reliable AI systems. The current multimodal evaluation only investigated image and text inputs, highlighting the necessity to explore other inputs and their combinations. For instance, incorporating audio, image, and text together could significantly contribute to cross-domain applications and arts (Shahriar & Al Roken, 2022), enhancing the model's utility in various fields. Lastly, incorporating qualitative assessments and human judgment in the evaluation process will provide a more nuanced understanding of the model's practical applicability and contextual performance. This can help identify areas where the model performs well in real-world scenarios and where it may require further enhancement.

References

  1. Ahamad, A., Anand, A., & Bhargava, P. (2020). AccentDB: a database of non-native english accents to assist neural speech recognition. Proceedings of the 12th Language Resources and Evaluation Conference, 5351–5358. https://www.aclweb.org/anthology/2020.lrec-1.659.
  2. Aher, G. V., Arriaga, R. I., & Kalai, A. T. (2023). Using Large Language Models to Simulate Multiple Humans and Replicate Human Subject Studies. Proceedings of the 40th International Conference on Machine Learning, 337–371. https://proceedings.mlr.press/v202/aher23a.html.
  3. Allyn, B. (2024). Scarlett Johansson says she is “shocked, angered” over new ChatGPT voice. https://www.npr.org/2024/05/20/1252495087/openai-pulls-ai-voice-that-was-compared-to-scarlett-johansson-in-the-movie-her.
  4. Bhagavatula, C., Bras, R. L., Malaviya, C., Sakaguchi, K., Holtzman, A., Rashkin, H., Downey, D., Yih, W., & Choi, Y. (2019, September 25). Abductive Commonsense Reasoning. International Conference on Learning Representations. https://openreview.net/forum?id=Byg1v1HKDB.
  5. Brin, D., Sorin, V., Vaid, A., Soroush, A., Glicksberg, B. S., Charney, A. W., Nadkarni, G., & Klang, E. (2023). Comparing ChatGPT and GPT-4 performance in USMLE soft skill assessments. Scientific Reports, 13(1), 16492. [CrossRef]
  6. Callanan, E., Mbakwe, A., Papadimitriou, A., Pei, Y., Sibue, M., Zhu, X., Ma, Z., Liu, X., & Shah, S. (2023). Can GPT models be Financial Analysts? An Evaluation of ChatGPT and GPT-4 on mock CFA Exams (arXiv:2310.08678). arXiv. [CrossRef]
  7. CFA Institute. (n.d.). CFA program curriculum level I. https://www.cfainstitute.org.
  8. College Board. (n.d.). The SAT suite of assessments. https://www.collegeboard.org.
  9. Coyne, S., Sakaguchi, K., Galvan-Sosa, D., Zock, M., & Inui, K. (2023). Analyzing the Performance of GPT-3.5 and GPT-4 in Grammatical Error Correction (arXiv:2303.14342). arXiv. [CrossRef]
  10. Cubero, S., Aleixos, N., Moltó, E., Gómez-Sanchis, J., & Blasco, J. (2011). Advances in machine vision applications for automatic inspection and quality evaluation of fruits and vegetables. Food and Bioprocess Technology, 4, 487–504. [CrossRef]
  11. Dalvi, B., Jansen, P., Tafjord, O., Xie, Z., Smith, H., Pipatanangkura, L., & Clark, P. (2021). Explaining Answers with Entailment Trees. In M.-F. Moens, X. Huang, L. Specia, & S. W. Yih (Eds.), Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (pp. 7358–7370). Association for Computational Linguistics. [CrossRef]
  12. Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 4171–4186.
  13. Dillion, D., Mondal, D., Tandon, N., & Gray, K. (2024). Large Language Models as Moral Experts? GPT-4o Outperforms Expert Ethicist in Providing Moral Guidance. OSF. [CrossRef]
  14. Floridi, L., & Chiriatti, M. (2020). GPT-3: Its Nature, Scope, Limits, and Consequences. Minds and Machines, 30(4), 681–694. [CrossRef]
  15. Gemini Team, Anil, R., Borgeaud, S., Alayrac, J.-B., Yu, J., Soricut, R., Schalkwyk, J., Dai, A. M., Hauth, A., Millican, K., Silver, D., Johnson, M., Antonoglou, I., Schrittwieser, J., Glaese, A., Chen, J., Pitler, E., Lillicrap, T., Lazaridou, A., … Vinyals, O. (2024). Gemini: A Family of Highly Capable Multimodal Models (arXiv:2312.11805). arXiv. [CrossRef]
  16. Gilson, A., Safranek, C. W., Huang, T., Socrates, V., Chi, L., Taylor, R. A., & Chartash, D. (2023). How Does ChatGPT Perform on the United States Medical Licensing Examination (USMLE)? The Implications of Large Language Models for Medical Education and Knowledge Assessment. JMIR Medical Education, 9(1), e45312. [CrossRef]
  17. Graham, C., & Roll, N. (2024). Evaluating OpenAI’s Whisper ASR: Performance analysis across diverse accents and speaker traits. JASA Express Letters, 4(2). [CrossRef]
  18. Griggs, M. (2019). Building a Better Bar Exam. Texas A&M Law Review, 7, 1.
  19. Haleem, A., Javaid, M., Qadri, M. A., & Suman, R. (2022). Understanding the role of digital technologies in education: A review. Sustainable Operations and Computers, 3, 275–285. [CrossRef]
  20. Hayawi, K., & Shahriar, S. (2024). AI Agents from Copilots to Coworkers: Historical Context, Challenges, Limitations, Implications, and Practical Guidelines (2024040709). Preprints. [CrossRef]
  21. Hayawi, K., Shahriar, S., & Mathew, S. S. (2024). The imitation game: Detecting human and AI-generated texts in the era of ChatGPT and BARD. Journal of Information Science, 01655515241227531. [CrossRef]
  22. Hayes, B. K., Heit, E., & Swendsen, H. (2010). Inductive reasoning. WIREs Cognitive Science, 1(2), 278–292. [CrossRef]
  23. Hu, B., Sheng, Q., Cao, J., Shi, Y., Li, Y., Wang, D., & Qi, P. (2024). Bad actor, good advisor: Exploring the role of large language models in fake news detection. Proceedings of the AAAI Conference on Artificial Intelligence, 38, Article 20. [CrossRef]
  24. Huang, J., & Chang, K. C.-C. (2023). Towards Reasoning in Large Language Models: A Survey. In A. Rogers, J. Boyd-Graber, & N. Okazaki (Eds.), Findings of the Association for Computational Linguistics: ACL 2023 (pp. 1049–1065). Association for Computational Linguistics. [CrossRef]
  25. Jebraeily, Y., Sharafi, Y., & Teshnehlab, M. (2024). Driver drowsiness detection based on convolutional neural network architecture optimization using genetic algorithm. IEEE Access : Practical Innovations, Open Solutions, 12, 45709–45726. [CrossRef]
  26. Johnson-Laird, P. (2010). Deductive reasoning. WIREs Cognitive Science, 1(1), 8–17. [CrossRef]
  27. Katz, D. M., Bommarito, M. J., Gao, S., & Arredondo, P. (2024). Gpt-4 passes the bar exam. Philosophical Transactions of the Royal Society A, 382(2270), 20230254. [CrossRef]
  28. Khoshafah, F. (2023). ChatGPT for Arabic-English Translation: Evaluating the Accuracy. [CrossRef]
  29. Korinek, A. (2023). Language models and cognitive automation for economic research. National Bureau of Economic Research. [CrossRef]
  30. Koubaa, A. (2023). GPT-4 vs. GPT-3.5: A Concise Showdown (2023030422). Preprints. [CrossRef]
  31. Kunchukuttan, A., Mehta, P., & Bhattacharyya, P. (2018). The IIT bombay english-hindi parallel corpus. Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018).
  32. Li, H., Ding, W., Kang, Y., Liu, T., Wu, Z., & Liu, Z. (2021). CTAL: Pre-training cross-modal transformer for audio-and-language representations. Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 3966–3977.
  33. López Espejel, J., Ettifouri, E. H., Yahaya Alassan, M. S., Chouham, E. M., & Dahhane, W. (2023). GPT-3.5, GPT-4, or BARD? Evaluating LLMs reasoning ability in zero-shot setting and performance boosting through prompts. Natural Language Processing Journal, 5, 100032. [CrossRef]
  34. Lund, B. D., Wang, T., Mannuru, N. R., Nie, B., Shimray, S., & Wang, Z. (2023). ChatGPT and a new academic reality: Artificial Intelligence-written research papers and the ethics of the large language models in scholarly publishing. Journal of the Association for Information Science and Technology, 74(5), 570–581. [CrossRef]
  35. Mannuru, N. R., Shahriar, S., Teel, Z. A., Wang, T., Lund, B. D., Tijani, S., Pohboon, C. O., Agbaji, D., Alhassan, J., Galley, J., & others. (2023). Artificial intelligence in developing countries: The impact of generative artificial intelligence (AI) technologies for development. Information Development, 02666669231200628. [CrossRef]
  36. National Conference of Bar Examiners. (n.d.). MBE sample test questions. https://www.ncbex.org.
  37. of State Medical Boards, F., & of Medical Examiners, N. B. (2024). USMLE step 1 content description and general information. https://www.usmle.org.
  38. Ongsulee, P. (2017). Artificial intelligence, machine learning and deep learning. 2017 15th International Conference on ICT and Knowledge Engineering (ICT&KE), 1–6.
  39. OpenAI, Achiam, J., Adler, S., Agarwal, S., Ahmad, L., Akkaya, I., Aleman, F. L., Almeida, D., Altenschmidt, J., Altman, S., Anadkat, S., Avila, R., Babuschkin, I., Balaji, S., Balcom, V., Baltescu, P., Bao, H., Bavarian, M., Belgum, J., … Zoph, B. (2024). GPT-4 Technical Report (arXiv:2303.08774). arXiv. [CrossRef]
  40. Oura, T., Tatekawa, H., Horiuchi, D., Matsushita, S., Takita, H., Atsukawa, N., Mitsuyama, Y., Yoshida, A., Murai, K., Tanaka, R., Shimono, T., Yamamoto, A., Miki, Y., & Ueda, D. (2024). Diagnostic Accuracy of Vision-Language Models on Japanese Diagnostic Radiology, Nuclear Medicine, and Interventional Radiology Specialty Board Examinations (p. 2024.05.31.24308072). medRxiv. [CrossRef]
  41. Pang, S., Nol, E., & Heng, K. (2024). ChatGPT-4o for English language teaching and learning: Features, applications, and future prospects (SSRN Scholarly Paper 4837988). [CrossRef]
  42. Ray, S. (2024). Google CEO says Gemini AI’s ‘unacceptable’ responses offended users and showed bias. https://www.forbes.com/sites/siladityaray/2024/02/28/google-ceo-says-gemini-ais-unacceptable-responses-offended-users-and-showed-bias/?sh=250e1a1b1103.
  43. Salman, S., & Liu, X. (2019). Overfitting Mechanism and Avoidance in Deep Neural Networks (arXiv:1901.06566). arXiv. [CrossRef]
  44. Shahriar, S. (2022). GAN computers generate arts? A survey on visual arts, music, and literary text generation using generative adversarial network. Displays, 73, 102237. [CrossRef]
  45. Shahriar, S., & Al Roken, N. (2022). How can generative adversarial networks impact computer generated art? Insights from poetry to melody conversion. International Journal of Information Management Data Insights, 2(1), 100066. [CrossRef]
  46. Shahriar, S., Al Roken, N., & Zualkernan, I. (2023). Classification of Arabic poetry emotions using deep learning. Computers, 12(5), 89. [CrossRef]
  47. Shen, X., Wu, Y., Backes, M., & Zhang, Y. (2024). Voice Jailbreak Attacks Against GPT-4o (arXiv:2405.19103). arXiv. [CrossRef]
  48. Singgalen, Y. A. (2024). Analyzing an Interest in GPT 4o through Sentiment Analysis using CRISP-DM. Journal of Information Systems and Informatics, 6(2), Article 2. [CrossRef]
  49. Sinha, K., Sodhani, S., Dong, J., Pineau, J., & Hamilton, W. L. (2019). CLUTRR: A Diagnostic Benchmark for Inductive Reasoning from Text. In K. Inui, J. Jiang, V. Ng, & X. Wan (Eds.), Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP) (pp. 4506–4515). Association for Computational Linguistics. [CrossRef]
  50. Sivarajkumar, S., Kelley, M., Samolyk-Mazzanti, A., Visweswaran, S., & Wang, Y. (2024). An Empirical Evaluation of Prompting Strategies for Large Language Models in Zero-Shot Clinical Natural Language Processing: Algorithm Development and Validation Study. JMIR Medical Informatics, 12(1), e55318. [CrossRef]
  51. Sonoda, Y., Kurokawa, R., Nakamura, Y., Kanzawa, J., Kurokawa, M., Ohizumi, Y., Gonoi, W., & Abe, O. (2024). Diagnostic Performances of GPT-4o, Claude 3 Opus, and Gemini 1.5 Pro in “Diagnosis Please” Cases (p. 2024.05.26.24307915). medRxiv. [CrossRef]
  52. Thirunavukarasu, A. J., Ting, D. S. J., Elangovan, K., Gutierrez, L., Tan, T. F., & Ting, D. S. W. (2023). Large language models in medicine. Nature Medicine, 29(8), 1930–1940. [CrossRef]
  53. Tiedemann, J. (2012). Parallel data, tools and interfaces in OPUS. Proceedings of the Eighth International Conference on Language Resources and Evaluation (LREC’12), 2214–2218.
  54. Walton, D. (2014). Abductive reasoning. University of Alabama Press.
  55. Wang, H., Xu, J., Xie, S., Wang, R., Li, J., Xie, Z., Zhang, B., Xiong, C., & Chen, X. (2024). M4U: Evaluating Multilingual Understanding and Reasoning for Large Multimodal Models (arXiv:2405.15638). arXiv. [CrossRef]
  56. Weston, J., Bordes, A., Chopra, S., Rush, A. M., van Merriënboer, B., Joulin, A., & Mikolov, T. (2015). Towards AI-Complete Question Answering: A Set of Prerequisite Toy Tasks (arXiv:1502.05698). arXiv. [CrossRef]
  57. Xu, S., Wang, Y., Liu, D., & Xu, C. (2024). Collage Prompting: Budget-Friendly Visual Recognition with GPT-4V (arXiv:2403.11468). arXiv. [CrossRef]
  58. Ying, Z., Liu, A., Liu, X., & Tao, D. (2024). Unveiling the Safety of GPT-4o: An Empirical Study using Jailbreak Attacks (arXiv:2406.06302). arXiv. [CrossRef]
  59. Yu, W., Yang, Z., Li, L., Wang, J., Lin, K., Liu, Z., Wang, X., & Wang, L. (2023). MM-Vet: Evaluating Large Multimodal Models for Integrated Capabilities (arXiv:2308.02490). arXiv. [CrossRef]
  60. Zhang, N., Sun, Z., Xie, Y., Wu, H., & Li, C. (n.d.). The latest version ChatGPT powered by GPT-4o: What will it bring to the medical field? International Journal of Surgery, 10.1097/JS9.0000000000001754. [CrossRef]
  61. Zhou, Y., Ong, H., Kennedy, P., Wu, C. C., Kazam, J., Hentel, K., Flanders, A., Shih, G., Peng, Y., Moy, L., & Atzen, S. (2024). Evaluating GPT-4V (GPT-4 with Vision) on Detection of Radiologic Findings on Chest Radiographs. Radiology, 311(2), e233270. [CrossRef]
  62. Zhu, N., Zhang, N., Shao, Q., Cheng, K., & Wu, H. (2024). OpenAI’s GPT-4o in surgical oncology: Revolutionary advances in generative artificial intelligence. European Journal of Cancer, 206. [CrossRef]
Figure 1. Visualization of the Relationship Between General AI and GPT 4o.
Figure 1. Visualization of the Relationship Between General AI and GPT 4o.
Preprints 110113 g001
Figure 2. Example of GPT-4o Incorrect Answer on SAT Reading & Writing: Question (Top), GPT-4o Response (Bottom Left), and Correct Answer (Bottom Right).
Figure 2. Example of GPT-4o Incorrect Answer on SAT Reading & Writing: Question (Top), GPT-4o Response (Bottom Left), and Correct Answer (Bottom Right).
Preprints 110113 g002
Figure 3. Example of GPT-4o Correct Answer on SAT Reading & Writing: Question (Top), GPT-4o Response (Bottom Left), and Correct Answer (Bottom Right).
Figure 3. Example of GPT-4o Correct Answer on SAT Reading & Writing: Question (Top), GPT-4o Response (Bottom Left), and Correct Answer (Bottom Right).
Preprints 110113 g003
Figure 4. Example of GPT-4o Incorrect Answer on SAT Math: Question (Top), GPT-4o Response (Bottom Left), and Correct Answer (Bottom Right).
Figure 4. Example of GPT-4o Incorrect Answer on SAT Math: Question (Top), GPT-4o Response (Bottom Left), and Correct Answer (Bottom Right).
Preprints 110113 g004
Figure 5. Example of GPT-4o Correct Answer on SAT Math: Question (Top), GPT-4o Response (Bottom Left), and Correct Answer (Bottom Right).
Figure 5. Example of GPT-4o Correct Answer on SAT Math: Question (Top), GPT-4o Response (Bottom Left), and Correct Answer (Bottom Right).
Preprints 110113 g005
Figure 6. Types of Reasoning.
Figure 6. Types of Reasoning.
Preprints 110113 g006
Figure 7. Logical Reasoning Categories and Datasets.
Figure 7. Logical Reasoning Categories and Datasets.
Preprints 110113 g007
Figure 8. Confusion Matrix (left) and Performance Comparison (right) for Fruits Classification.
Figure 8. Confusion Matrix (left) and Performance Comparison (right) for Fruits Classification.
Preprints 110113 g008
Figure 9. Confusion Matrix (left) and Performance Comparison (right) for Drowsiness Detection.
Figure 9. Confusion Matrix (left) and Performance Comparison (right) for Drowsiness Detection.
Preprints 110113 g009
Figure 10. Confusion Matrix (top) and performance comparison (bottom) for Crop Disease Classification.
Figure 10. Confusion Matrix (top) and performance comparison (bottom) for Crop Disease Classification.
Preprints 110113 g010aPreprints 110113 g010b
Figure 11. Confusion Matrix (left) and Performance Comparison (right) for Glaucoma Detection.
Figure 11. Confusion Matrix (left) and Performance Comparison (right) for Glaucoma Detection.
Preprints 110113 g011
Figure 12. Performance Evolution Against Number of Shots for Glaucoma Detection.
Figure 12. Performance Evolution Against Number of Shots for Glaucoma Detection.
Preprints 110113 g012
Figure 13. Confusion Matrix for Cancer, Tumor, and Aneurysm Detection Task.
Figure 13. Confusion Matrix for Cancer, Tumor, and Aneurysm Detection Task.
Preprints 110113 g013
Figure 14. Confusion Matrix (top) and Performance Comparison (bottom) for Audio Emotion Detection.
Figure 14. Confusion Matrix (top) and Performance Comparison (bottom) for Audio Emotion Detection.
Preprints 110113 g014
Figure 15. Confusion Matrix (Top) and Performance Comparison (Bottom) for Accent Detection.
Figure 15. Confusion Matrix (Top) and Performance Comparison (Bottom) for Accent Detection.
Preprints 110113 g015aPreprints 110113 g015b
Figure 16. Accuracy by Numbers of Shots in Visual Question Answering (VQA).
Figure 16. Accuracy by Numbers of Shots in Visual Question Answering (VQA).
Preprints 110113 g016
Table 1. Performance Comparison of GPT models on USMLE.
Table 1. Performance Comparison of GPT models on USMLE.
Model Total Questions Correct Answers Accuracy
GPT-3.5 389 201 51.67%
GPT-4 80 72 90.00%
GPT-4o 118 98 83.05%
Table 2. Performance Comparison of GPT models on CFA Level 1 Exam.
Table 2. Performance Comparison of GPT models on CFA Level 1 Exam.
Model Accuracy
GPT-3.5 58.80%
GPT-4 73.20%
GPT-4o 85.39%
Table 3. GPT-4o Performance on SAT.
Table 3. GPT-4o Performance on SAT.
Test Total Questions Correct Answers Accuracy
Reading & Writing M1 33 31 93.94%
Reading & Writing M2 33 29 87.88%
Math M1 27 25 92.59%
Math M2 27 22 81.48%
Table 4. Performance Comparison of GPT models on SAT.
Table 4. Performance Comparison of GPT models on SAT.
Model Reading & Writing Math
GPT-3.5 83.75% 73.75%
GPT-4 88.75% 87.50%
GPT-4 (no vision) 88.75% 86.25%
GPT-4o 90.91% 87.04%
Table 5. Performance Comparison of GPT models on MBE.
Table 5. Performance Comparison of GPT models on MBE.
Model Accuracy
GPT-3.5 45.10%
GPT-4 75.70%
GPT-4o 75.00%
Table 6. Performance of GPT Models on Logical Reasoning Tasks.
Table 6. Performance of GPT Models on Logical Reasoning Tasks.
Category/
Model
Deductive Reasoning Inductive Reasoning Abductive Reasoning
Entailment Bank bAbI (task 15) CLUTRR bAbI (task 15) αNLI
GPT 3.5 25/30 26/30 2/30 14/30 19/30
GPT 4 27/30 30/30 11/30 28/30 25/30
GPT 4o 29/30 30/30 17/30 30/30 27/30
Table 7. GPT-4o Translation Accuracy Across Languages.
Table 7. GPT-4o Translation Accuracy Across Languages.
Language Translation Accuracy (%)
Spanish 88
Arabic 78
Hindi 82
French 75
Portuguese 86
Russian 80
Table 8. GPT-4o Performance on Fruit Classification.
Table 8. GPT-4o Performance on Fruit Classification.
Class Precision Recall F1-Score
Banana 0.91 1.00 0.95
Papaya 1.00 1.00 1.00
Apple 1.00 1.00 1.00
Litchi 1.00 1.00 1.00
Jackfruit 1.00 0.90 0.95
Hog Plum 1.00 1.00 1.00
Grapes 1.00 1.00 1.00
Guava 1.00 1.00 1.00
Mango 0.91 1.00 0.95
Orange 1.00 0.90 0.95
Table 9. GPT-4o Performance on Drowsiness Detection.
Table 9. GPT-4o Performance on Drowsiness Detection.
Class Precision Recall F1-Score
Drowsy 0.8 0.8 0.8
Natural 0.8 0.8 0.8
Table 10. GPT-4o Performance on Crop Disease Detection.
Table 10. GPT-4o Performance on Crop Disease Detection.
Class Precision Recall F1-score
Anthracnose 0.60 0.60 0.60
Apple Scab 1.00 0.80 0.89
Black Spot 0.67 1.00 0.80
Blight 0.38 0.75 0.50
Blossom End Rot 1.00 1.08 1.00
Botrytis 1.00 0.20 0.33
Brown Rot 1.00 0.20 0.33
Canker 0.25 0.25 0.25
Cedar Apple Rust 0.83 1.00 0.91
Clubroot 1.00 1.00 1.00
Crown Gall 1.00 1.00 1.00
Downy Mildew 1.00 0.20 0.33
Fire Blight 0.80 0.80 0.80
Fusarium 1.00 0.60 0.75
Gray Mold 0.43 0.75 0.55
Leaf Spots 0.40 0.80 0.53
Mosaic Virus 0.67 0.80 0.72
Nematodes 1.00 1.00 1.00
Powdery Mildew 0.71 1.00 0.83
Verticillium 0.67 0.40 0.50
Table 11. GPT-4o Performance on Glaucoma Detection.
Table 11. GPT-4o Performance on Glaucoma Detection.
Class Precision Recall F1-Score
Glaucoma 0.58 0.86 0.69
Normal 0.73 0.38 0.50
Table 12. GPT-4o Performance on Cancer, Tumor, and Aneurysm Detection.
Table 12. GPT-4o Performance on Cancer, Tumor, and Aneurysm Detection.
Class Precision Recall F1-Score
Aneurysm 0.3725 0.5758 0.4524
Cancer 0.0000 0.0000 0.0000
Tumor 0.2708 0.3939 0.3210
Table 13. GPT-4o Performance on Image Captioning.
Table 13. GPT-4o Performance on Image Captioning.
Metric Score
BLEU-1 0.193
BLEU-2 0.095
BLEU-3 0.058
BLEU-4 0.031
Table 14. Performance Comparison of GPT Models on VL capabilities.
Table 14. Performance Comparison of GPT Models on VL capabilities.
Model Recognition OCR Knowledge Language Generation Spatial Awareness Math Total
GPT-3.5 22.3 31.4 15.6 16.6 32.9 24.0 27.6
GPT-4 34.3 66.3 25.6 36.6 60.6 72.0 48.1
GPT-4V 67.5 68.3 56.2 60.7 69.4 58.6 67.7
GPT-4o 80.6 87.5 75.7 77.1 84.7 88.1 83.9
Table 15. Instances of GPT-4o Responses with Visual Inputs (incorrect answers highlighted in red).
Table 15. Instances of GPT-4o Responses with Visual Inputs (incorrect answers highlighted in red).
Input Image Questions|Assessed Capabilities GPT-4o Responses
Preprints 110113 i001 a) Q: How many gallons of supreme gasoline can I get with $50? GPT-4o: a) With $50, you can get approximately 13.70 gallons of supreme gasoline.
b) Q: What is the price difference between diesel and regular gasoline?

Required capabilities: OCR, Math
GPT-4o: The price difference between diesel and regular gasoline is $0.11 per gallon.
Preprints 110113 i002 a) Q: Based on the image, what is the total international awareness of Tina Fey? GPT-4o: Based on the image, the total international awareness of Tina Fey is 35%. This percentage is located at the bottom row under the column labeled "Total" for Tina Fey.
b) Q: Based on the image, in which country does Tina Fey have the highest level of total awareness?

Required capabilities: OCR, Spatial awareness
GPT-4o: Based on the image, Tina Fey has the highest level of total awareness in Australia, with an awareness level of 62%.
Preprints 110113 i003 a) Q: What is the price for tomatoes? GPT-4o: The price for tomatoes is 8.00, as indicated by the sign placed above the pile of tomatoes in the image.
b) Q: What fruit is the the right of plums?

Required capabilities: Recognition, Spatial awareness
GPT-4o: The fruit to the right of the plums in the image is pears.
Preprints 110113 i004 a) Q: Which car is on the parking spot 33? GPT-4o: The parking spot labeled 33 is empty.
b) Q: Which spot is the red car parking on?

Required capabilities: OCR, Recognition, Spatial awareness
GPT-4o: The red car is parked in spot number 30.
Preprints 110113 i005 a) Q: Which category did the worst job in plastic recycling, considering the amount of plastic made and thrown out? GPT-4o: The category that did the worst job in plastic recycling, considering the amount of plastic made and thrown out, is "Packaging." In 2015, 146 million metric tonnes of plastic were made for packaging, and 141 million metric tonnes were thrown out. This shows that almost all the plastic made for packaging was discarded as waste.
b) Q: What is the difference in metric tonnes between the amount of plastic produced and the amount discarded?

Required capabilities: OCR, Spatial awareness
The difference between the amount of plastic produced (407 million metric tonnes) and the amount discarded (302 million metric tonnes) is:

407 million metric tonnes – 302 million metric tonnes = 105 million metric tonnes

So, the difference is 105 million metric tonnes.
Preprints 110113 i006 Q: What does this factory make?

Required capabilities: OCR
GPT-4o: The factory in the image belongs to Leica, which is a company known for manufacturing cameras, lenses, and other optical equipment.
Preprints 110113 i007 Q: In which country was this photo taken?

Required capabilities: Recognition, Knowledge
GPT-4o: The photo shows Novak Djokovic kissing a trophy that appears to be the Australian Open trophy. Therefore, this photo was likely taken in Australia.
Preprints 110113 i008 Q: Which are producers in this food web?

Required capabilities: OCR, Knowledge, Spatial Awareness
GPT-4o: In this food web, the producers are:

Phytoplankton
Seaweed
These are the organisms that produce energy through photosynthesis and serve as the base of the food web.
Preprints 110113 i009 Q: Is it summertime in the scene? Answer it and give the rationale.

Required capabilities: Recognition, Knowledge, Language generation
GPT-4o: No, it is not summertime in the scene. The presence of large pumpkins suggests that it is likely fall, as pumpkins are typically associated with the autumn harvest season. Additionally, the characters are wearing long-sleeved clothing, which is more common in cooler weather.
Preprints 110113 i010 Q: Can you explain this meme?

Required capabilities: Recognition, OCR, Knowledge, Language generation
GPT-4o: This meme humorously illustrates the common tendency to procrastinate. The text shows an intention to start a task at a specific time (8:00), but when the time is slightly missed (8:05), the person decides to delay the task further, waiting until the next hour (9:00) to start. The image of Kermit the Frog relaxing on a bed adds to the humor by depicting the relaxed attitude towards procrastination. The meme pokes fun at the irrational yet relatable habit of postponing tasks due to missing an arbitrary starting time by a few minutes.
Preprints 110113 i011 Q: What is funny about this image?

Required capabilities: Recognition, Knowledge, Language generation
GPT-4o: This image is a popular meme that combines two unrelated photos to create a humorous juxtaposition.
On the left side, there is an image of a woman (Taylor Armstrong from "The Real Housewives of Beverly Hills") angrily pointing and shouting, while another woman (Kyle Richards) tries to calm her down.
On the right side, there is a photo of a white cat (known as "Smudge the Cat") sitting at a dinner table with a confused and displeased expression in front of a plate of vegetables.
The humor comes from the contrast between the woman's intense emotion and the cat's indifferent and bemused reaction, as if the cat is unbothered by the woman's outburst. This combination creates a relatable and funny depiction of exaggerated reactions and misunderstandings.
Preprints 110113 i012 Q: How to cook this dish?

Required capabilities: Recognition, Knowledge, Language generation
GPT-4o: This image shows a pie with a lattice crust. Here is a basic recipe to make a lattice-top pie, such as a strawberry or mixed berry pie:
Ingredients: For the Filling: 4 cups fresh strawberries or mixed berries, 3/4 cup granulated sugar, 1/4 cup cornstarch, 1 tablespoon lemon juice, 1 teaspoon vanilla extract, A pinch of salt For the Crust: 2 1/2 cups all-purpose flour, 1 teaspoon salt, 1 teaspoon sugar, 1 cup unsalted butter, chilled and cut into small pieces, 1/4 to 1/2 cup ice water
Instructions: In a large bowl, combine the flour, salt, and sugar. Add the chilled butter pieces and cut them into the flour mixture using a pastry blender or your fingers until the mixture resembles coarse crumbs. Gradually add ice water, a tablespoon at a time, mixing until the dough begins to come together. Do not overwork the dough. Divide the dough into two equal parts, shape them into disks, wrap in plastic wrap, and refrigerate for at least 1 hour. In a large bowl, combine the berries, sugar, cornstarch, lemon juice, vanilla extract, and salt. Stir gently to combine. On a lightly floured surface, roll out one disk of dough to fit a 9-inch pie pan. Place the rolled-out dough into the pie pan and trim the edges. Pour the berry mixture into the crust-lined pie pan. Roll out the second disk of dough and cut it into 1-inch-wide strips. Lay 4 to 5 strips over the filling, parallel to each other. Lay another 4 to 5 strips perpendicular to the first set, weaving them to create a lattice pattern. Trim and crimp the edges to seal. Preheat the oven to 375°F (190°C). Bake: Place the pie on a baking sheet to catch any drips and bake for 45-55 minutes, or until the crust is golden brown and the filling is bubbly. Cool: Let the pie cool completely before serving to allow the filling to set. Enjoy your homemade lattice-top pie!
Preprints 110113 i013 Q: Can you give a short introduction to this painting?

Required capabilities: Recognition, Knowledge, Language generation
GPT-4o: This painting is a historical Chinese artwork titled "Emperor Taizong Receiving the Tibetan Envoy," created by Yan Liben during the Tang Dynasty. The piece illustrates a significant diplomatic event where Emperor Taizong of the Tang Dynasty meets a Tibetan envoy. This event highlights the political and cultural exchanges between the Tang Dynasty and Tibet.The painting is notable for its detailed depiction of the figures, the hierarchical arrangement, and the ceremonial aspects of the meeting. Yan Liben, a renowned court artist of the Tang Dynasty, is celebrated for his ability to capture the grandeur and solemnity of imperial events. This artwork not only serves as a historical record but also exemplifies the artistic sophistication of the Tang Dynasty.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated