Preprint
Article

Large Language Models in Entrepreneurship: A Survey

Altmetrics

Downloads

347

Views

151

Comments

0

Submitted:

13 September 2024

Posted:

14 September 2024

Read the latest preprint version here

Alerts
Abstract
Large Language Models (LLMs) have emerged as powerful tools in the domain of artificial intelligence, specifically enhancing natural language processing capabilities. These models leverage extensive datasets and sophisticated algorithms to automate and enhance tasks traditionally performed by humans, thereby revolutionizing innovation and entrepreneurship. LLMs accelerate product development, streamline business operations, and enable precise and rapid decision-making, all of which are crucial for maintaining competitiveness in dynamic markets. This paper categorizes the applications of LLMs in innovation and entrepreneurship into three main areas: technological innovation, strategic decision-making, and process automation. By exploring various integrations of LLMs in entrepreneurial ventures, this work offers both theoretical insights and practical examples, underscoring the transformative impact of LLMs in shaping modern business landscapes.
Keywords: 
Subject: Computer Science and Mathematics  -   Artificial Intelligence and Machine Learning

1. Introduction

Large Language Models (LLMs) represent a significant advancement in artificial intelligence, particularly in natural language processing capabilities. These models, equipped with vast amounts of data and sophisticated algorithms, are revolutionizing various sectors by automating complex tasks that traditionally require human intelligence [1,2].In the context of innovation and entrepreneurship, LLMs facilitate the rapid development of new technologies and streamline business operations, thereby enhancing productivity and fostering innovation [3,4].
The impact of LLMs on innovation and entrepreneurship is profound. They accelerate the ideation and product development processes, enable smarter and faster decision-making, and improve customer interactions through personalized services [5,6]. By analyzing large datasets, LLMs provide insights that help businesses anticipate market trends and consumer needs more accurately [7,8]. This capability is crucial for maintaining competitiveness in fast-paced industries.
Currently, the application of LLMs in innovation and entrepreneurship can be categorized into several areas: technological innovation, strategic decision-making, and process automation [9,10,11]. Each category utilizes the strengths of LLMs to tackle specific challenges, from enhancing product design with rapid prototyping to optimizing strategic plans with predictive analytics. These applications demonstrate the versatile role of LLMs in driving business growth and operational efficiency.
This article explores the integration of LLMs in various entrepreneurial ventures, following a structured approach. It begins by defining LLMs and discussing their core functionalities. Subsequent sections delve into the specific applications and benefits of LLMs in business settings, supported by case studies and examples that illustrate successful integrations. Overall, the contribution of this article lies in its comprehensive analysis of how LLMs are being employed to reshape the landscape of innovation and entrepreneurship, providing readers with both theoretical insights and practical examples of this transformative technology.
Figure 1. LLMs in Innovation and Entrepreneurship.
Figure 1. LLMs in Innovation and Entrepreneurship.
Preprints 118115 g001

2. Technological Innovation through LLMs

2.1. LLMs: Definitions and Capabilities

Large Language Models (LLMs) represent a sophisticated advancement in Natural Language Processing (NLP) technology [12,13,14,15,16,17,18]. Built on deep learning neural network architectures, particularly the Transformer model, these models are engineered to perform various language understanding and generation tasks [19,20]. The defining features of LLMs are their scale and capability, which allow them to handle everything from simple text generation to complex language comprehension tasks [21,22].
LLMs leverage the Transformer architecture, which is highly efficient and powerful in processing language tasks. This architecture relies heavily on the self-attention mechanism, enabling the model to weigh the importance of different pieces of information during text generation or comprehension, thereby optimizing its processing pipeline [23,24,25,26,27]. Additionally, the training process of LLMs involves millions of diverse text examples, equipping the model to learn the extensive uses and complexities of language, thus enabling it to generate or comprehend text independently, without needing task-specific directions [17,28,29,30,31].
With their advanced technical features and training methodologies, LLMs excel in various NLP tasks such as summarization, translation, and sentiment analysis [32,33,34,35]. This broad applicability has made LLMs a critical tool not only in academic research but also in commercial and industrial applications [36]. Developers and businesses utilize LLMs to create chatbots, enhance content creation, and improve user interaction experiences, showcasing their substantial potential and value in practical applications [37,38,39].

2.2. LLMs Implementation Basic Process

LLMs are implemented through two main stages: pre-training and fine-tuning [40,41,42]. Together, these stages ensure that the model performs well on specific tasks and is applicable across various language processing scenarios.

Pre-training Stage.

During the pre-training phase, large language models are trained on a vast and diverse corpus of texts, which may include books, articles, and web pages [43,44,45,46]. The goal of pre-training is to enable the model to learn the basic structures of language, grammatical rules, and the relationships between vocabulary [47]. This stage does not rely on task-specific guidance. Instead, the model autonomously learns language features through self-generated methods, such as predicting the next word or reconstructing sentences.

Fine-tuning Stage.

After pre-training, LLMs typically undergo fine-tuning for specific tasks [16,48,49]. During this stage, the model is further trained on smaller, more focused datasets specific to the task at hand, often with annotated information like question-answer pairs or sentiment labels [15]. Fine-tuning allows the model to adjust its parameters to better respond to and optimize for specific tasks, such as sentiment analysis, text classification, or question-answering systems [50,51]. This stage ensures that while the model retains the linguistic knowledge acquired during pre-training, it can also adapt effectively to specific application scenarios.

2.3. LLM Capabilities and Model Technologies

Basic Functions.

LLMs often derive their core functionalities from early model innovations. For instance, the GPT (Generative Pre-trained Transformer) series first demonstrated robust text generation capabilities, autonomously creating coherent and logical text passages based on given prompts [52,53]. The BERT (Bidirectional Encoder Representations from Transformers) model achieved breakthroughs in language understanding by employing bidirectional training to comprehend language in context better. This approach has been widely applied to information retrieval and sentiment analysis tasks [54,55].

Recent Advancements.

As technology progresses, the development of new functionalities often closely aligns with specific model innovations. For example, OpenAI’s GPT-3.5 Turbo [56] has shown exceptional ability in automatic programming assistance, generating programming code from natural language instructions [57,58,59,60,61]. Additionally, BERT and its variants like RoBERTa excel in sentiment analysis, accurately identifying emotions and sentiments from text, which is extensively used in social media analysis and customer feedback systems [46,62,63]. Personalized content creation is driven by more advanced transformer models like T5 (Text-to-Text Transfer Transformer) [64,65,66,67] and the latest versions of GPT, which can tailor highly personalized content responses or suggestions based on user history.

2.4. Detailed Functions and Business Applications

LLMs are crucial tools in innovation and entrepreneurship due to their advanced language-processing capabilities [68,69,70,71,72,73,74]. These models automate and optimize business processes that require language understanding and generation, continuously evolving to provide state-of-the-art technological support and open new opportunities for businesses [75,76]. These abilities help companies maintain a competitive edge in a fiercely competitive market environment:
  • Customer Service Automation. Automating response systems in customer service to handle inquiries, complaints, and interactions without human intervention [77,78]. This reduces the workload on human agents, lowers operational costs, and improves response times and customer satisfaction.
  • Market Analysis. Analyzing market trends, consumer feedback, and competitive dynamics through sentiment analysis and data aggregation [79,80,81,82,83,84]. This enables swift, data-driven decisions that adapt to market changes effectively.
  • Content Creation. Generating high-quality, engaging content for blogs, articles, social media posts, and marketing materials [85,86,87,88]. This helps in scaling content production without compromising quality.
  • Personalization. Personalizing user experiences on digital platforms based on user behavior and preferences [89]. This increases customer loyalty and potentially higher sales through personalized recommendations and communications.
  • Research and Development. Accelerating research processes by summarizing existing research, generating new ideas, or drafting protocols [90,91]. This supports faster literature reviews and hypothesis generation in fields like pharmaceuticals and engineering.

3. LLMs in Business Operations and Strategy

3.1. Operational Efficiency

LLMs significantly enhance operational efficiency in businesses by automating complex language-based tasks. These AI-driven systems, adept in tasks ranging from customer service automation to sophisticated data analysis, allow businesses to handle customer inquiries, analyze vast amounts of unstructured data for strategic insights, and streamline processes like report generation and compliance monitoring [92,93,94,95,96]. The ability to perform these functions with high accuracy and consistency reduces the workload on human employees. It ensures that businesses can scale operations efficiently, thereby gaining a competitive advantage in their respective markets.

Automated Customer Service.

LLMs are transforming customer service by powering sophisticated chatbots and virtual assistants that offer 24/7 support, enhancing the customer experience and reducing operational costs [97]. For instance, Amazon has effectively utilized these technologies to create chatbots that help customers with everything from tracking orders to managing returns, streamlining the customer service process significantly. Similarly, Bank of America has integrated its virtual assistant “Erica”, which provides personalized banking advice and facilitates transactions autonomously [98,99,100,101]. These LLM applications improve service delivery efficiency and further bolster customer engagement through fast, reliable, and context-sensitive responses, setting a new industry standard for automated customer support.

Data Analysis.

LLMs can help businesses handle and interpret large datasets to refine their operations. These models extract meaningful insights from extensive unstructured data—customer feedback, transaction records, or market trends—enabling companies to make informed decisions quickly [102,103,104,105,106,107]. The application of these insights significantly impacts resource allocation; for example, LLMs can identify operational bottlenecks or underutilized assets, allowing businesses to reallocate resources more efficiently. Moreover, by predicting future trends and customer behaviors, LLMs enable companies to adjust their inventory levels strategically, optimize staffing, and better plan budget allocations [108,109]. This data-driven approach enhances operational efficiency and improves overall business agility and competitive edge in the market.

Process Automation.

LLMs are highly effective in enhancing business process automation and are particularly suited for handling repetitive tasks such as email filtering and report generation [81,110,111]. LLMs efficiently sift through and categorize vast quantities of emails, enabling companies to concentrate on high-priority communications and automatically respond to routine inquiries. In report generation, these models aggregate and synthesize data from multiple sources into coherent summaries, delivering clear and actionable reports that significantly reduce the need for manual labor [112]. Automating these tasks minimizes the demand for manual labor, thus lowering operational costs and increasing the accuracy and speed of business processes.

3.2. Strategic Decision-Making

LLMs significantly enhance strategic decision-making in organizations by leveraging their advanced analytical capabilities [113,114,115,116]. These models process and analyze extensive datasets, uncovering trends and patterns that are not easily detected by humans, which supports more informed and proactive strategic planning [117,118,119,120]. LLMs enable businesses to anticipate market shifts and consumer trends, facilitating a forward-thinking approach to decision-making. They also aid in scenario analysis, allowing executives to assess various strategic options’ implications and potential outcomes. These applications sharpen strategic insights and minimize risks, boosting the company’s agility and alignment with long-term goals [121,122,123].

Predictive Analytics with LLMs.

LLMs excel at leveraging historical data to forecast trends and support proactive decision-making, making them invaluable in dynamic sectors like finance and marketing [124,125]. In finance, LLMs predict market trends, evaluate credit risk, and automate trading strategies by analyzing vast amounts of financial data and market conditions, thus enhancing investment strategies and risk management for institutions [126]. In the realm of marketing, LLMs assess consumer behaviors, predict outcomes of campaigns, and gauge customer engagement, enabling marketers to devise precisely targeted strategies that optimize budget allocation and improve marketing efficacy [127]. This predictive prowess of LLMs helps businesses in these sectors anticipate future trends and strategically plan to maximize outcomes and efficiencies.

Market Trend Analysis.

In terms of market trend analysis, LLMs employ advanced data processing techniques to derive actionable insights from diverse sources such as social media, economic reports, and news articles [128,129]. For instance, by performing sentiment analysis, LLMs assess public opinions on products or brands, while topic modeling helps identify prevalent themes in market discussions. These capabilities have empowered businesses to strategically tailor their operations [130,131]. A notable example includes a retail giant that used LLM-driven insights from social media trends to adjust its inventory ahead of emerging fashion trends, significantly boosting sales by aligning offerings with consumer demands early on [132,133]. Similarly, there are studies utilizing LLMs to monitor regional technology adoption rates and strategically plan their product launches to maximize market penetration effectively.

3.3. Competitive Advantage

Innovation and Product Development.

Startups leveraging LLMs are gaining a competitive edge by accelerating and enhancing their product development cycles [134,135,136]. LLMs enable rapid prototyping and iterative testing by automating significant portions of the development process, from initial design to market readiness. This capability allows startups to bring innovative products to market much faster than traditional methods. For example, companies like Character.ai1 are using LLMs to develop advanced conversational AI, disrupting markets that were traditionally dominated by less interactive technologies. Innovations also include Google’s Gemini2, originally known as Bard, which showcases Google’s commitment to developing generative AI technologies in response to competitive advancements. Similarly, Google’s LaMDA3 represents a significant stride in conversational models, originally introduced as Meena and evolved into a more robust dialogue system. Meta’s Llama4, with its various versions offering different parameter sizes, demonstrates an expansive approach to AI scalability and accessibility. Additionally, ByteDance’s Doubao5 focuses on building industry-leading AI foundation models, highlighting the drive for technological and social progress. Lastly, Microsoft’s Copilot6, based on the Microsoft Prometheus model, integrates directly into consumer software, offering advanced AI functionalities directly from the Windows taskbar. These cases exemplify how LLMs are not only entering established markets but also creating entirely new niches, redefining consumer expectations and engagement.

Advanced Language Models in Industry.

Anthropic’s Claude series demonstrates the rapid evolution in LLM capabilities, with the latest version, Claude 3, introducing image analysis features in March 2024, broadening its utility in multimedia applications 7. Similarly, Google’s PaLM (Pathways Language Model), a 540 billion parameter model, showcases versatility across various tasks including reasoning and code generation, significantly enhanced when combined with chain-of-thought prompting to improve performance on complex datasets 8. These models exemplify how advanced training techniques and expansive parameter scaling are pushing the boundaries of what AI can achieve in practical settings.
Cohere and T5 from Google highlight the industry’s focus on generative AI for both consumer and enterprise applications. Cohere’s platform specializes in embedding large language models into enterprise solutions, enhancing functionalities like chatbots and content generation, and is known for its flexibility across various cloud services 9. Google’s T5 model, exploring transfer learning within a unified text-to-text framework, has set benchmarks across multiple NLP tasks, illustrating the effectiveness of large-scale datasets in training LLMs 10. Meanwhile, Falcon LLM introduces the Falcon Mamba 7B, an open-source State Space Language Model, which highlights advancements in AI architecture that reduce memory costs and enhance performance, marking significant progress in the field 11.

Emerging LLMs Showcasing Advanced Capabilities.

StableLM by Stability AI introduces the StableLM 3B 4E1T, a decoder-only model trained on a diverse dataset including English and code, demonstrating enhanced versatility with features like partial Rotary Position Embeddings and SwiGLU activation for optimized performance 12. NVIDIA’s Megatron-LM focuses on model parallelism to train multi-billion parameter models efficiently, showcasing a methodology that significantly advances NLP applications by enabling the handling of vast models without extensive system overhead [137,138]. These models exemplify the rapid advancements in hardware and training techniques that are pushing the boundaries of what LLMs can achieve.
In the realm of specialized and hybrid models, Jamba from AI21 Labs represents a novel approach by combining Transformer and Mamba architectures, adding mixture-of-experts to enhance model capacity while maintaining efficiency on standard benchmarks [139]. Meanwhile, DeepSeek-v2 develops upon its predecessor by incorporating a Mixture-of-Experts framework, which achieves substantial performance improvements and efficiency in training, proving the efficacy of MoE models in practical applications [140]. These developments not only highlight the innovative architectural evolutions in the field of LLMs but also underscore the ongoing efforts to refine and optimize AI systems for both general and specialized applications.

Customization and Personalization.

Customization and personalization are areas where LLMs significantly impact customer retention and market differentiation [141,142]. By analyzing customer data, LLMs enable companies to tailor experiences and products to individual preferences, enhancing customer satisfaction and loyalty. This personalization extends from customized marketing messages to personalized product recommendations, transforming how businesses interact with their customers [143,144,145]. The ability to deliver these personalized experiences efficiently helps businesses build a competitive advantage, as customers are more likely to return to services that resonate with their unique preferences and needs [146].

Cost Reduction and Scalability.

Furthermore, LLMs contribute to cost reduction and scalability within businesses by automating routine tasks and complex processes. This automation extends across various business functions, including customer service, content generation, and even complex decision-making processes, reducing the need for extensive human intervention [147,148]. As a result, businesses can scale their operations without a corresponding increase in costs. This scalability is crucial for startups and growing businesses that need to manage their resources wisely while expanding their market presence and operational capabilities. Through LLMs, companies not only save on operational costs but also enhance their agility and responsiveness to market dynamics.

4. Challenges and Ethical Considerations

4.1. Data Privacy and Security

In deploying LLMs, data privacy and security are crucial, especially given these systems’ frequent handling of sensitive business and customer data [12,149,150,151,152]. Risks include unauthorized data access, data leakage, and the misuse of personal information, potentially leading to significant financial and reputational damage [153,154]. To mitigate these risks, robust security measures are essential. Encrypting data both at rest and in transit protects sensitive information from interception. Access controls are critical for defining who can interact with the LLM and access specific data, while data anonymization techniques help reduce risks by ensuring personal identifiers are removed from datasets used for training and operating LLMs.
Additionally, newer technologies such as federated learning can enhance data privacy by allowing LLMs to be trained directly on user devices, ensuring that sensitive data does not leave the device [155,156,157]. Implementing advanced machine learning techniques like differential privacy during the training phase can also help in minimizing the risk of data exposure [150,158,159,160,161]. These preventive strategies are vital for maintaining data integrity, building user trust, and ensuring compliance with stringent data protection standards, making them indispensable in the responsible deployment of LLMs [162,163].

4.2. Bias and Fairness

Addressing bias in LLMs is essential for ensuring ethical AI deployment [38,164,165]. Bias in these models can lead to skewed outputs, affecting fairness and accuracy in applications ranging from hiring practices to loan approvals. To counteract this, organizations implement various strategies to detect and mitigate bias. One effective approach is to diversify the data used for training LLMs [166,167,168]. By incorporating a wide array of sources and demographics, models can learn a broader spectrum of language usage and context, reducing the risk of biased outcomes. Additionally, conducting thorough bias audits involves analyzing model decisions across different groups to identify any discrepancies that might disadvantage specific demographics.
Another crucial strategy is inclusive model training, which involves adjusting model parameters and training processes to accommodate and correctly interpret data from underrepresented groups [169,170]. This might include re-weighting training examples or modifying model architectures to improve fairness. Companies like Google and IBM have demonstrated commitment to these practices by developing tools and methodologies that enhance the transparency and fairness of their LLMs [12,171,172]. These efforts not only help in aligning AI technologies with societal norms but also improve customer trust and regulatory compliance, showcasing a proactive approach to managing AI ethics effectively [173].

4.3. Regulatory Compliance

Navigating the complex regulatory landscape for AI and LLMs is crucial for companies aiming to deploy these technologies responsibly. Key international regulations like the General Data Protection Regulation (GDPR) in Europe and the California Consumer Privacy Act (CCPA) set stringent guidelines on data privacy and the ethical use of AI [174,175]. These regulations require that companies handling personal data implement robust measures to protect it and ensure transparency in how AI systems make decisions. Compliance challenges often arise from the varying requirements across different jurisdictions, making it difficult for global companies to maintain a consistent approach to AI governance. To address these challenges, businesses must stay informed about new and evolving laws and adjust their compliance strategies accordingly [176,177].
To ensure adherence to these legal standards, companies should adopt best practices such as maintaining detailed documentation of data handling and AI decision processes, conducting regular audits to assess compliance, and implementing ongoing compliance training for employees. These steps help businesses not only meet regulatory requirements but also build trust with users and stakeholders. For instance, some leading tech companies have shared case studies showing how they navigate these challenges, providing valuable lessons on integrating compliance into their operations without stifling innovation. By proactively addressing regulatory compliance, companies can avoid significant penalties and foster an environment of ethical AI use.

5. Conclusion

Throughout this exploration of LLMs in the context of innovation and entrepreneurship, it’s evident that LLMs are not just technological tools, but transformative agents that redefine how businesses operate and innovate. From accelerating product development to enhancing strategic decision-making and automating routine processes, LLMs offer a range of applications that drive business efficiency and competitiveness. The insights gathered from various sectors underscore the adaptability and impact of LLMs, proving that they are instrumental in both responding to and shaping market demands.
In conclusion, as businesses continue to harness the capabilities of LLMs, the challenges of data privacy, bias, fairness, and regulatory compliance remain critical considerations. Successfully navigating these challenges will be essential for businesses to fully realize the potential of LLMs. The ongoing development and integration of LLMs in business practices not only enhance operational capabilities but also set new standards in the digital economy. The future of LLMs in innovation and entrepreneurship is poised to further evolve, promising even greater advancements and more refined applications in the coming years.

References

  1. Bengio, Y.; Ducharme, R.; Vincent, P. A neural probabilistic language model. Advances in neural information processing systems 2000, 13. [Google Scholar]
  2. Collobert, R.; Weston, J.; Bottou, L.; Karlen, M.; Kavukcuoglu, K.; Kuksa, P. Natural language processing (almost) from scratch. Journal of machine learning research 2011, 12, 2493–2537. [Google Scholar]
  3. Pinker, S. The language instinct: How the mind creates language; Penguin uK, 2003. [Google Scholar]
  4. Hauser, M.D.; Chomsky, N.; Fitch, W.T. The faculty of language: what is it, who has it, and how did it evolve? science 2002, 298, 1569–1579. [Google Scholar] [CrossRef]
  5. Turing, A.M. Computing machinery and intelligence; Springer, 2009. [Google Scholar]
  6. Stolcke, A.; others. SRILM-an extensible language modeling toolkit. Interspeech, 2002, Vol. 2002, p. 2002.
  7. Gao, J.; Lin, C.Y. Introduction to the special issue on statistical language modeling, 2004. [CrossRef]
  8. Rosenfeld, R. Two decades of statistical language modeling: Where do we go from here? Proceedings of the IEEE 2000, 88, 1270–1278. [Google Scholar] [CrossRef]
  9. Liu, X.; Croft, W.B. Statistical language modeling for information retrieval. Annu. Rev. Inf. Sci. Technol. 2005, 39, 1–31. [Google Scholar] [CrossRef]
  10. Thede, S.M.; Harper, M. A second-order hidden Markov model for part-of-speech tagging. Proceedings of the 37th annual meeting of the Association for Computational Linguistics, 1999, pp. 175–182.
  11. Bahl, L.R.; Brown, P.F.; De Souza, P.V.; Mercer, R.L. A tree-based statistical language model for natural language speech recognition. IEEE Transactions on Acoustics, Speech, and Signal Processing 1989, 37, 1001–1008. [Google Scholar] [CrossRef]
  12. Yao, Y.; Duan, J.; Xu, K.; Cai, Y.; Sun, Z.; Zhang, Y. A survey on large language model (llm) security and privacy: The good, the bad, and the ugly. High-Confidence Computing 2024, p. 100211. [CrossRef]
  13. Wang, Z.; Wang, P.; Liu, K.; Wang, P.; Fu, Y.; Lu, C.T.; Aggarwal, C.C.; Pei, J.; Zhou, Y. A Comprehensive Survey on Data Augmentation. arXiv 2024, arXiv:2405.09591. [Google Scholar]
  14. Kalyan, K.S. A survey of GPT-3 family large language models including ChatGPT and GPT-4. Natural Language Processing Journal 2023, 100048. [Google Scholar] [CrossRef]
  15. Zhao, W.X.; Zhou, K.; Li, J.; Tang, T.; Wang, X.; Hou, Y.; Min, Y.; Zhang, B.; Zhang, J.; Dong, Z. A survey of large language models. arXiv others. 2023, arXiv:2303.18223. [Google Scholar]
  16. Chang, Y.; Wang, X.; Wang, J.; Wu, Y.; Yang, L.; Zhu, K.; Chen, H.; Yi, X.; Wang, C.; Wang, Y.; others. A survey on evaluation of large language models. ACM Transactions on Intelligent Systems and Technology 2024, 15, 1–45. [Google Scholar] [CrossRef]
  17. Zhang, X.; Zhang, J.; Rekabdar, B.; Zhou, Y.; Wang, P.; Liu, K. Dynamic and Adaptive Feature Generation with LLM. arXiv 2024, arXiv:2406.03505. [Google Scholar]
  18. Minaee, S.; Mikolov, T.; Nikzad, N.; Chenaghlu, M.; Socher, R.; Amatriain, X.; Gao, J. Large language models: A survey. arXiv 2024, arXiv:2402.06196. [Google Scholar]
  19. Zhao, H.; Chen, H.; Yang, F.; Liu, N.; Deng, H.; Cai, H.; Wang, S.; Yin, D.; Du, M. Explainability for large language models: A survey. ACM Transactions on Intelligent Systems and Technology 2024, 15, 1–38. [Google Scholar] [CrossRef]
  20. Vaswani, A. Attention is all you need. Advances in Neural Information Processing Systems 2017. [Google Scholar]
  21. Huang, K.; Mo, F.; Li, H.; Li, Y.; Zhang, Y.; Yi, W.; Mao, Y.; Liu, J.; Xu, Y.; Xu, J.; others. A Survey on Large Language Models with Multilingualism: Recent Advances and New Frontiers. arXiv 2024, arXiv:2405.10936. [Google Scholar]
  22. Su, Z.; Zhou, Y.; Mo, F.; Simonsen, J.G. Language modeling using tensor trains. arXiv 2024, arXiv:2405.04590. [Google Scholar]
  23. Wang, C.; Li, M.; Smola, A.J. Language models with transformers. arXiv 2019, arXiv:1904.09408. [Google Scholar]
  24. Sun, Y.; Dong, L.; Huang, S.; Ma, S.; Xia, Y.; Xue, J.; Wang, J.; Wei, F. Retentive network: A successor to transformer for large language models. arXiv 2023, arXiv:2307.08621. [Google Scholar]
  25. Wang, J.; Mo, F.; Ma, W.; Sun, P.; Zhang, M.; Nie, J.Y. A User-Centric Benchmark for Evaluating Large Language Models. arXiv 2024, arXiv:2404.13940. [Google Scholar]
  26. Mao, K.; Dou, Z.; Mo, F.; Hou, J.; Chen, H.; Qian, H. Large language models know your contextual search intent: A prompting framework for conversational search. arXiv 2023, arXiv:2303.06573. [Google Scholar]
  27. Dai, Z.; Yang, Z.; Yang, Y.; Carbonell, J.; Le, Q.V.; Salakhutdinov, R. Transformer-xl: Attentive language models beyond a fixed-length context. arXiv 2019, arXiv:1901.02860. [Google Scholar]
  28. Thirunavukarasu, A.J.; Ting, D.S.J.; Elangovan, K.; Gutierrez, L.; Tan, T.F.; Ting, D.S.W. Large language models in medicine. Nature medicine 2023, 29, 1930–1940. [Google Scholar] [CrossRef] [PubMed]
  29. Meyer, J.G.; Urbanowicz, R.J.; Martin, P.C.; O’Connor, K.; Li, R.; Peng, P.C.; Bright, T.J.; Tatonetti, N.; Won, K.J.; Gonzalez-Hernandez, G.; others. ChatGPT and large language models in academia: opportunities and challenges. BioData Mining 2023, 16, 20. [Google Scholar] [CrossRef]
  30. Goldberg, A. AI in Finance: Leveraging Large Language Models for Enhanced Decision-Making and Risk Management. Social Science Journal for Advanced Research 2024, 4, 33–40. [Google Scholar]
  31. Li, H.; Gao, H.; Wu, C.; Vasarhelyi, M.A. Extracting financial data from unstructured sources: Leveraging large language models. Journal of Information Systems 2023, 1–22. [Google Scholar] [CrossRef]
  32. Birhane, A.; Kasirzadeh, A.; Leslie, D.; Wachter, S. Science in the age of large language models. Nature Reviews Physics 2023, 5, 277–280. [Google Scholar] [CrossRef]
  33. Liu, Y.; Han, T.; Ma, S.; Zhang, J.; Yang, Y.; Tian, J.; He, H.; Li, A.; He, M.; Liu, Z.; others. Summary of chatgpt-related research and perspective towards the future of large language models. Meta-Radiology 2023, 100017. [Google Scholar] [CrossRef]
  34. Zhang, X.; Wang, Z.; Jiang, L.; Gao, W.; Wang, P.; Liu, K. TFWT: Tabular Feature Weighting with Transformer. arXiv 2024, arXiv:2405.08403. [Google Scholar]
  35. Abd-Alrazaq, A.; AlSaad, R.; Alhuwail, D.; Ahmed, A.; Healy, P.M.; Latifi, S.; Aziz, S.; Damseh, R.; Alrazak, S.A.; Sheikh, J.; others. Large language models in medical education: opportunities, challenges, and future directions. JMIR Medical Education 2023, 9, e48291. [Google Scholar] [CrossRef]
  36. Zeng, J.; Chen, B.; Deng, Y.; Chen, W.; Mao, Y.; Li, J. Fine-tuning of financial Large Language Model and application at edge device. Proceedings of the 3rd International Conference on Computer, Artificial Intelligence and Control Engineering, 2024, pp. 42–47.
  37. Zhao, H.; Liu, Z.; Wu, Z.; Li, Y.; Yang, T.; Shu, P.; Xu, S.; Dai, H.; Zhao, L.; Mai, G.; others. Revolutionizing finance with llms: An overview of applications and insights. arXiv 2024, arXiv:2401.11641. [Google Scholar]
  38. Hadi, M.U.; Al Tashi, Q.; Shah, A.; Qureshi, R.; Muneer, A.; Irfan, M.; Zafar, A.; Shaikh, M.B.; Akhtar, N.; Wu, J.; others. Large language models: a comprehensive survey of its applications, challenges, limitations, and future prospects. Authorea Preprints 2024.
  39. Li, X.V.; Passino, F.S. FinDKG: Dynamic Knowledge Graphs with Large Language Models for Detecting Global Trends in Financial Markets. arXiv 2024, arXiv:2407.10909. [Google Scholar]
  40. Carlini, N.; Tramer, F.; Wallace, E.; Jagielski, M.; Herbert-Voss, A.; Lee, K.; Roberts, A.; Brown, T.; Song, D.; Erlingsson, U.; others. Extracting training data from large language models. 30th USENIX Security Symposium (USENIX Security 21), 2021, pp. 2633–2650.
  41. Naveed, H.; Khan, A.U.; Qiu, S.; Saqib, M.; Anwar, S.; Usman, M.; Akhtar, N.; Barnes, N.; Mian, A. A comprehensive overview of large language models. arXiv 2023, arXiv:2307.06435. [Google Scholar]
  42. Shi, W.; Ajith, A.; Xia, M.; Huang, Y.; Liu, D.; Blevins, T.; Chen, D.; Zettlemoyer, L. Detecting pretraining data from large language models. arXiv 2023, arXiv:2310.16789. [Google Scholar]
  43. Korbak, T.; Shi, K.; Chen, A.; Bhalerao, R.V.; Buckley, C.; Phang, J.; Bowman, S.R.; Perez, E. Pretraining language models with human preferences. International Conference on Machine Learning. PMLR, 2023, pp. 17506–17533.
  44. Gururangan, S.; Marasović, A.; Swayamdipta, S.; Lo, K.; Beltagy, I.; Downey, D.; Smith, N.A. Don’t stop pretraining: Adapt language models to domains and tasks. arXiv 2020, arXiv:2004.10964. [Google Scholar]
  45. Petroni, F.; Rocktäschel, T.; Lewis, P.; Bakhtin, A.; Wu, Y.; Miller, A.H.; Riedel, S. Language models as knowledge bases? arXiv 2019, arXiv:1909.01066. [Google Scholar]
  46. Bert, C.W.; Malik, M. Differential quadrature method in computational mechanics: a review 1996.
  47. Chowdhery, A.; Narang, S.; Devlin, J.; Bosma, M.; Mishra, G.; Roberts, A.; Barham, P.; Chung, H.W.; Sutton, C.; Gehrmann, S.; others. Palm: Scaling language modeling with pathways. Journal of Machine Learning Research 2023, 24, 1–113. [Google Scholar] [CrossRef]
  48. Zhang, S.; Dong, L.; Li, X.; Zhang, S.; Sun, X.; Wang, S.; Li, J.; Hu, R.; Zhang, T.; Wu, F.; others. Instruction tuning for large language models: A survey. arXiv 2023, arXiv:2308.10792. [Google Scholar]
  49. Wang, Y.; Zhong, W.; Li, L.; Mi, F.; Zeng, X.; Huang, W.; Shang, L.; Jiang, X.; Liu, Q. Aligning large language models with human: A survey. arXiv 2023, arXiv:2307.12966. [Google Scholar]
  50. Zhang, J.; Wang, X.; Jin, Y.; Chen, C.; Zhang, X.; Liu, K. Prototypical Reward Network for Data-Efficient RLHF. arXiv 2024, arXiv:2406.06606. [Google Scholar]
  51. Min, B.; Ross, H.; Sulem, E.; Veyseh, A.P.B.; Nguyen, T.H.; Sainz, O.; Agirre, E.; Heintz, I.; Roth, D. Recent advances in natural language processing via large pre-trained language models: A survey. ACM Computing Surveys 2023, 56, 1–40. [Google Scholar] [CrossRef]
  52. Mahowald, K.; Ivanova, A.A.; Blank, I.A.; Kanwisher, N.; Tenenbaum, J.B.; Fedorenko, E. Dissociating language and thought in large language models. Trends in Cognitive Sciences 2024. [Google Scholar] [CrossRef]
  53. Meier, J.; Rao, R.; Verkuil, R.; Liu, J.; Sercu, T.; Rives, A. Language models enable zero-shot prediction of the effects of mutations on protein function. Advances in neural information processing systems 2021, 34, 29287–29303. [Google Scholar]
  54. Madani, A.; Krause, B.; Greene, E.R.; Subramanian, S.; Mohr, B.P.; Holton, J.M.; Olmos, J.L.; Xiong, C.; Sun, Z.Z.; Socher, R.; others. Large language models generate functional protein sequences across diverse families. Nature Biotechnology 2023, 41, 1099–1106. [Google Scholar] [CrossRef]
  55. Rafailov, R.; Sharma, A.; Mitchell, E.; Manning, C.D.; Ermon, S.; Finn, C. Direct preference optimization: Your language model is secretly a reward model. Advances in Neural Information Processing Systems 2024, 36. [Google Scholar]
  56. OpenAI. GPT-3.5 Turbo Fine-Tuning and API Updates. https://openai.com/index/gpt-3-5-turbo-fine-tuning-and-api-updates/, 2023. Accessed: 2024-05-20.
  57. Huang, W.; Abbeel, P.; Pathak, D.; Mordatch, I. Language models as zero-shot planners: Extracting actionable knowledge for embodied agents. International Conference on Machine Learning. PMLR, 2022, pp. 9118–9147.
  58. Zhang, S.; Chen, Z.; Shen, Y.; Ding, M.; Tenenbaum, J.B.; Gan, C. Planning with large language models for code generation. arXiv 2023, arXiv:2303.05510. [Google Scholar]
  59. Zhang, J.; Wang, X.; Ren, W.; Jiang, L.; Wang, D.; Liu, K. RATT: AThought Structure for Coherent and Correct LLMReasoning. arXiv 2024, arXiv:2406.02746. [Google Scholar]
  60. Zhao, P.; Zhang, H.; Yu, Q.; Wang, Z.; Geng, Y.; Fu, F.; Yang, L.; Zhang, W.; Cui, B. Retrieval-augmented generation for ai-generated content: A survey. arXiv 2024, arXiv:2402.19473. [Google Scholar]
  61. Renze, M.; Guven, E. Self-Reflection in LLM Agents: Effects on Problem-Solving Performance. arXiv 2024, arXiv:2405.06682. [Google Scholar]
  62. Noor, A.K.; Burton, W.S.; Bert, C.W. Computational models for sandwich panels and shells 1996. [CrossRef]
  63. Adoma, A.F.; Henry, N.M.; Chen, W. Comparative analyses of bert, roberta, distilbert, and xlnet for text-based emotion recognition. 2020 17th International Computer Conference on Wavelet Active Media Technology and Information Processing (ICCWAMTIP). IEEE, 2020, pp. 117–121.
  64. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of machine learning research 2020, 21, 1–67. [Google Scholar]
  65. Guo, M.; Ainslie, J.; Uthus, D.; Ontanon, S.; Ni, J.; Sung, Y.H.; Yang, Y. LongT5: Efficient text-to-text transformer for long sequences. arXiv 2021, arXiv:2112.07916. [Google Scholar]
  66. Mastropaolo, A.; Scalabrino, S.; Cooper, N.; Palacio, D.N.; Poshyvanyk, D.; Oliveto, R.; Bavota, G. Studying the usage of text-to-text transfer transformer to support code-related tasks. 2021 IEEE/ACM 43rd International Conference on Software Engineering (ICSE). IEEE, 2021, pp. 336–347.
  67. Li, Z.; Guan, B.; Wei, Y.; Zhou, Y.; Zhang, J.; Xu, J. Mapping new realities: Ground truth image creation with pix2pix image-to-image translation. arXiv 2024, arXiv:2404.19265. [Google Scholar]
  68. Zhang, B.; Yang, H.; Zhou, T.; Ali Babar, M.; Liu, X.Y. Enhancing financial sentiment analysis via retrieval augmented large language models. Proceedings of the fourth ACM international conference on AI in finance, 2023, pp. 349–356.
  69. Ding, H.; Li, Y.; Wang, J.; Chen, H. Large Language Model Agent in Financial Trading: A Survey. arXiv 2024, arXiv:2408.06361. [Google Scholar]
  70. Xue, L. mt5: A massively multilingual pre-trained text-to-text transformer. arXiv 2020, arXiv:2010.11934. [Google Scholar] [CrossRef]
  71. Wang, L.; Zhang, X.; Li, K.; Zhang, S. Semi-supervised learning for k-dependence Bayesian classifiers. Applied Intelligence 2022, 1–19. [Google Scholar] [CrossRef]
  72. Wang, L.; Zhang, S.; Mammadov, M.; Li, K.; Zhang, X.; Wu, S. Semi-supervised weighting for averaged one-dependence estimators. Applied Intelligence 2022, 1–17. [Google Scholar] [CrossRef]
  73. Yuan, Y.; Huang, Y.; Ma, Y.; Li, X.; Li, Z.; Shi, Y.; Zhou, H. Rhyme-aware Chinese lyric generator based on GPT. arXiv 2024, arXiv:2408.10130. [Google Scholar]
  74. Zhang, J.; Cao, J.; Chang, J.; Li, X.; Liu, H.; Li, Z. Research on the Application of Computer Vision Based on Deep Learning in Autonomous Driving Technology. arXiv 2024, arXiv:2406.00490. [Google Scholar]
  75. Xie, Q.; Han, W.; Zhang, X.; Lai, Y.; Peng, M.; Lopez-Lira, A.; Huang, J. Pixiu: A large language model, instruction data and evaluation benchmark for finance. arXiv 2023, arXiv:2306.05443. [Google Scholar]
  76. Kang, H.; Liu, X.Y. Deficiency of large language models in finance: An empirical examination of hallucination. I Can’t Believe It’s Not Better Workshop: Failure Modes in the Age of Foundation Models, 2023.
  77. Robles-Serrano, S.; Rios-Perez, J.; Sanchez-Torres, G. Integration of Large Language Models in Mobile Applications for Statutory Auditing and Finance. Prospectiva (1692-8261) 2024, 22. [Google Scholar]
  78. Xie, Q.; Huang, J.; Li, D.; Chen, Z.; Xiang, R.; Xiao, M.; Yu, Y.; Somasundaram, V.; Yang, K.; Yuan, C.; others. FinNLP-AgentScen-2024 Shared Task: Financial Challenges in Large Language Models-FinLLMs. Proceedings of the Eighth Financial Technology and Natural Language Processing and the 1st Agent AI for Scenario Planning, 2024, pp. 119–126.
  79. Wu, S.; Irsoy, O.; Lu, S.; Dabravolski, V.; Dredze, M.; Gehrmann, S.; Kambadur, P.; Rosenberg, D.; Mann, G. Bloomberggpt: A large language model for finance. arXiv 2023, arXiv:2303.17564. [Google Scholar]
  80. Li, Y.; Wang, S.; Ding, H.; Chen, H. Large language models in finance: A survey. Proceedings of the fourth ACM international conference on AI in finance, 2023, pp. 374–382.
  81. Yang, Y.; Uy, M.C.S.; Huang, A. Finbert: A pretrained language model for financial communications. arXiv 2020, arXiv:2006.08097. [Google Scholar]
  82. Huang, A.H.; Wang, H.; Yang, Y. FinBERT: A large language model for extracting information from financial text. Contemporary Accounting Research 2023, 40, 806–841. [Google Scholar] [CrossRef]
  83. Park, T. Enhancing Anomaly Detection in Financial Markets with an LLM-based Multi-Agent Framework. arXiv 2024, arXiv:2403.19735. [Google Scholar]
  84. Kim, D.; Lee, D.; Park, J.; Oh, S.; Kwon, S.; Lee, I.; Choi, D. KB-BERT: Training and Application of Korean Pre-trained Language Model in Financial Domain. Journal of Intelligence and Information Systems 2022, 28, 191–206. [Google Scholar]
  85. Krause, D. Large language models and generative AI in finance: an analysis of ChatGPT, Bard, and Bing AI. Bard, and Bing AI (July 15, 2023) 2023.
  86. Liu, X.Y.; Wang, G.; Yang, H.; Zha, D. Fingpt: Democratizing internet-scale data for financial large language models. arXiv 2023, arXiv:2307.10485. [Google Scholar]
  87. Mo, F.; Qu, C.; Mao, K.; Zhu, T.; Su, Z.; Huang, K.; Nie, J.Y. History-Aware Conversational Dense Retrieval. arXiv 2024, arXiv:2401.16659. [Google Scholar]
  88. Zhang, C.; Liu, X.; Jin, M.; Zhang, Z.; Li, L.; Wang, Z.; Hua, W.; Shu, D.; Zhu, S.; Jin, X.; others. When AI Meets Finance (StockAgent): Large Language Model-based Stock Trading in Simulated Real-world Environments. arXiv 2024, arXiv:2407.18957. [Google Scholar]
  89. Krumdick, M.; Koncel-Kedziorski, R.; Lai, V.; Reddy, V.; Lovering, C.; Tanner, C. BizBench: A Quantitative Reasoning Benchmark for Business and Finance. Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2024, pp. 8309–8332.
  90. Wang, L.; Ma, C.; Feng, X.; Zhang, Z.; Yang, H.; Zhang, J.; Chen, Z.; Tang, J.; Chen, X.; Lin, Y.; others. A survey on large language model based autonomous agents. Frontiers of Computer Science 2024, 18, 186345. [Google Scholar] [CrossRef]
  91. Peng, C.; Yang, X.; Chen, A.; Smith, K.E.; PourNejatian, N.; Costa, A.B.; Martin, C.; Flores, M.G.; Zhang, Y.; Magoc, T.; others. A study of generative large language model for medical research and healthcare. NPJ digital medicine 2023, 6, 210. [Google Scholar] [CrossRef]
  92. Obschonka, M.; Audretsch, D.B. Artificial intelligence and big data in entrepreneurship: a new era has begun. Small Business Economics 2020, 55, 529–539. [Google Scholar] [CrossRef]
  93. Weber, M.; Beutter, M.; Weking, J.; Böhm, M.; Krcmar, H. AI startup business models: Key characteristics and directions for entrepreneurship research. Business & Information Systems Engineering 2022, 64, 91–109. [Google Scholar]
  94. Hu, T.; Zhu, W.; Yan, Y. Artificial intelligence aspect of transportation analysis using large scale systems. Proceedings of the 2023 6th Artificial Intelligence and Cloud Computing Conference, 2023, pp. 54–59.
  95. Zhu, W. Optimizing distributed networking with big data scheduling and cloud computing. International Conference on Cloud Computing, Internet of Things, and Computer Applications (CICA 2022). SPIE, 2022, Vol. 12303, pp. 23–28.
  96. Yan, Y. Influencing Factors of Housing Price in New York-analysis: Based on Excel Multi-regression Model 2022.
  97. Khang, A.; Jadhav, B.; Dave, T. Enhancing Financial Services. Synergy of AI and Fintech in the Digital Gig Economy 2024, p. 147.
  98. Crosman, P. Banking 2025: The rise of the invisible bank. American Banker 2019. [Google Scholar]
  99. Wendlinger, B. The challenge of FinTech from the perspective of german incumbent banks: an exploratory study investigating industry trends and considering the future of banking. PhD thesis, 2022.
  100. Kang, Y.; Xu, Y.; Chen, C.P.; Li, G.; Cheng, Z. 6: Simultaneous Tracking, Tagging and Mapping for Augmented Reality. SID Symposium Digest of Technical Papers. Wiley Online Library, 2021, Vol. 52, pp. 31–33. [CrossRef]
  101. Zhu, W.; Hu, T. Twitter Sentiment analysis of covid vaccines. 2021 5th International Conference on Artificial Intelligence and Virtual Reality (AIVR), 2021, pp. 118–122.
  102. Zhang, X.; Zhang, J.; Mo, F.; Chen, Y.; Liu, K. TIFG: Text-Informed Feature Generation with Large Language Models. arXiv 2024, arXiv:2406.11177. [Google Scholar]
  103. Mo, F.; Mao, K.; Zhu, Y.; Wu, Y.; Huang, K.; Nie, J.Y. ConvGQR: generative query reformulation for conversational search. arXiv 2023, arXiv:2305.15645. [Google Scholar]
  104. Mao, K.; Deng, C.; Chen, H.; Mo, F.; Liu, Z.; Sakai, T.; Dou, Z. ChatRetriever: Adapting Large Language Models for Generalized and Robust Conversational Dense Retrieval. arXiv 2024, arXiv:2404.13556. [Google Scholar]
  105. Zhu, Y.; Honnet, C.; Kang, Y.; Zhu, J.; Zheng, A.J.; Heinz, K.; Tang, G.; Musk, L.; Wessely, M.; Mueller, S. Demonstration of ChromoCloth: Re-Programmable Multi-Color Textures through Flexible and Portable Light Source. Adjunct Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology, 2023, pp. 1–3.
  106. Yang, X.; Kang, Y.; Yang, X. Retargeting destinations of passive props for enhancing haptic feedback in virtual reality. 2022 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW). IEEE, 2022, pp. 618–619.
  107. Kang, Y.; Zhang, Z.; Zhao, M.; Yang, X.; Yang, X. Tie Memories to E-souvenirs: Hybrid Tangible AR Souvenirs in the Museum. Adjunct Proceedings of the 35th Annual ACM Symposium on User Interface Software and Technology, 2022, pp. 1–3.
  108. Mo, F.; Nie, J.Y.; Huang, K.; Mao, K.; Zhu, Y.; Li, P.; Liu, Y. Learning to relate to previous turns in conversational search. Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2023, pp. 1722–1732.
  109. Mo, F.; Yi, B.; Mao, K.; Qu, C.; Huang, K.; Nie, J.Y. Convsdg: Session data generation for conversational search. Companion Proceedings of the ACM on Web Conference 2024, 2024, pp. 1634–1642. [Google Scholar]
  110. Grohs, M.; Abb, L.; Elsayed, N.; Rehse, J.R. Large language models can accomplish business process management tasks. International Conference on Business Process Management. Springer, 2023, pp. 453–465.
  111. Bellan, P.; Dragoni, M.; Ghidini, C. Extracting business process entities and relations from text using pre-trained language models and in-context learning. International Conference on Enterprise Design, Operations, and Computing. Springer, 2022, pp. 182–199.
  112. George, A.S.; George, A.H. A review of ChatGPT AI’s impact on several business sectors. Partners universal international innovation journal 2023, 1, 9–23. [Google Scholar]
  113. Sha, H.; Mu, Y.; Jiang, Y.; Chen, L.; Xu, C.; Luo, P.; Li, S.E.; Tomizuka, M.; Zhan, W.; Ding, M. Languagempc: Large language models as decision makers for autonomous driving. arXiv 2023, arXiv:2310.03026. [Google Scholar]
  114. Li, S.; Puig, X.; Paxton, C.; Du, Y.; Wang, C.; Fan, L.; Chen, T.; Huang, D.A.; Akyürek, E.; Anandkumar, A.; others. Pre-trained language models for interactive decision-making. Advances in Neural Information Processing Systems 2022, 35, 31199–31212. [Google Scholar]
  115. Song, Y.; Arora, P.; Singh, R.; Varadharajan, S.T.; Haynes, M.; Starner, T. Going Blank Comfortably: Positioning Monocular Head-Worn Displays When They are Inactive. Proceedings of the 2023 International Symposium on Wearable Computers; ACM: Cancun, Quintana Roo Mexico, 2023; pp. 114–118. [Google Scholar] [CrossRef]
  116. Ni, H.; Meng, S.; Chen, X.; Zhao, Z.; Chen, A.; Li, P.; Zhang, S.; Yin, Q.; Wang, Y.; Chan, Y. Harnessing Earnings Reports for Stock Predictions: A QLoRA-Enhanced LLM Approach. arXiv 2024, arXiv:2408.06634. [Google Scholar]
  117. Chen, L.; Zhang, Y.; Ren, S.; Zhao, H.; Cai, Z.; Wang, Y.; Wang, P.; Liu, T.; Chang, B. Towards end-to-end embodied decision making via multi-modal large language model: Explorations with gpt4-vision and beyond. arXiv 2023, arXiv:2310.02071. [Google Scholar]
  118. Zhou, Y.; Zeng, Z.; Chen, A.; Zhou, X.; Ni, H.; Zhang, S.; Li, P.; Liu, L.; Zheng, M.; Chen, X. Evaluating Modern Approaches in 3D Scene Reconstruction: NeRF vs Gaussian-Based Methods. arXiv 2024, arXiv:2408.04268. [Google Scholar]
  119. Yao, S.; Yu, D.; Zhao, J.; Shafran, I.; Griffiths, T.; Cao, Y.; Narasimhan, K. Tree of thoughts: Deliberate problem solving with large language models. Advances in Neural Information Processing Systems 2024, 36. [Google Scholar]
  120. Rao, A.; Kim, J.; Kamineni, M.; Pang, M.; Lie, W.; Succi, M.D. Evaluating ChatGPT as an adjunct for radiologic decision-making. MedRxiv 2023, pp. 2023–02.
  121. Shen, Y.; Heacock, L.; Elias, J.; Hentel, K.D.; Reig, B.; Shih, G.; Moy, L. ChatGPT and other large language models are double-edged swords, 2023. [CrossRef]
  122. Song, Y.; Arora, P.; Varadharajan, S.T.; Singh, R.; Haynes, M.; Starner, T. Looking From a Different Angle: Placing Head-Worn Displays Near the Nose. Proceedings of the Augmented Humans International Conference 2024; ACM: Melbourne VIC Australia, 2024; pp. 28–45. [Google Scholar] [CrossRef]
  123. Yao, S.; Zhao, J.; Yu, D.; Du, N.; Shafran, I.; Narasimhan, K.; Cao, Y. React: Synergizing reasoning and acting in language models. arXiv 2022, arXiv:2210.03629. [Google Scholar]
  124. Chowdhury, R.; Bouatta, N.; Biswas, S.; Floristean, C.; Kharkar, A.; Roy, K.; Rochereau, C.; Ahdritz, G.; Zhang, J.; Church, G.M.; others. Single-sequence protein structure prediction using a language model and deep learning. Nature Biotechnology 2022, 40, 1617–1623. [Google Scholar] [CrossRef]
  125. Christian, H.; Suhartono, D.; Chowanda, A.; Zamli, K.Z. Text based personality prediction from multiple social media data sources using pre-trained language model and model averaging. Journal of Big Data 2021, 8, 68. [Google Scholar] [CrossRef]
  126. Lin, Z.; Akin, H.; Rao, R.; Hie, B.; Zhu, Z.; Lu, W.; dos Santos Costa, A.; Fazel-Zarandi, M.; Sercu, T.; Candido, S.; others. Language models of protein sequences at the scale of evolution enable accurate structure prediction. BioRxiv 2022, 2022, 500902. [Google Scholar]
  127. Guu, K.; Lee, K.; Tung, Z.; Pasupat, P.; Chang, M. Retrieval augmented language model pre-training. International conference on machine learning. PMLR, 2020, pp. 3929–3938.
  128. Eloundou, T.; Manning, S.; Mishkin, P.; Rock, D. Gpts are gpts: An early look at the labor market impact potential of large language models. arXiv 2023, arXiv:2303.10130. [Google Scholar]
  129. Horton, J.J. Large language models as simulated economic agents: What can we learn from homo silicus? Technical report, National Bureau of Economic Research, 2023.
  130. Zhang, H.; Hua, F.; Xu, C.; Kong, H.; Zuo, R.; Guo, J. Unveiling the Potential of Sentiment: Can Large Language Models Predict Chinese Stock Price Movements? arXiv 2023, arXiv:2306.14222. [Google Scholar]
  131. Tang, X.; Lei, N.; Dong, M.; Ma, D. Stock Price Prediction Based on Natural Language Processing1. Complexity 2022, 2022, 9031900. [Google Scholar] [CrossRef]
  132. Muhammad, T.; Aftab, A.B.; Ibrahim, M.; Ahsan, M.M.; Muhu, M.M.; Khan, S.I.; Alam, M.S. Transformer-based deep learning model for stock price prediction: A case study on Bangladesh stock market. International Journal of Computational Intelligence and Applications 2023, 22, 2350013. [Google Scholar] [CrossRef]
  133. Fan, X.; Tao, C.; Zhao, J. Advanced Stock Price Prediction with xLSTM-Based Models: Improving Long-Term Forecasting. Preprints 2024. [Google Scholar]
  134. Yan, C. Predict lightning location and movement with atmospherical electrical field instrument. 2019 IEEE 10th Annual Information Technology, Electronics and Mobile Communication Conference (IEMCON). IEEE, 2019, pp. 0535–0537.
  135. Gu, W.; Zhong, Y.; Li, S.; Wei, C.; Dong, L.; Wang, Z.; Yan, C. Predicting stock prices with FinBERT-LSTM: Integrating news sentiment analysis. arXiv 2024, arXiv:2407.16150. [Google Scholar]
  136. Xing, Y.; Yan, C.; Xie, C.C. Predicting NVIDIA’s Next-Day Stock Price: A Comparative Analysis of LSTM, MLP, ARIMA, and ARIMA-GARCH Models. arXiv 2024, arXiv:2405.08284. [Google Scholar]
  137. Shoeybi, M.; Patwary, M.; Puri, R.; LeGresley, P.; Casper, J.; Catanzaro, B. Megatron-lm: Training multi-billion parameter language models using model parallelism. arXiv 2019, arXiv:1909.08053. [Google Scholar]
  138. Fan, X.; Tao, C. Towards Resilient and Efficient LLMs: A Comparative Study of Efficiency, Performance, and Adversarial Robustness. arXiv 2024, arXiv:2408.04585. [Google Scholar]
  139. Lieber, O.; Lenz, B.; Bata, H.; Cohen, G.; Osin, J.; Dalmedigos, I.; Safahi, E.; Meirom, S.; Belinkov, Y.; Shalev-Shwartz, S.; others. Jamba: A hybrid transformer-mamba language model. arXiv 2024, arXiv:2403.19887. [Google Scholar]
  140. Liu, A.; Feng, B.; Wang, B.; Wang, B.; Liu, B.; Zhao, C.; Dengr, C.; Ruan, C.; Dai, D.; Guo, D.; others. DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model. arXiv 2024, arXiv:2405.04434. [Google Scholar]
  141. Chen, J.; Liu, Z.; Huang, X.; Wu, C.; Liu, Q.; Jiang, G.; Pu, Y.; Lei, Y.; Chen, X.; Wang, X.; others. When large language models meet personalization: Perspectives of challenges and opportunities. World Wide Web 2024, 27, 42. [Google Scholar] [CrossRef]
  142. Wu, J.; Antonova, R.; Kan, A.; Lepert, M.; Zeng, A.; Song, S.; Bohg, J.; Rusinkiewicz, S.; Funkhouser, T. Tidybot: Personalized robot assistance with large language models. Autonomous Robots 2023, 47, 1087–1102. [Google Scholar] [CrossRef]
  143. Salemi, A.; Mysore, S.; Bendersky, M.; Zamani, H. Lamp: When large language models meet personalization. arXiv 2023, arXiv:2304.11406. [Google Scholar]
  144. Geng, S.; Liu, S.; Fu, Z.; Ge, Y.; Zhang, Y. Recommendation as language processing (rlp): A unified pretrain, personalized prompt & predict paradigm (p5). Proceedings of the 16th ACM Conference on Recommender Systems, 2022, pp. 299–315.
  145. Li, L.; Zhang, Y.; Chen, L. Personalized transformer for explainable recommendation. arXiv 2021, arXiv:2105.11601. [Google Scholar]
  146. Zheng, Y.; Zhang, R.; Huang, M.; Mao, X. A pre-training based personalized dialogue generation model with persona-sparse data. Proceedings of the AAAI Conference on Artificial Intelligence, 2020, Vol. 34, pp. 9693–9700. [CrossRef]
  147. Nasr, M.; Carlini, N.; Hayase, J.; Jagielski, M.; Cooper, A.F.; Ippolito, D.; Choquette-Choo, C.A.; Wallace, E.; Tramèr, F.; Lee, K. Scalable extraction of training data from (production) language models. arXiv 2023, arXiv:2311.17035. [Google Scholar]
  148. Choromanski, K.; Likhosherstov, V.; Dohan, D.; Song, X.; Gane, A.; Sarlos, T.; Hawkins, P.; Davis, J.; Belanger, D.; Colwell, L.; others. Masked language modeling for proteins via linearly scalable long-context transformers. arXiv 2020, arXiv:2006.03555. [Google Scholar]
  149. Peris, C.; Dupuy, C.; Majmudar, J.; Parikh, R.; Smaili, S.; Zemel, R.; Gupta, R. Privacy in the time of language models. Proceedings of the sixteenth ACM international conference on web search and data mining, 2023, pp. 1291–1292.
  150. Pan, X.; Zhang, M.; Ji, S.; Yang, M. Privacy risks of general-purpose language models. 2020 IEEE Symposium on Security and Privacy (SP). IEEE, 2020, pp. 1314–1331.
  151. Huang, Y.; Song, Z.; Chen, D.; Li, K.; Arora, S. TextHide: Tackling data privacy in language understanding tasks. Findings of the Association for Computational Linguistics: EMNLP 2020, 2020, pp. 1368–1382. [Google Scholar]
  152. Kandpal, N.; Wallace, E.; Raffel, C. Deduplicating training data mitigates privacy risks in language models. International Conference on Machine Learning. PMLR, 2022, pp. 10697–10707.
  153. Das, B.C.; Amini, M.H.; Wu, Y. Security and privacy challenges of large language models: A survey. arXiv 2024, arXiv:2402.00888. [Google Scholar]
  154. Shi, W.; Cui, A.; Li, E.; Jia, R.; Yu, Z. Selective differential privacy for language modeling. arXiv 2021, arXiv:2108.12944. [Google Scholar]
  155. Raeini, M. Privacy-preserving large language models (PPLLMs). Available at SSRN 4512071 2023. [Google Scholar] [CrossRef]
  156. Sommestad, T.; Ekstedt, M.; Holm, H. The cyber security modeling language: A tool for assessing the vulnerability of enterprise system architectures. IEEE Systems Journal 2012, 7, 363–373. [Google Scholar] [CrossRef]
  157. Geren, C.; Board, A.; Dagher, G.G.; Andersen, T.; Zhuang, J. Blockchain for large language model security and safety: A holistic survey. arXiv 2024, arXiv:2407.20181. [Google Scholar]
  158. Chen, H.; Mo, F.; Wang, Y.; Chen, C.; Nie, J.Y.; Wang, C.; Cui, J. A customized text sanitization mechanism with differential privacy. arXiv 2022, arXiv:2207.01193. [Google Scholar]
  159. Huang, K.; Huang, D.; Liu, Z.; Mo, F. A joint multiple criteria model in transfer learning for cross-domain chinese word segmentation. Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP), 2020, pp. 3873–3882.
  160. Mao, K.; Qian, H.; Mo, F.; Dou, Z.; Liu, B.; Cheng, X.; Cao, Z. Learning denoised and interpretable session representation for conversational search. Proceedings of the ACM Web Conference 2023, 2023, pp. 3193–3202. [Google Scholar]
  161. Huang, K.; Xiao, K.; Mo, F.; Jin, B.; Liu, Z.; Huang, D. Domain-aware word segmentation for Chinese language: A document-level context-aware model. Transactions on Asian and Low-Resource Language Information Processing 2021, 21, 1–16. [Google Scholar] [CrossRef]
  162. Zhang, L.; Wu, Y.; Mo, F.; Nie, J.Y.; Agrawal, A. MoqaGPT: Zero-Shot Multi-modal Open-domain Question Answering with Large Language Model. arXiv 2023, arXiv:2310.13265. [Google Scholar]
  163. Wang, P.; Fu, Y.; Zhou, Y.; Liu, K.; Li, X.; Hua, K.A. Exploiting Mutual Information for Substructure-aware Graph Representation Learning. IJCAI, 2020, pp. 3415–3421.
  164. Gallegos, I.O.; Rossi, R.A.; Barrow, J.; Tanjim, M.M.; Kim, S.; Dernoncourt, F.; Yu, T.; Zhang, R.; Ahmed, N.K. Bias and fairness in large language models: A survey. Computational Linguistics 2024, pp. 1–79. [CrossRef]
  165. Liang, P.P.; Wu, C.; Morency, L.P.; Salakhutdinov, R. Towards understanding and mitigating social biases in language models. International Conference on Machine Learning. PMLR, 2021, pp. 6565–6576.
  166. Delobelle, P.; Tokpo, E.K.; Calders, T.; Berendt, B. Measuring fairness with biased rulers: A survey on quantifying biases in pretrained language models. arXiv 2021, arXiv:2112.07447. [Google Scholar]
  167. Xu, Y.; Hu, L.; Zhao, J.; Qiu, Z.; Ye, Y.; Gu, H. A Survey on Multilingual Large Language Models: Corpora, Alignment, and Bias. arXiv 2024, arXiv:2404.00929. [Google Scholar]
  168. Navigli, R.; Conia, S.; Ross, B. Biases in large language models: origins, inventory, and discussion. ACM Journal of Data and Information Quality 2023, 15, 1–21. [Google Scholar] [CrossRef]
  169. Garrido-Muñoz, I.; Montejo-Ráez, A.; Martínez-Santiago, F.; Ureña-López, L.A. A survey on bias in deep NLP. Applied Sciences 2021, 11, 3184. [Google Scholar] [CrossRef]
  170. Wu, L.; Zheng, Z.; Qiu, Z.; Wang, H.; Gu, H.; Shen, T.; Qin, C.; Zhu, C.; Zhu, H.; Liu, Q.; others. A survey on large language models for recommendation. World Wide Web 2024, 27, 60. [Google Scholar] [CrossRef]
  171. Lin, L.; Wang, L.; Guo, J.; Wong, K.F. Investigating Bias in LLM-Based Bias Detection: Disparities between LLMs and Human Perception. arXiv 2024, arXiv:2403.14896. [Google Scholar]
  172. Rubinstein, I.S.; Good, N. Privacy by design: A counterfactual analysis of Google and Facebook privacy incidents. Berkeley Tech. LJ 2013, 28, 1333. [Google Scholar] [CrossRef]
  173. Yan, B.; Li, K.; Xu, M.; Dong, Y.; Zhang, Y.; Ren, Z.; Cheng, X. On protecting the data privacy of large language models (llms): A survey. arXiv 2024, arXiv:2403.05156. [Google Scholar]
  174. Voigt, P.; Von dem Bussche, A. The eu general data protection regulation (gdpr). A Practical Guide, 1st Ed., Cham: Springer International Publishing 2017, 10, 10–5555. [Google Scholar] [CrossRef]
  175. Goldman, E. An introduction to the california consumer privacy act (ccpa). Santa Clara Univ. Legal Studies Research Paper 2020. [Google Scholar] [CrossRef]
  176. Song, Y. Deep Learning Applications in the Medical Image Recognition. American Journal of Computer Science and Technology 2019, 2, 22–26. [Google Scholar] [CrossRef]
  177. Ni, H.; Meng, S.; Geng, X.; Li, P.; Li, Z.; Chen, X.; Wang, X.; Zhang, S. Time Series Modeling for Heart Rate Prediction: From ARIMA to Transformers. arXiv 2024, arXiv:2406.12199. [Google Scholar]
1
2
3
4
5
6
7
8
9
10
11
12
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated