From Statistical Models to LLMs: A Comprehensive Survey of Language Model Evolution

Authors

    Hamid Hassanpour * 1. Department of Computer Engineering and IT, Shahrood University of Technology, Shahrood, Iran. h.hassanpour@shahroodut.ac.ir
    Maryam Majidi 1. Department of Computer Engineering and IT, Shahrood University of Technology, Shahrood, Iran.
https://doi.org/10.61838/jaiai.1.4.5

Keywords:

Language Modeling, Natural Language Processing, Statistical Language Models, Recurrent Neural Networks, Transformer Models, Large Language Models

Abstract

The evolution of language models marks one of the most transformative trajectories in the history of Natural Language Processing (NLP). This survey aims to provide a structured overview of key developments, tracing the progression from early statistical models to deep learning approaches, and culminating in the rise of Transformer-based architectures and Large Language Models (LLMs). We categorize and synthesize key contributions based on algorithmic paradigms, performance metrics, and systemic challenges. Specifically, we examine contributions from foundational models such as n-gram and Hidden Markov Models (HMMs), advances enabled by Recurrent Neural Networks (RNNs) and Long Short-Term Memory (LSTM) networks, and the paradigm shift introduced by self-attention mechanisms in Transformer architectures. Additionally, the survey discusses how LLMs have expanded the capabilities of NLP systems in tasks including text generation, translation, and dialogue modeling. Alongside these achievements, we critically highlight ongoing challenges, including model bias, interpretability, computational costs, and environmental impacts, drawing on recent literature and evaluation frameworks. Emerging trends toward improving model efficiency, fairness, and societal alignment are also explored. By mapping historical progress and identifying open questions, this article offers a comprehensive reference for researchers and practitioners interested in the evolving landscape of language models.

Downloads

Download data is not yet available.

Downloads

Published

2024-10-01

Submitted

2024-06-07

Revised

2024-07-25

Accepted

2024-09-02

How to Cite

Hassanpour, H., & Majidi, M. (2024). From Statistical Models to LLMs: A Comprehensive Survey of Language Model Evolution. Journal of Artificial Intelligence, Applications and Innovations, 1(4), 55-75. https://doi.org/10.61838/jaiai.1.4.5

Similar Articles

1-10 of 16

You may also start an advanced similarity search for this article.