Bridging Large Language Models and Reinforcement Learning: Innovations and Real-World Applications

Authors

  • Pranav Yadav Department of Computer Science, University of East Anglia (UEA)
  • Jia Li Tan School of Computing, University of Wolverhampton

Keywords:

Large Language Models, Reinforcement Learning, Natural Language Processing, Artificial Intelligence, Deep Learning, Neural Networks, Language Generation

Abstract

The combination of large language models and reinforcement learning represents a burgeoning area of research and application. Large language models, such as GPT (Generative Pre-trained Transformer), have demonstrated remarkable capabilities in natural language understanding, generation, and translation tasks. Reinforcement learning, on the other hand, is a paradigm in machine learning where agents learn to make decisions by interacting with an environment to maximize cumulative rewards. Research in this area aims to leverage the strengths of both large language models and reinforcement learning to create more robust, context-aware, and adaptive AI systems for diverse applications ranging from dialogue systems to content generation and beyond.

References

[1] Q. Lu, B. Qiu, L. Ding, L. Xie, and D. Tao, "Error analysis prompting enables human-like translation evaluation in large language models: A case study on chatgpt," arXiv preprint arXiv:2303.13809, 2023.

[2] C. Li et al., "Llava-med: Training a large language-and-vision assistant for biomedicine in one day," Advances in Neural Information Processing Systems, vol. 36, 2024.

[3] E. Kasneci et al., "ChatGPT for good? On opportunities and challenges of large language models for education," Learning and individual differences, vol. 103, p. 102274, 2023.

[4] N. Kandpal, H. Deng, A. Roberts, E. Wallace, and C. Raffel, "Large language models struggle to learn long-tail knowledge," in International Conference on Machine Learning, 2023: PMLR, pp. 15696-15707.

[5] J. Hoffmann et al., "Training compute-optimal large language models," arXiv preprint arXiv:2203.15556, 2022.

[6] Q. He et al., "Can Large Language Models Understand Real-World Complex Instructions?," in Proceedings of the AAAI Conference on Artificial Intelligence, 2024, vol. 38, no. 16, pp. 18188-18196.

[7] L. Floridi, "AI as agency without intelligence: On ChatGPT, large language models, and other generative models," Philosophy & Technology, vol. 36, no. 1, p. 15, 2023.

[8] E. Ferrara, "Should chatgpt be biased? challenges and risks of bias in large language models," arXiv preprint arXiv:2304.03738, 2023.

[9] Z. Chen et al., "Exploring the potential of large language models (llms) in learning on graphs," ACM SIGKDD Explorations Newsletter, vol. 25, no. 2, pp. 42-61, 2024.

[10] J. Austin et al., "Program synthesis with large language models," arXiv preprint arXiv:2108.07732, 2021.

[11] K. Patil and B. Desai, "From Remote Outback to Urban Jungle: Achieving Universal 6G Connectivity through Hybrid Terrestrial-Aerial-Satellite Networks," Advances in Computer Sciences, vol. 6, no. 1, pp. 1− 13-1− 13, 2023.

[12] D. Myers et al., "Foundation and large language models: fundamentals, challenges, opportunities, and social impacts," Cluster Computing, vol. 27, no. 1, pp. 1-26, 2024.

[13] B. Desai and K. Patel, "Reinforcement Learning-Based Load Balancing with Large Language Models and Edge Intelligence for Dynamic Cloud Environments," Journal of Innovative Technologies, vol. 6, no. 1, pp. 1− 13-1− 13, 2023.

[14] S. Pal, M. Bhattacharya, S.-S. Lee, and C. Chakraborty, "A domain-specific next-generation large language model (LLM) or ChatGPT is required for biomedical engineering and research," Annals of Biomedical Engineering, vol. 52, no. 3, pp. 451-454, 2024.

Downloads

Published

2025-01-14

Similar Articles

<< < 1 2 3 4 > >> 

You may also start an advanced similarity search for this article.