Hybrid Intelligence for Industry 4.0: Integrating Large Language Models and Reinforcement Learning
DOI:
https://doi.org/10.70917/ijcisim-2025-0038Abstract
The future of industry is driven by intelligent systems capable of autonomous decision-making, dynamic adaptation, and integrating human knowledge. In this context, hybrid approaches emerge that combine data-driven methodologies with external knowledge sources. This paper introduces OPRA-RL, a hybrid framework that integrates Reinforcement Learning (RL) with the OPRA (Observation-Prompt-Response-Action) framework. OPRA-RL integrates the self-learning capabilities of RL with the contextual expertise and adaptive reasoning of Large Language Models, such as ChatGPT, to tackle challenges in complex, real-world environments. We present an analytical formulation of OPRA-RL, highlighting its complex reward structure designed to balance internal learning with external guidance through prompts. The proposed OPRA-Q-Learning variant is implemented and validated experimentally through a simulated decision-making game, illustrating how knowledge-informed autonomy can outperform traditional RL in scenarios characterized by sparse data, high complexity, or novel challenges. Our findings reveal how multifaceted reward systems, external knowledge integration, and dynamic decision-making enhance agent performance in unpredictable environments. By bridging the gap between knowledge-informed and data-driven AI, OPRA-RL contributes to smarter and resilient autonomous systems.
Downloads
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2025 Vagan Terziyan, Oleksandra Vitko, Oleksandr Terziyan, Artur Terziian

This work is licensed under a Creative Commons Attribution 4.0 International License.