Affiliation:
1. Petroleum Engineering Department, Texas A&M University, College Station, Texas, Brazos
2. Teale Engineering LLC., Houston, Texas, Harris
Abstract
Abstract
The efficiency of modern drilling operations depends on the planning phase to determine possible well trajectories and the ability of the directional driller to traverse them accurately. Deviations from the planned trajectory while drilling often require updates to the original well plan, involving drilling engineers and rig personnel, which can be time-consuming due to several uncertainties, such as formation tendencies, survey measurement inaccuracy, or estimation errors. To address these challenges, this paper proposes an innovative solution that leverages artificial intelligence (AI) methods, specifically deep reinforcement learning (DRL) to dramatically reduce the need for continuous corrections to the well plan while drilling. In the DRL paradigm, the proposed approach eliminates the need for constant plan adjustments by training a drilling agent to imitate the driller's ability to dynamically adjust the well trajectory in real time based on information from previous drilling logs, well plans, and near-the-bit measurements.
This research utilizes a physics-based simulation engine to model a directional drilling environment as a Markov Decision Process (MDP). The MDP is intended for an autonomous system based on geological data models and real-time measurements obtained while drilling (MWD) to train deep reinforcement learning (DRL) agents to drill directional wellbores that maintain maximum contact with the target formation. The simulator incorporates uncertainties of the real-world drilling environment, such as bit walk, formation properties, and drilling speed which, combined, help derive the actions performed by the drilling agent. Our findings reveal that using the proposed methodology in a virtual drilling environment, the drilling agent effectively navigates well paths in the presence of uncertainties and successfully tackle challenges such as avoiding excessive tortuosity and doglegs while maximizing contact with the target formation. Furthermore, the use of domain randomization during training enabled the RL agents to exhibit exceptional generalizability to a wide range of drilling scenarios through the random selection of drilling test sites from a set of unseen sites during training — demonstrating the ability of the agent to adapt and reach the target formation even when the initial well-plan is inaccurate — resulting in a 90% success rate.
This self-correcting approach demonstrates the potential for automated, proactive, self-contained steering operations with minimal human involvement. The developed simulation framework is a pioneering approach to enhancing real-time adjustments of drilling well paths using reinforcement learning and optimizing drilling operations. It is the first-of-its-kind method to augment drillers’ ability to navigate through drilling uncertainties and maximize pay zone contact and pave the way for the creation of robust, scalable, and practical autonomous drilling systems.
Reference18 articles.
1. ArnØ, Mikkel, Godhavn, John-Morten, and Aamo, Ole Morten. 2020. Deep Reinforcement Learning Applied to Managed Pressure Drilling. Proc., SPE Norway Subsurface Conference. https://doi.org/10.2118/200757-MS.
2. The theory of dynamic programming;Bellman;Bulletin of the American Mathematical Society,1954
3. Autonomous Decision-Making While Drilling;Cayeux;Energies,2021
4. Twin-delayed ddpg: A deep reinforcement learning technique to model a continuous movement of an intelligent robot agent;Dankwa;Proc., Proceedings of the 3rd international conference on vision, image and signal processing,2019
5. Dey, Sheelabhadra, Pendurkar, Sumedh, Sharon, Guni. 2022. A Joint Imitation-Reinforcement Learning Framework for Reduced Baseline Regret. https://ui.adsabs.harvard.edu/abs/2022arXiv220909446D.