توسعه مدل ریاضی الگوریتم یادگیری تقویتی برای خودکارسازی آب‏بندهای کشویی در کانال‌ها

نوع مقاله : مقالات پژوهشی

نویسندگان

دانشگاه تربیت مدرس

چکیده

امروزه با توجه به کمبود آب و ضعف مدیریت در بخش آب کشاورزی و به منظور استفاده بهینه از آب، بهبود عملکرد شبکه های آبیاری ضروری است. در سالهای اخیر، بکارگیری سامانه‏های کنترل خودکار پیشرفته به منظور بهبود عملکرد شبکه های آبیاری مورد توجه محققان قرارگرفته است. برای استفاده از این سامانه‏ها، تهیه مدل ریاضی سیستم های کنترل خودکار و سازه های مربوط بصورت تلفیقی با مدل‏های هیدرودینامیکی ضروری است. هدف اصلی این تحقیق، توسعه مدل ریاضی سیستم کنترل خودکار بالادست دریچه کشویی با استفاده از الگوریتم یادگیری تقویتی بصورت یک زیربرنامه در مدل هیدرودینامیکی ICSS است. دو سناریوی افزایشی و کاهشی برای جریان ورودی با دبی اولیه 25 لیتر در ثانیه در نظر گرفته شد و شبیه سازی انجام گرفت. شاخصهای حداکثر خطای مطلق، میانگین متوسط خطاها و زمان پاسخگویی سیستم برای ارزیابی عملکرد سیستم کنترل توسعه داده شد. مقادیر این شاخصها برای سناریوی کاهش جریان برابر صفر بدست آمد. نتایج حاصل نشان می دهد مدل توسعه یافته عملکرد مناسبی داشته و سطح آب را بخوبی کنترل می کند. از این رو توسعه آن برای کانالهای آبیاری پیشنهاد می شود.

کلیدواژه‌ها


عنوان مقاله [English]

Development of Reinforcement Learning Algorithm for Automation of Slide Gate Check Structure in Canals

نویسندگان [English]

  • K. Shahverdi
  • M.J. Monem
Tarbiat Modares University
چکیده [English]

Introduction: Nowadays considering water shortage and weak management in agricultural water sector and for optimal uses of water, irrigation networks performance need to be improveed. Recently, intelligent management of water conveyance and delivery, and better control technologies have been considered for improving the performance of irrigation networks and their operation. For this affair, providing of mathematical model of automatic control system and related structures, which connected with hydrodynamic models, is necessary. The main objective of this research, is development of mathematical model of RL upstream control algorithm inside ICSS hydrodynamic model as a subroutine.
Materials and Methods: In the learning systems, a set of state-action rules called classifiers compete to control the system based on the system's receipt from the environment. One could be identified five main elements of the RL: an agent, an environment, a policy, a reward function, and a simulator. The learner (decision-maker) is called the agent. The thing it interacts with, comprising everything outside the agent, is called the environment. The agent selects an action based on existing state in the environment. When the agent takes an action and performs on environment, the environment goes new state and reward is assigned based on it. The agent and the environment continually interact to maximize the reward. The policy is a set of state-action pair, which have higher rewards. It defines the agent's behavior and says which action must be taken in which state. The reward function defines the goal in a RL problem. The reward function defines what the good and bad events are for the agent. The higher the reward, the better the action. The simulator provides environment information. In irrigation canals, the agent is the check structures. The action and state are the check structures adjustment and the water depth, respectively. The environment comprises the hydraulic information existing in the canal. Policy is a map of water depth-check structure pairs. Reward function is defined based on the difference between water depth and target depth, and the simulator is a hydrodynamic model which, in the present study, was Irrigation Conveyance System Simulation (ICSS). In the developed RL, the RL begins with required initializations, and then the canal structures are operated. While the maximum reward is reached at the time step of t, the agent receives some representation of the state of the environment and, on that basis, selects an action. The simulator performs the action and provides information on the state of the new environment by simulating the canal system. Finally, the reward is assigned. Maximizing the reward, the RL goes on to the next time step. This process is continued until the final simulation time step is reached. The learning process is similar for all operations. The ICSS hydrodynamic model was used to simulate the canal system and provide the environmental information. Input to the ICSS was the new selected action. The ICSS performed the action and simulated the canal system. The output from the ICSS was information on the new environment for use in the next time step. Two scenarios of flow increase and decrease with initial flow of 25 l/s were simulated. MAE (Maximum Absolute Error), IAE (Integral of Absolute magnitude of Error) and SRT (System Response Time) indicators have been used to assess developed model. For flow decrease scenario, the indicators value are obtained zero.
Results and Discussion: Results were obtained from the performed scenarios. In the flow increase scenario, water depth variations were inside the dead band, therefore, SRT indicator was obtained zero. The MAE and IAE indicators were obtained 3.5% and 2.57%, respectively, which showed the water depth deviations from target depth was very low. In the flow decrease scenario, the all indicators values were obtained zero. At time zero in two scenarios, 1000 populations were generated and tested. As the RL controlled the water depth, it generated new populations, too. The reason for this is that the RL generates a new population if there is no classifier with maximum reward in the population. There are no new generation after 0.03 hr and 0.16 hr in flow increase and flow decrease scenarios, respectively. Considering the results, it could be concluded that the developed control system is a powerful technique in terms of accuracy and response time for water depth control.
Conclusion: In this research, the RL upstream control system was developed and connected with ICSS hydrodynamic model and evaluated in two scenarios of flow increase and flow decrease. The results showed an ability to control of deviations, short response time and accurate performance of the developed RL control system, which could be used for further study in irrigation canals.

کلیدواژه‌ها [English]

  • Learning
  • Model development
  • Performance improvement
Alpaydin E. 2004. Introduction to Machine Learning. The MIT Press.
2- Bennett S. 1993. A History of Control Engineering. Peter Peregrinus Ltd.
3- Burt C.M. 1983. Regulation of Sloping Canals by Automatic Downstream Control. Dept. of Agricultural and Irrigation Engineering, Utah State University.
4- Butz M., and Wilson S. 2001a. An algorithmic description of XCS. Advances in Learning Classifier Systems, 267-274.
5- Butz M.V., Kovacs T., Lanzi P.L., and Wilson S.W. 2001b. How XCS evolves accurate classifiers. Proceedings of the Third Genetic and Evolutionary Computation Conference (GECCO).
6- Buyalski C.P., Ehler D., Falvey, H.T., Rogers D.C., and Serfozo E.A. 1991. Canal Systems Automation Manual: Volume 1. US Department of Interior.
7- Clemmens A. and Replogle J. 1989. Control of irrigation canal networks. Journal of irrigation and drainage engineering, 115(1): 96-110.
8-Clemmens A.J., Kacerek T. F., Grawitz B., and Schuurmans W. 1998. Test cases for canal control algorithms. Journal of irrigation and drainage engineering, 124(1): 23-30.
9- Hernandez J. and Merkley G. 2011. Canal structure automation rules using an accuracy-based learning classifier system, a genetic algorithm, and a hydraulic simulation model. I: result. Journal of irrigation and drainage engineering, 137(1), 12–16.
10- Malaterre P.O., Rogers D. C., and Schuurmans J. 1998. Classification of canal control algorithms. Journal of irrigation and drainage engineering, 124(1): 3-10.
11- Monem M.J., and Manz D. H. 1994. Application of simulation techniques for improving the performance of irrigation conveyance systems. Journal of Water Resources Engineering, 2: 1-22.
12-Rogers D.C., and Goussard J. 1998. Canal control algorithms currently in use. Journal of irrigation and drainage engineering, 124(1): 11-15.
13- Sawadogo S., Faye R., Benhammou A., and Akouz K. 2000. Decentralized adaptive predictive control of multireach irrigation canal. Institute of Electrical and Electronics Engineers: 5: 3438-3442.
14-Sutton R.S., and Barto A. G. 1998. Reinforcement Learning: An Introduction. Cambridge University Press.
15- Wagemaker R. 2005. Model predictive control on irrigation canals application of various internal models. Delft University of Technology, Netherland.
16- Zhang Q., Wu C.H., and Tilt K. 1996. Application of fuzzy logic in an irrigation control system. Industrial Technology, Proceedings of Institute of Electrical and Electronics Engineers International Conference on, Institute of Electrical and Electronics Engineers, New York.
CAPTCHA Image