This paper presents a novel model predictive adaptive cruise control strategy of intelligent electric vehicles based on deep reinforcement learning algorithm for driver characteristics. Firstly, the influence mechanis...
详细信息
This paper presents a novel model predictive adaptive cruise control strategy of intelligent electric vehicles based on deep reinforcement learning algorithm for driver characteristics. Firstly, the influence mechanism of factors such as inter-vehicle distance, relative speed and time headway (THW) on the driver's behavior in the process of car following is analyzed by the correlation coefficient method. Then, the driver behavior in the process of car following is learned from the natural driving data, the car following model is established by the deep deterministic policy gradient (DDPG) algorithm, and the output acceleration of the DDPG model is used as the reference trajectory of the ego vehicle's acceleration. Next, in order to reflect the driver behavior and achieve multi performance objective optimization of adaptive cruise control of intelligent electric vehicles, the model predictive controller (MPC) is designed and used for tracking the desired acceleration produced by the car following DDPG model. Finally, the performance of the proposed adaptive cruise control strategy is evaluated by the experimental tests, and the results demonstrate the effectiveness of proposed control strategy.
Mobile-edge computing (MEC) has emerged as a promising paradigm that moves tasks running in the cloud to edge servers. In MEC systems, there are various individual requirements, such as less user-perceived time and lo...
详细信息
ISBN:
(纸本)9783031243851;9783031243868
Mobile-edge computing (MEC) has emerged as a promising paradigm that moves tasks running in the cloud to edge servers. In MEC systems, there are various individual requirements, such as less user-perceived time and lower energy consumption. In this case, substantial efforts have been paid to task allocation, aiming at enabling lower latency and higher resource utilization. However existing studies on multiple-objectives task allocation algorithms rarely consider the Pareto efficient problem, where no objective could be further improved without vitiating the other objectives optimization. In this paper, we propose a Pareto-efficient task-allocation framework based on a deep reinforcement learning algorithm. We give the formal formulations for objectives and construct a multi-objectives' optimization model for task allocation. Then a Pareto efficient algorithm is proposed to solve the problem of conflicting among multi-objectives. By coordinating multi-objectives parameters get from Pareto efficient algorithm, the deepreinforcementlearning model takes a Pareto-efficient task allocation to improve real-time and resource utilization performance. We evaluate the proposed framework over various real-world tasks and compare it with existing allocating tasks models in edge computing networks. By using the proposed framework, we can get an accuracy that not be lower than 90% under the 0.6 s latency requirement. The simulation results also show that the proposed framework achieves lower latency and higher resource utilization compared to other task allocation methods.
The past decade witnessed an explosive growth on the number of IoT devices (objects/suppliers), including portable mobile devices, autonomous vehicles, sensors and intelligence appliances. To realize the digital repre...
详细信息
The past decade witnessed an explosive growth on the number of IoT devices (objects/suppliers), including portable mobile devices, autonomous vehicles, sensors and intelligence appliances. To realize the digital representations of objects, Digital Twins (DTs) are key enablers to provide real-time monitoring, behavior simulations and predictive decisions for objects. On the other hand, Mobile Edge Computing (MEC) has been envisioned as a promising paradigm to provide delay-sensitive services for mobile users (consumers) at the network edge, e.g., real-time healthcare, AR/VR, online gaming, smart cities, and so on. In this paper, we study a novel DT migration problem for high quality service provisioning in an MEC network with the mobility of both suppliers and consumers for a finite time horizon, with the aim to minimize the sum of the accumulative DT synchronization cost of all suppliers and the total service cost of all consumers requesting for different DT services. To this end, we first show that the problem is NP-hard, and formulate an integer linear programming solution to the offline version of the problem. We then develop a deepreinforcementlearning (DRL) algorithm for the DT migration problem, by considering the system dynamics and heterogeneity of different resource consumptions, mobility traces of both suppliers and consumers, and workloads of cloudlets. We finally evaluate the performance of the proposed algorithms through experimental simulations. Simulation results demonstrate that the proposed algorithms are promising.
The increasing use of distributed generation (DG) in power systems can result in frequent online voltage problems. In scenarios in which substantial DG prediction errors occur because of high DG accommodation levels, ...
详细信息
The increasing use of distributed generation (DG) in power systems can result in frequent online voltage problems. In scenarios in which substantial DG prediction errors occur because of high DG accommodation levels, traditional technical solutions cannot meet the online voltage regulation requirements. Hence, new resources for online voltage regulation are needed. Here, flexible network reconfiguration is proposed to coordinate with the existing resources for severe online voltage deviations. For online topology-based voltage regulation (OTVR), the authors develop a deepreinforcementlearning (DRL) algorithm based on the following specially designed modelling to enhance the computational performance. The mechanism of action incorporates the concepts of local research, branch exchange, and action separation, and it effectively simplifies the action dimension and action space. In addition, for the graph data in OTVR, a graph convolution network (GCN) is applied to obtain better feature extraction. Case studies performed on IEEE 14-bus, 33-bus, 141-bus systems and a practical system verify that our proposed algorithm can obtain close to optimal solutions in 2 s which can meet the needs of online voltage regulation. Moreover, we verify that the developed OTVR effectively increases DG penetration and decreases the need for investment in additional regulating devices.
Channel selection is a challenging task in cognitive radio vehicular networks. Vehicles have to sense the channels periodically. Due to this, a lot of time is wasted which could have been utilised for transmission of ...
详细信息
Channel selection is a challenging task in cognitive radio vehicular networks. Vehicles have to sense the channels periodically. Due to this, a lot of time is wasted which could have been utilised for transmission of data. Employing road side units (RSUs) in sensing can prove to be useful for this purpose. The RSUs may select the channel and allocate it to the vehicles on demand. However, this sensing should be proactive. RSUs should know in advance the channel to be allocated when requested. For this purpose, a deep reinforcement learning algorithm namely deepreinforcementlearning based optimal channel selection is proposed in this study for training the network according to the previously sensed data. Proposed protocol is simulated and results are compared with the existing methods. The packet delivery ratio is increased by 2%, throughput is increased by 1.8%, average delay is decreased by 2% and primary user collision ratio is reduced by 3.2% when compared with similar recent work by varying number of vehicles. On the other hand, when compared with similar recent work by varying channel availability, the packet delivery ratio is increased by 4.5 %, throughput by 4.3%, average delay is decreased by 3% and PU collision ratio by 5.5%.
As the proton exchange membrane fuel cell (PEMFC) is a nonlinear, time-varying, multiple-input multiple-output system, an advanced controller with strong robustness and adaptability is required for controlling PEMFC s...
详细信息
As the proton exchange membrane fuel cell (PEMFC) is a nonlinear, time-varying, multiple-input multiple-output system, an advanced controller with strong robustness and adaptability is required for controlling PEMFC stack temperature and achieve a high operation efficiency. In this paper, a data driven optimal controller is proposed for controlling the stack temperature, which is based on large-scale deepreinforcementlearning. In addition, a new deep reinforcement learning algorithm termed curriculum guidance strategy large-scale dual-delay deep deterministic policy gradient (CGS-L4DPG) algorithm is proposed for this controller. The design of this algorithm introduces the concepts of the curriculum guidance strategy and imitation learning, and its inclusion improves the performance and robustness of the proposed controller. The simulation results show that, taking advantage of the high adaptability and robustness of CGS-L4DPG algorithm, the proposed controller can more effectively control the PEMFC stack temperature than existing control algorithms.
The past decade experienced an explosive growth on the number of IoT devices connected to the Internet. Digital twins (DTs) emerge as key enablers to provide digital representations of objects for their monitoring, si...
详细信息
ISBN:
(纸本)9798350390605;9783903176638
The past decade experienced an explosive growth on the number of IoT devices connected to the Internet. Digital twins (DTs) emerge as key enablers to provide digital representations of objects for their monitoring, simulation, prediction and maintenance. At the same time, mobile edge computing (MEC) is envisioned as a promising paradigm to provide various delay-sensitive services for mobile users at the edge of core networks. In this paper, we study a novel cost-aware DT migration problem for effective service provisioning in an DT-empowered MEC network within a finite time horizon, with the aim to minimize the service cost. We first show the NP-hardness of the problem, and formulate an integer linear programming solution to the offline version of the problem, assuming that the mobility information of both objects and users for the given time horizon is given. Considering the system dynamics and heterogeneity of various resource usages, the mobility of objects and users, we then develop an efficient deep reinforcement learning algorithm for the online DT migration problem. We finally evaluate the performance of the proposed algorithms. Experimental results demonstrate that the proposed algorithms are promising.
We propose the optical deep Q network (ODQN) algorithm based on optical neural networks (ONNs) to accelerate calculation in 2D grid path planning task. The calculated results demonstrate that the innovative algorithm ...
详细信息
In view of the serious problems of increasing delay energy consumption and decreasing service quality caused by complex network state and massive computing data in Internet of vehicles (IOT), a high reliable computing...
详细信息
In view of the serious problems of increasing delay energy consumption and decreasing service quality caused by complex network state and massive computing data in Internet of vehicles (IOT), a high reliable computing offloading strategy based on edge computing is proposed in this paper. Firstly, the architecture of the Internet of vehicles is designed. The real-time business of terminal vehicles is directly unloaded to the mobile edge computing (MEC) equipment for processing, which reduces the high transmission delay of data on the core network. The combination of software defined network (SDN) and MEC is used to provide flexible network control and centralized resource management for the Internet of vehicles. Then, according to the computing model, communication model and privacy protection model, a joint computing offload and resource allocation strategy is proposed. Finally, taking the shortest time delay and minimum computing cost as the optimization objectives, Q-learning is used to solve the problem to achieve the optimization of unloading strategy, that is, the optimal allocation of communication and computing resources, and the system security is the best. Based on the Matlab simulation platform, the system model is built to carry out the experimental test. The results show that compared with other strategies, the unloading strategy can achieve fast convergence and reduce the system overhead effectively, and the computation complexity, data size and the number of computation nodes have the least impact on the delay.
Considering the importance of the energy management strategy for hybrid electric vehicles, this paper is aiming at addressing the energy optimization control issue using reinforcementlearningalgorithms. Firstly, thi...
详细信息
Considering the importance of the energy management strategy for hybrid electric vehicles, this paper is aiming at addressing the energy optimization control issue using reinforcementlearningalgorithms. Firstly, this paper establishes a hybrid electric vehicle power system model. Secondly, a hierarchical energy optimization control architecture based on networked information is designed, and a traffic signal timing model is used for vehicle target speed range planning in the upper system. More specifically, the optimal vehicle speed is optimized by a model predictive control algorithm. Thirdly, a mathematical model of vehicle speed variation in connected and unconnected states is established to analyze the effect of vehicle speed planning on fuel economy. Finally, three learning-based energy optimization control strategies, namely Q-learning, deep Q network (DQN), and deep deterministic policy gradient (DDPG) algorithms, are designed under the hierarchical energy optimization control architecture. It is shown that the Q-learningalgorithm is able to optimize energy control;however, the agent will meet the "dimension disaster" once it faces a high-dimensional state space issue. Then, a DQN control strategy is introduced to address the problem. Due to the limitation of the discrete output of DQN, the DDPG algorithm is put forward to achieve continuous action control. In the simulation, the superiority of the DDPG algorithm over Q-learning and DQN algorithms in hybrid electric vehicles is illustrated in terms of its robustness and faster convergence for better energy management purposes.
暂无评论