Complex network cognition-based federated reinforcement learning end-to-end autonomous driving control system, method, and vehicular device
12415528 ยท 2025-09-16
Assignee
Inventors
- Yingfeng CAI (Zhenjiang, CN)
- Sikai Lu (Zhenjiang, CN)
- Hai WANG (Zhenjiang, CN)
- Yubo LIAN (Zhenjiang, CN)
- Long CHEN (Zhenjiang, CN)
- Qingchao Liu (Zhenjiang, CN)
Cpc classification
B60W50/08
PERFORMING OPERATIONS; TRANSPORTING
G06V10/774
PHYSICS
B60W60/001
PERFORMING OPERATIONS; TRANSPORTING
Y02T10/40
GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
G06V20/56
PHYSICS
International classification
B60W50/08
PERFORMING OPERATIONS; TRANSPORTING
B60W60/00
PERFORMING OPERATIONS; TRANSPORTING
Abstract
The provided are a federated reinforcement learning (FRL) end-to-end autonomous driving control system and method, as well as vehicular equipment, based on complex network cognition. An FRL algorithm framework is provided, designated as FLDPPO, for dense urban traffic. This framework combines rule-based complex network cognition with end-to-end FRL through the design of a loss function. FLDPPO employs a dynamic driving guidance system to assist agents in learning rules, thereby enabling them to navigate complex urban driving environments and dense traffic scenarios. Moreover, the provided framework utilizes a multi-agent FRL architecture, whereby models are trained through parameter aggregation to safeguard vehicle-side privacy, accelerate network convergence, reduce communication consumption, and achieve a balance between sampling efficiency and high robustness of the model.
Claims
1. A complex network cognition-based federated reinforcement learning (FRL) end-to-end autonomous driving control system comprising: a measurement encoder, an image encoder, a complex network cognition module, a reinforcement learning module, and a federated learning module, wherein: the measurement encoder is implemented by at least one processor and is configured to obtain state quantities required by the complex network cognition module, which is implemented by the at least one processor, and the reinforcement learning module, which is implemented by the at least one processor, the state quantities required by the complex network cognition module comprise a x-coordinate, a y-coordinate, a heading angle change and a speed of a given driving agent of a plurality of driving agents, the state quantities required by the complex network cognition module are handed over to the complex network cognition module as inputs, the state quantities required by the reinforcement learning module comprise a steering wheel angle, a throttle, a brake, a gear, a lateral speed and a longitudinal speed, the state quantities required by the reinforcement learning module are given to the reinforcement learning module after extracting features from a two-layer fully connected network; the image encoder is implemented by the at least one processor and is configured to obtain an amount of image implicit state required by the reinforcement learning module, an image used is a 15-channel semantic bird's eye view (BEV), i.sub.RL[0, 1].sup.192*192*15, 192 is in pixels and the BEV used is 5 px/m, 15 channels contain a drivable domain, a desired path, a road edge, 4 frames of other vehicles, 4 frames of pedestrians, and 4 frames of traffic signs, wherein the desired path is calculated using an A* algorithm, the semantic BEV is extracted by multilayer convolutional layers to extract implicit features and then passed to the reinforcement learning module as another part of the inputs; the complex network cognition module is configured to model a driving situation of a driving subject, and to obtain a maximum risk value of the driving subject in a current driving situation according to the state quantity provided by the measurement encoder, and finally to output dynamic driving suggestions based on a risk value through an activation function; the reinforcement learning module is configured to integrate the state quantities output from the measurement encoder and the image encoder, output corresponding strategies according to integrated network inputs, and interact with an environment to generate experience samples stored in a local replay buffer in the federated learning module, which is implemented by the at least one processor, when a number of the experience samples reaches a certain threshold, a batch of sample is taken from the local replay buffer for training, and finally trained neural network parameters are uploaded to the federated learning module; and the federated learning module is configured to receive the finally trained neural network parameters uploaded by the reinforcement learning module of the plurality of driving agents, and to aggregate a set of global parameters based on the finally trained neural network parameters, and finally to send the global parameters to the plurality of driving agents until a neural network converges, a global parameter aggregation is performed by a following equation:
2. The complex network cognition-based FRL end-to-end autonomous driving control system according to claim 1, wherein a modeling process of the complex network cognition module constructs a dynamic complex network model with a traffic participant and a road infrastructure as nodes:
3. The complex network cognition-based FRL end-to-end autonomous driving control system according to claim 2, wherein a Gaussian function is used in the dynamic complex network to reveal a static property between nodes:
4. The complex network cognition-based FRL end-to-end autonomous driving control system according to claim 3, wherein a series of isofield lines are used to delineate the safety field, a top view projection of the series of isofield lines is a region covered by a series of ellipses, with a region covered by a smallest center ellipse being a core domain, a region between the smallest center ellipse and a second ellipse being a limited domain, and a region between the second ellipse and a largest ellipse being an extended domain, size and shape of the region are determined by the series of isofield lines, are related to a vehicle shape and a motion state, and are described based on the Gaussian function, a direction of the safety field is aligned with a direction of vehicle motion, when a vehicle is in motion, the risk center O(x.sub.0, y.sub.0) of the safety field will be transferred to a new risk center O(x.sub.0, y.sub.0): {(1, 0)(0,1)}, k.sub.v's positive or negative is related to the direction of vehicle motion, and denotes an angle of a transfer vector k.sub.v|{right arrow over (v)}| with axes in a Cartesian coordinate system, a virtual vehicle, with a vehicle length l.sub.v and a vehicle width w.sub.v, is formed under risk center transfer, a dynamic safety field:
5. The complex network cognition-based FRL end-to-end autonomous driving control system according to claim 4, wherein a risk perception is categorized into three types on a planar scale based on different levels of human driving reaction time: a first cognitive domain, a second cognitive domain and an extra-domain space, wherein: the first cognitive domain:
6. The complex network cognition-based FRL end-to-end autonomous driving control system according to claim 1, wherein a CARLA simulator is used as an interaction environment, the CARLA simulator realizes vehicle control by inputting control quantities of a steering, a throttle and a brake, wherein steering [1, 1], throttle [0, 1] and brake [0, 1], based on a CARLA simulator's control method, a reinforcement learning action space [1, 1].sup.2, is categorized into the steering and a throttle-brake, when outputting the throttle-brake, [1, 0] denotes the brake and [0, 1] denotes the throttle, the driving control system outputs two parameters of the beta distribution by the reinforcement learning module, and then obtains a policy action by sampling:
7. The complex network cognition-based FRL end-to-end autonomous driving control system according to claim 6, wherein in a training process, a parameter updating is performed through following loss functions for the reinforcement learning module: .sub.ppo denotes a clipped policy gradient loss with advantages estimated using a generalized advantage estimation,
.sub.exp denotes a maximum entropy loss, H(.sub.(.Math.|i.sub.RL, m.sub.RL)) denotes an entropy of a policy .sub. under an image input i.sub.RL and a measurement input m.sub.RL, and
(1, 1) denotes a uniform distribution, L.sub.exp encourage the given driving agent to explore by converging an action distribution to the uniform distribution, .sub.exp denotes a weight of the maximum entropy loss,
.sub.risk denotes dynamic risk suggestions based loss, and
.sub.{TN.sub.
8. The complex network cognition-based FRL end-to-end autonomous driving control system according to claim 1, wherein in a training process, a parameter updating is performed through following loss functions for the reinforcement learning module: .sub.ppo denotes a clipped policy gradient loss with advantages estimated using a generalized advantage estimation,
.sub.exp denotes a maximum entropy loss, H(.sub.(.Math.|i.sub.RL, m.sub.RL)) denotes an entropy of a policy .sub. under an image input i.sub.RL and a measurement input m.sub.RL, and
(1, 1) denotes a uniform distribution,
.sub.exp encourage the given driving agent to explore by converging an action distribution to the uniform distribution, .sub.exp denotes a weight of the maximum entropy loss,
.sub.risk denotes dynamic risk suggestions based loss, and
.sub.{TN.sub.
9. A complex network cognition-based FRL end-to-end autonomous driving control method, comprising the following steps: step 1: building an urban dense traffic simulation environment in a CARLA simulator, wherein the simulation environment contains a driving subject, traffic participants, and a road infrastructure, the driving subject is a plurality of agents, modeled as Markov decision processes, respectively, and using a reinforcement learning module implemented by at least one processor for a steering wheel, a throttle and a brake control, the Markov decision process is described by a tuple (S, A, P, R, ), wherein S denotes a state set, corresponding to state quantities acquired by a measurement encoder implemented by the at least one processor and an image encoder implemented by the at least one processor, and contains a steering wheel angle, a throttle, a brake, a gear, lateral and longitudinal speeds, and a 15-channel semantic BEV; A denotes an action set, corresponding to the steering wheel, the throttle, and the brake control quantities of the driving subject; P denotes a state transfer equation p: SA.fwdarw.P(S), each state-action pair (s, a)SA has a probability distribution p(.Math.|s, a) of entering a new state after adopting an action a, in a state s; R denotes a reward function R: SXA.fwdarw.R, R(s.sub.t+1, s.sub.t, a.sub.t) denotes a reward obtained after entering the new state s.sub.t+1 from an original state s.sub.t, a goodness of performing the action is defined by the reward function; denotes a discount factor, [0, 1], is configured to compute a cumulative reward (.sub.)=.sub.i=0.sup.T.sup.ir.sub.i, wherein T denotes a current moment, .sup.i denotes a discount factor of moment i, and r.sub.i denotes an immediate reward of moment i, a solution to the Markov decision process is to find a strategy : S.fwdarw.A maximize the cumulative reward *:=argmax.sub.(.sub.), the reinforcement learning module integrates implicit state quantities output by the measurement encoder and the image encoder and outputs a corresponding optimal control policy; step 2: building the complex network cognition module implemented by the at least one processor to model a driving situation of the driving subject, establish a complex network model, and output dynamic driving suggestions based on state quantities provided by the measurement encoder through an activation function, the complex network model represents a dynamic relationship between nodes within a field range through a variable Gaussian safety field based on risk center transfer, the nodes contain the driving subject, the traffic participants, and the road infrastructure; step 3: constructing an end-to-end neural network, comprising 2 fully connected layers used by the measurement encoder, 6 convolutional layers used by the image encoder and 6 fully connected layers used by the reinforcement learning module, the neural network has two output heads, an action head and a value head, the action head outputs two parameters of a beta distribution and the value head outputs a value of the action; step 4: interacting the driving subject with a CARLA simulation environment and storing experiences in respective local replay buffers, wherein when the a number of samples reaches a certain threshold, samples of the number of samples are sampled from the respective local replay buffer according to a mini-batch, and then neural network parameters are updated according to a designed loss function; step 5: uploading the neural network parameters corresponding to each driving subject to a federated learning module implemented by the at least one processor, aggregating global parameters based on the neural network parameters according to an aggregation interval, and sending the global parameters to each agent until the neural network converges; wherein in step 1, the traffic participants comprise other vehicles and pedestrians, the road infrastructure contains traffic lights and traffic signs, represented in the image encoder input using a 4-frame semantic BEV; in step 2, the complex network cognition module takes different activation functions according to different driving suggestions: .sub.risk, a guidance of the agent is realized by calculating a KL-divergence of the strategy and the dynamic driving suggestions output by the driving subject N.sub.z=100 steps before a termination state; in step 5, the federated learning module is a multi-agent framework that uses a local replay buffer architecture between agents; and in step 5, a global parameter aggregation uses a parameter-averaged aggregation method with the aggregation interval of 256.
10. A vehicular device, wherein the vehicular device includes A complex network cognition-based FRL end-to-end autonomous driving control system comprising: a measurement encoder, an image encoder, a complex network cognition module, a reinforcement learning module, and a federated learning module, wherein: the measurement encoder is implemented by at least one processor and is configured to obtain state quantities required by the complex network cognition module, which is implemented by the at least one processor, and the reinforcement learning module, which is implemented by the at least one processor, the state quantities required by the complex network cognition module comprise a x-coordinate, a y-coordinate, a heading angle change and a speed of a given driving agent of a plurality of driving agents, the state quantities required by the complex network cognition module are handed over to the complex network cognition module as inputs, the state quantities required by the reinforcement learning module comprise a steering wheel angle, a throttle, a brake, a gear, a lateral speed and a longitudinal speed, the state quantities required by the reinforcement learning module are given to the reinforcement learning module after extracting features from a two-layer fully connected network; the image encoder is implemented by the at least one processor and is configured to obtain an amount of image implicit state required by the reinforcement learning module, an image used is a 15-channel semantic bird's eye view (BEV), i.sub.RL[0, 1].sup.192*192*15, 192 is in pixels and the BEV used is 5 px/m, 15 channels contain a drivable domain, a desired path, a road edge, 4 frames of other vehicles, 4 frames of pedestrians, and 4 frames of traffic signs, wherein the desired path is calculated using an A* algorithm, the semantic BEV is extracted by multilayer convolutional layers to extract implicit features and then passed to the reinforcement learning module as another part of the inputs; the complex network cognition module is configured to model a driving situation of a driving subject, and to obtain a maximum risk value of the driving subject in a current driving situation according to the state quantity provided by the measurement encoder, and finally to output dynamic driving suggestions based on a risk value through an activation function; the reinforcement learning module is configured to integrate the state quantities output from the measurement encoder and the image encoder, output corresponding strategies according to integrated network inputs, and interact with an environment to generate experience samples stored in a local replay buffer in the federated learning module, which is implemented by the at least one processor, when a number of the experience samples reaches a certain threshold, a batch of sample is taken from the local replay buffer for training, and finally trained neural network parameters are uploaded to the federated learning module; and the federated learning module is configured to receive the finally trained neural network parameters uploaded by the reinforcement learning module of the plurality of driving agents, and to aggregate a set of global parameters based on the finally trained neural network parameters, and finally to send the global parameters to the plurality of driving agents until a neural network converges, a global parameter aggregation is performed by a following equation:
11. The vehicular device according to claim 10, wherein in the complex network cognition-based FRL end-to-end autonomous driving control system, a modeling process of the complex network cognition module constructs a dynamic complex network model with the traffic participant and the road infrastructure as nodes;
12. The vehicular device according to claim 11, wherein in the complex network cognition-based FRL end-to-end autonomous driving control system, a Gaussian function is used in the dynamic complex network to reveal a static property between nodes:
13. The vehicular device according to claim 12, wherein in the complex network cognition-based FRL end-to-end autonomous driving control system, a series of isofield lines are used to delineate the safety field, a top view projection of the series of isofield lines is a region covered by a series of ellipses, with a region covered by a smallest center ellipse being a core domain, a region between the smallest center ellipse and a second ellipse being a limited domain, and a region between the second ellipse and a largest ellipse being an extended domain, size and shape of the region are determined by the series of isofield lines, are related to a vehicle shape and a motion state, and are described based on the Gaussian function, a direction of the safety field is aligned with a direction of vehicle motion, when a vehicle is in motion, the risk center O(x.sub.0, y.sub.0) of the safety field will be transferred to a new risk center O(x.sub.0, y.sub.0): {(1, 0)(0, 1)}, k.sub.v's positive or negative is related to the direction of vehicle motion, and denotes an angle of a transfer vector k.sub.v|{right arrow over (v)}| with axes in a Cartesian coordinate system, a virtual vehicle, with a vehicle length l.sub.v and a vehicle width w.sub.v, is formed under the risk center transfer, a dynamic safety field:
14. The vehicular device according to claim 13, wherein in the complex network cognition-based FRL end-to-end autonomous driving control system, a risk perception is categorized into three types on a planar scale based on different levels of human driving reaction time: a first cognitive domain, a second cognitive domain and an extra-domain space, wherein: the first cognitive domain:
15. The vehicular device according to claim 10, wherein in the complex network cognition-based FRL end-to-end autonomous driving control system, the CARLA simulator is used as an interaction environment, the CARLA simulator realizes vehicle control by inputting control quantities of a steering, the throttle and the brake, wherein steering [1, 1], throttle [0, 1] and brake [0, 1], based on a CARLA simulator's control method, a reinforcement learning action space [1, 1].sup.2, is categorized into the steering and a throttle-brake, when outputting the throttle-brake, [1, 0] denotes the brake and [0, 1] denotes the throttle, the driving control system outputs two parameters of the beta distribution by the reinforcement learning module, and then obtains a policy action by sampling:
16. The vehicular device according to claim 10, wherein in the complex network cognition-based FRL end-to-end autonomous driving control system, in a training process, a parameter updating is performed through following loss functions for the reinforcement learning module: .sub.ppo denotes a clipped policy gradient loss with advantages estimated using a generalized advantage estimation,
.sub.exp denotes a maximum entropy loss, H(.sub.(.Math.|i.sub.RL, m.sub.RL)) denotes an entropy of a policy .sub. under an image input i.sub.RL and a measurement input m.sub.RL, and
(1, 1) denotes a uniform distribution,
.sub.exp encourage the given driving agent to explore by converging an action distribution to the uniform distribution, .sub.exp denotes a weight of the maximum entropy loss,
.sub.risk denotes dynamic risk suggestions based loss, and
.sub.{TN.sub.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1)
(2)
(3)
(4)
(5)
(6)
DETAILED DESCRIPTION OF THE EMBODIMENTS
(7) The technical solution of the present disclosure is described in detail below in conjunction with the drawings, but is not limited to the contents of the present disclosure.
(8) The present disclosure provides a complex network cognition-based FRL end-to-end algorithmic framework that enables autonomous driving under dense urban traffic, specifically including the following steps:
(9) (1) The framework of FRL algorithm based on complex network cognition is built in CARLA simulator, as shown in
(10) (2) Model the driving situation of the driving subject, as shown in
(11)
(12) Where, G.sub.t denotes the dynamic complex network at the moment t, P={p.sub.1, p.sub.2, . . . , p.sub.N} is the set of nodes, and the number of nodes is N; E={e.sub.1,2, e.sub.1,3, . . . , e.sub.i,j} is the set of edges, and the number of edges is
(13)
and e.sub.i,j stands for the connectivity between nodes p.sub.i and p.sub.j; W={w.sub.1,2, w.sub.1,3, . . . , w.sub.i,j} is the set of weights of the edges, w.sub.i,j represents the coupling strength between nodes p.sub.i and p.sub.j; is the active region of the nodes, representing the dynamic constraints on the nodes in the network. Model as a smooth bounded surface:
(14)
(15) Where, is the boundary of the representation of the slip surface. Consider a continuous time dynamic network with N nodes on with a node state equation of the form:
(16)
(17) Where, X.sub.iR.sup.m denotes the state vector of node p.sub.i, R.sup.m denotes the vector space consisting of m-dimensional real numbers R, U.sub.iR.sup.q is the input vector, R.sup.q denotes the vector space consisting of q-dimensional real numbers R, A.sub.i denotes the dynamics matrix, B.sub.i denotes the input matrix. Based on the node state equation, the output vector of node p.sub.i can be obtained:
(18)
(19) Where, f.sub.i denotes the output function of the node. Then the weight function between nodes p.sub.i and p.sub.j is:
(20)
(21) Where, F denotes the weight function between the nodes. The present disclosure uses a Gaussian function to reveal the static properties between nodes:
(22)
(23) Where, S.sub.sta denotes the static field strength, C.sub.a denotes the field strength coefficients, x.sub.0 and y.sub.0 denote the coordinates of the risk center O(x.sub.0, y.sub.0), and a.sub.x and b.sub.y denote the vehicle appearance coefficients, respectively. The safety field is characterized by shape anisotropy:
(24)
(25) Where, is the aspect ratio, l.sub.v denotes the vehicle length, and w.sub.v denotes the vehicle width. The safety field is delineated by a series of isofields, and the top view projection is the region covered by a series of ellipses as shown in
(26) When the vehicle is in motion, the risk center O(x.sub.0, y.sub.0) of the safety field will be transferred to a new risk center O(x.sub.0, y.sub.0):
(27)
(28) Where, k.sub.v denotes the moderating factor and k.sub.v{(1, 0)(0,1)}, k.sub.v's positive or negative is related to the direction of motion, and denotes the angle of the transfer vector k.sub.v|{right arrow over (v)}| with the axes in the Cartesian coordinate system. A virtual vehicle, with length l.sub.v and width w.sub.v, is formed under the risk center transfer. The dynamic safety field:
(29)
(30) Where, S.sub.dyn denotes the dynamic field strength and the new aspect ratio is denoted as =a.sub.x/b.sub.y=l.sub.v/w.sub.v. As the motion state of the vehicle changes, the shape of the Gaussian safety field changes, thus changing the three fields covered by the safety field: the core region, the limited region and the extended region.
(31) The present disclosure categorizes risk perception into three layers on a planar scale based on different levels of human driving reaction time: the first cognitive domain, the second cognitive domain and the extra-domain space.
(32) The first cognitive domain:
(33)
(34) The second cognitive domain:
(35)
(36) The extra-domain space:
(37)
(38) Where, s.sub.th1 denotes the first cognitive domain threshold, obtained from the human driving the first reaction time t.sub.c1 and the maximum approach speed v.sub.e of the other nodes relative to the self-vehicle. s.sub.th2 denotes the second cognitive domain threshold, obtained from the human driving the second reaction time t.sub.c2 and the maximum approach speed v.sub.e of the other nodes relative to the self-vehicle.
(39) Establish a risk perception function between nodes under a variable safety field model:
(40)
(41) Where, |{right arrow over (S.sub.i,j)}| denotes the field strength of node p.sub.i at node p.sub.j, k.sub.c denotes the risk-adjustment cognitive coefficient, |{right arrow over (v.sub.j)}| denotes the scalar velocity of node p.sub.j, and .sub.i,j denotes the angle (positive in the clockwise direction) between the velocity vector {right arrow over (v.sub.j)} of node p.sub.j and the field strength vector {right arrow over (S.sub.i,j)}. The risk value Risk(p.sub.i, p.sub.j), obtained through the risk perception function, indicates the coupling strength between the nodes, and the higher the risk value, the higher the coupling strength, implying the higher correlation between the nodes.
(42) The activation function is configured to map the risk value, Activate(Risk) represents different activation functions according to different driving suggestions, and the mapped risk value will be used as the basis for guiding the output strategy of the reinforcement learning module:
(43)
(44) Where, Activate.sub.go(Risk) denotes the activation function when the suggestion is forward, Activate.sub.stop(Risk) denotes the activation function when the suggestion is stop, and Risk denotes the current risk value of the self-vehicle. The dynamic risk suggestion B.sub.risk:
(45)
(46) Where, B denotes the beta distribution with .sub.stop=.sub.go=1.
(47) (3) Construct the reinforcement learning model of the driving subject. Based on CARLA's control method, the reinforcement learning action space [1, 1].sup.2, are categorized into steering wheel corner and throttle brake. When outputting the throttle brake, [1,0] denotes the brake and [0,1] denotes the throttle. The present disclosure outputs the two parameters of the beta distribution by reinforcement learning, and then obtains the policy actions by sampling:
(48)
(49) In contrast to Gaussian distributions, are commonly used for model-free reinforcement learning, beta distributions are bounded and do not require mandatory constraints.
(50) The reward function setting, as shown in
(51)
(52) Where, r.sub.speed denotes the speed-related reward function, r.sub.position denotes the position-related reward function, r.sub.action denotes the action-related reward function, r.sub.terminal denotes the termination state-related reward function, v denotes the vehicle speed, v.sub.desire denotes the desired speed, and v.sub.max=6 m/s denotes the maximum speed. d denotes the vehicle lateral distance from the desired path, and denotes the angle between the vehicle traveling direction and the tangent line of the desired path. Table 1 describes the values of r.sub.action and r.sub.terminal in detail, where steering denotes the amount of steering wheel angle change in two frames.
(53) Construct an end-to-end neural network, as shown in
(54) (4) The driving subjects interact with the CARLA simulation environment and the experiences are stored in their respective local replay buffers. As shown in
(55)
(56) Where, .sub.ppo denotes the clipped policy gradient loss with advantages estimated using generalized advantage estimation.
.sub.exp denotes the maximum entropy loss, H(.sub.(.Math.|i.sub.RL, m.sub.RL)) denotes the entropy of the policy .sub. under the image input i.sub.RL and the measurement input m.sub.RL, and
(1, 1) denotes the uniform distribution.
.sub.exp encourage the agent to explore by converging the action distribution to a uniform distribution, .sub.exp denotes the weight of the maximum entropy loss.
.sub.risk denotes the loss based on the dynamic risk suggestions, and
.sub.{TN.sub.
(57) (5) The federated learning module is configured to receive the neural network parameters uploaded by the reinforcement learning module of each agent, and to aggregate the global parameters based on the plurality of neural network parameters, and finally to send the global parameters to each agent until the network converges. The global parameter aggregation is performed by the following equation:
(58)
(59) Where, .sub.m* denotes the global parameters at time m, N denotes the number of agents, and .sub.m.sup.n denotes the neural network parameters at time m of the nth agent.
(60) Overall, the present disclosure proposes FLDPPO, a complex network cognition based FRL algorithmic framework for urban autonomous driving in dense traffic scenarios. The FLDPPO algorithm realizes the combination of rule-based complex network cognition and end-to-end FRL by designing the loss function. The dynamic driving suggestions guide the agents to learn the rules, enabling them to cope with complex urban driving environments and dense traffic scenarios. The present disclosure introduces federated learning to train models by the method of parameter aggregation. The federated learning architecture accelerates network convergence, reduces communication consumption. The multi-agent architecture in the algorithmic framework not only improves the sample efficiency, but the trained models also exhibit high robustness and generalization.
(61) The present disclosure also proposes a vehicular device, the vehicular device being capable of executing the contents of the complex network cognition-based FRL end-to-end autonomous driving control system, or complex network cognition-based FRL end-to-end autonomous driving control method.
(62) A serious of detailed descriptions above are only specific descriptions of the practicable mode of implementation of the present disclosure and are not intended to limit the scope of protection of the present disclosure. Any equivalent mode or modification that does not depart from the technology of the present disclosure is included in the scope of protection of the present disclosure.