Skip to main content
Erschienen in: Complex & Intelligent Systems 2/2023

Open Access 05.11.2022 | Original Article

AGDS: adaptive goal-directed strategy for swarm drones flying through unknown environments

verfasst von: Fakui Wang, Jialei Huang, Kin Huat Low, Zisen Nie, Tianjiang Hu

Erschienen in: Complex & Intelligent Systems | Ausgabe 2/2023

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

This paper aims to address a challenging problem of a drone swarm for a specific mission by reaching a desired region, through an unknown environment. A bio-inspired flocking algorithm with adaptive goal-directed strategy (AGDS) is proposed and developed for the drones swarmed across unknown environments. Each drone employs a biological visual mechanism to sense obstacles in within local perceptible scopes. Task information of the destination is only given to a few specified drones (named as informed agents), rather than to all other individual drones (uninformed agents). With the proposed flocking swarm, the informed agents operate collectively with the remaining uninformed agents to achieve a common and overall mission. By virtue of numerical simulation, the AGDS and non-adaptive goal-directed strategy (non-AGDS) are both presented and evaluated. Experiments by flying six DJI Tello quadrotors indoor are conducted to validate the developed flocking algorithm. Additional validations within canyon-like complicated scenarios have also been carried out. Both simulation and experimental results demonstrate the efficiency of the proposed swarm flocking algorithm with AGDS.
Hinweise

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Introduction

With the advent of the artificial intelligence era, unmanned aerial vehicles (UAVs) have become substantially more autonomous and intelligent, which have become popular for various tasks [1, 2]. However, a single aircraft is with limited capability and resources [3], which results in its low efficiency and success rate on complex tasks such as transportation of large-scale supplies. The multiple unmanned aerial system is an unmanned intelligent cluster inspired by the self-organizing behaviors of biological groups, which significantly improves task efficiency through cooperation and complementary capabilities. Therefore, the cooperative control technology of the aerial cluster with distribution, autonomy and robustness is a promising solution. The aerial swarm has numerous potential applications, such as monitoring missions [4], UAV-assisted wireless coverage [5], search and rescue [6], target tracking [7], cooperative exploration [8], air traffic [9]. However, some drones in a swarm of drones have limited perceptual and computational capabilities, making it difficult to obtain abundant information about the environment and their companions. It is a challenge to make a drone swarm without collision, while passing an unknown environment, by relying only on the local information obtained by each individual drone.
Although the swarm has advantages in number and function, these advantages bring great challenges to formation control. The phenomenon of collective motion in nature provides inspiration for solving this problem [10]. Although individual animals have limited perceptual ability and seldom obtain global information, they can form coordinated and orderly collective motion without apparent leaders. This magical and fascinating collective behavior comes from the local interaction between individuals in gregarious living beings. Currently, the earliest collective intelligence algorithm recognized by academic circles is the Boids model proposed by Reynolds [11], which is a microscopic agent-based model constructed from a phenomenological perspective. It follows the separation-alignment-cohesion (SAC) rules. Compared with centralized control, this decentralized self-organization control provides great advantages for the flying robot swarm [12]. The former depends on a central computing node, while the latter each agent is an independent computing node, thus improving the robustness to individual faults. Self-organization means that any global motion mode results from local decisions. The decisions of each agent are only related to a limited number of neighbors, independent of the aerial swarm size [13]. This is highly critical for drone swarm control, which means that we can deploy dozens or hundreds of drones with limited computing resources.
Although the self-organizing control strategy based on biological inspiration has many advantages, it still faces some challenges when applied to the aerial swarm operating in complex environments. A concerning problem is to translate the phenomena of collective motion into interpretable mathematical models and deploy them on the multi-robot system. Vásárhelyi et al. [14] proposed a flocking model for real drones and carried out field experiments with a self-organizing swarm of 30 drones. This work realized flocking flight and obstacle avoidance of the aerial swarm like flocks of birds in predefined confined environments. However, such studies focus on reproducing the collective behavior of social animals in nature without considering the application in unknown environments and complex tasks [15]. The visual perception mechanism that individuals use the visual cortex to obtain information provides inspiration for the flying robot swarm to operate in an unknown environment [16, 17]. Perception and operation of drones in unknown environments is difficult, thus we use a bionic vision mechanism to achieve perception of the unknown environment. Besides, in practical applications, it is not necessary or even impossible to ensure that all individuals are informed of desired region information, such as the aerial swarm operating in substantial interference or dangerous environments. In most animal groups, only a small number of individuals have knowledge of food source or migration route [18, 19].
In this paper, we focus on the problem of a swarm of drones traversing an unknown environment without collision. A novel adaptive goal-directed bio-inspired flocking algorithm is designed. Each drone in the swarm works independently as a computing node and cooperates with others through local information interaction. Instead of prior knowledge of the environment, drones perceive the environment in real time through visual mechanisms. The flocking swarm is composed of informed agents and uniformed agents that operate equally in the overall missions. In addition, each drone moves with the same self-organizing rules. We propose an adaptive goal-directed strategy based on visual perception, which can significantly improve the efficiency of the drone swarm traversing the unknown environment. Finally, the feasibility of the proposed algorithm is verified by numerical simulation and physical experiments of 6 DJI Tello quadrotors.
The rest of this paper is organized as follows. “Related work” section covers some existing related works. The drone model, communication topology and evaluation metrics are presented in “Problem statement” section. In “Proposed flocking algorithm for drone swarm” section, the flocking algorithm suitable for the drone swarm traversing unknown environments is proposed, including non-AGDS and AGDS. In “Validation proposed flocking algorithms: typical scenario” section, the flocking algorithm is verified by simulation and experiment in a typical scenario. The validation and analysis of flocking algorithm in an extended scenario are given in “Extended simulation validation: complicated scenarios” section. Finally, this paper is concluded in “Conclusions” section.

Collective behavior modelling

The research on the mechanism of self-organizing behaviors promotes the development of the flying robot swarm. Vicsek et al. [20] studied the minimalist conditions of collective motion and proposed the Vicsek model, which only considered the velocity alignment rules between neighbors. Couzin et al. [21] proposed the Couzin model, which defined the zones and priorities of SAC rules. Levine et al. [22] proposed a social force model to study vortex motion based on Newtonian mechanics, which considered both velocity synergy and position synergy. Wu et al. [23] proposed an autonomous cooperative flocking algorithm for heterogeneous swarm, which solved the escort problem in the group cruise. By analyzing biological data, some scholars found unique behavior mechanisms, such as topological communication [24], hierarchical interaction [25], scale-free behavioral correlation [26], etc. Besides, the development of perceptual neuroscience has also promoted the study of collective biological behavior. Pearce et al. [17] proposed a model for long-range information exchange in bird flocks based on the projected view of each individual out through the flock and provided a solution for density control. Bastien et al. [27] proposed a purely vision-based model of collective behavior that showed the organized collective behavior that can be generated by relying solely on individual visual information. The model mainly focused on the perceptual information within the group without considering the perception of the environment. However, most of these works only explain the mechanism of collective motion from mathematical or biological data and have been seldom verified in artificial systems.

Reproducing natural collective motion with swarm robots

With the miniaturization of hardware, some scholars devoted themselves to deploying self-organizing algorithms in swarm robots. Slavkov et al. [28] created emergent morphologies in large swarm of real robots and reproduced the self-organized morphogenetic phenomenon of cells. Berlinger et al. [29] designed a fish-inspired robot swarm with implicit communication, which realized a variety of fish school behaviors such as synchrony, dispersion/aggregation, dynamic circle formation, and search-capture. Werfel et al. [30] developed a termite-inspired robot construction team construct complex predetermined structures. Rubenstein et al. [31] used a thousand-robot swarm to complete the self-assembly of complex two-dimensional shapes. Vásárhelyi et al. [14] deployed the flocking algorithm on 30 autonomous drones to achieve flocking flight and obstacle avoidance like a flock of birds. Balázs et al. [32] proposed a WillFull model for the persistence-responsivity trade-off in the swarm and used 52 autonomous drones to achieve abrupt collective turns like schools of fish or flocks of birds. Thus far, this is the largest self-organizing decentralized aerial swarm in an outdoor environment. The above works are mainly concerned with the reproduction of collective motions in nature on artificial collective systems.

Passing-through experiments using swarm robots and collective models

Based on the above research, swarm robots are expected to perform given missions in various scenarios. Quan et al. [33] proposed a distributed virtual pipeline control algorithm by pre-planning the airspace as virtual pipelines without obstacles. This work addressed the pipeline passage problem of the drone swarm and was experimentally verified using 6 multicopters. Liu et al. [34] developed a pragmatic distributed flexible formation protocol to pass through a narrow and irregular channel. In this study, two leaders tracked predefined channel boundaries, followers tracked leaders, and a field experiment was carried out with 3 unmanned surface vessels. Nevertheless, the premise of study is that the environment needs to be predefined and known to swarm individuals. Virágh et al. [35] provided a model of a general autonomous flying robot integrated into a realistic simulation framework and achieved collective target tracking using 9 autonomous aerial robots in the real world. Balázs et al. [36] proposed a decentralized air traffic control solution for dense traffic problems and implemented crosswalk and package-delivery applications on 30 autonomous drones. Soria et al. [37, 38] deployed the proposed predictive model on the multicopter swarm, which ensured rapid and safe collective migration in cluttered environments. Schilling et al. [39] proposed to learn vision-based swarm behaviors in an end-to-end manner directly from raw images and implemented collision-free navigation of the drone swarm in the real world. In these works, each agent in the drone swarm has knowledge of the desired region.
To sum up, the research on the self-organizing flocking algorithm of the aerial swarm in various applications is still in its infancy. This paper mainly solves the problem of the drone swarm traversing an unknown environment. Different from the above works, we propose a self-organizing flocking algorithm, which does not require every agent to have knowledge of the desired region. This is particularly beneficial for the drone swarm operating in interference or electromagnetic environments, where knowledge of the desired region cannot be guaranteed for each aircraft. Based on using the mathematical bionic vision mechanism to perceive the environment, we propose a vision-based adaptive goal-oriented strategy, which can effectively improve the efficiency of the drone swarm traversing a complex canyon-like environment.

Problem statement

In this section, a drone dynamics model is introduced first, including constraints on velocity and acceleration. Then, the communication topology of the drone swarm is presented. Finally, several evaluation metrics of the algorithm are proposed.

Individual drone dynamics

Consider aerial swarm consists of N individuals denoted by 1, 2, ..., N. The dynamics model of single drone can be simplified as a second-order integral model [40]. Therefore, we assume that the i-th drone with altitude hold mode satisfies the following model:
$$\begin{aligned} \begin{aligned} \varvec{\dot{p}}_i(t)&={\varvec{v}}_i(t) \\ \varvec{\dot{v}}_i(t)&= -\frac{1}{\tau _i}({\varvec{v}}_i(t)-{\varvec{v}}_{i\_\text {d}}(t)) \end{aligned} \end{aligned}$$
(1)
where \({\varvec{p}}_i(t) \in {\mathbb {R}}^2\) and \({\varvec{v}}_i(t) \in {\mathbb {R}}^2\) are respectively the position vector and velocity vector of the i-th drone at time t, \(\tau _i \in {\mathbb {R}}^+\) is the characteristic time of velocity response, \({\varvec{v}}_{i\_\text {d}}(t) \in {\mathbb {R}}^2\) is the desired velocity vector of the i-th drone at time t and is also the output of the flocking algorithm. We do not consider the differences between drones, and thus assume that \(\tau _i = \tau \) for all drones. Because of the physical constraints of the drone, the magnitude of velocity and acceleration are limited to \(v_{\max }\) and \(a_{\max }\), respectively. The rules are given by
$$\begin{aligned}{} & {} {\varvec{v}}_{i\_\text {d}}(t)=\frac{{\varvec{v}}_{i\_\text {d}}(t)}{\vert {\varvec{v}}_{i\_\text {d}}(t)\vert }\cdot \min \left\{ \vert {\varvec{v}}_{i\_\text {d}}(t)\vert , v_{\max } \right\} \end{aligned}$$
(2)
$$\begin{aligned}{} & {} {\varvec{a}}_i(t)=\frac{{\varvec{v}}_i(t) -{\varvec{v}}_{i\_\text {d}}(t)}{\vert {\varvec{v}}_i(t)-{\varvec{v}}_{i\_\text {d}}(t)\vert }\cdot \min \left\{ \frac{\vert {\varvec{v}}_i(t)-{\varvec{v}}_{i\_\text {d}}(t)\vert }{\delta t}, a_{\max } \right\} \nonumber \\ \end{aligned}$$
(3)
where \(\delta t\) is the interval of velocity updates.

Communication network between drones

To describe the model of the drone swarm in mathematical language, we use an undirected graph [41] \({\mathcal {G}}(t)=({\mathcal {V}},{\mathcal {E}}(t))\) to establish the communication network among N drones, where \({\mathcal {V}}=\{v_1,v_2,..,v_N\}\) is the set of vertices representing the set of drones, \({\mathcal {E}}(t) \in {\mathcal {V}} \times {\mathcal {V}}\) is the edge set that defines the communication at time t.
Definition 1
The neighbor set \({\mathcal {N}}_i(t)\) of i-th drone at time t is denoted by
$$\begin{aligned} {\mathcal {N}}_i(t)=\{v_j\in {\mathcal {V}}\vert (v_j,v_i)\in {\mathcal {E}}(t)\} \end{aligned}$$
(4)
Definition 2
The adjacency matrix \(\varvec{{\mathcal {A}}}_i(t)=[a_{ij}(t)]\in {\mathbb {R}}^{N \times N}\) represents the relationship between nodes, which is expressed as
$$\begin{aligned} {a_{ij}(t)} = {\left\{ \begin{array}{ll} 1,&{}{\text {if}}\ (v_j,v_i)\in {\mathcal {E}}(t) \\ {0,}&{}{\text {otherwise.}} \end{array}\right. } \end{aligned}$$
(5)
If drone i can communicate directly with drone j, \(a_{ij}(t)=1\), otherwise \(a_{ij}(t)=0\). Note that \({\mathcal {G}}(t)\) is an undirected graph, thus \(a_{ij}(t)=a_{ji}(t)\) and \(i \in {\mathcal {N}}_j(t) \Leftrightarrow j \in {\mathcal {N}}_i(t)\) hold.

Evaluation metrics for swarm flocking

We consider three different categories of metrics to evaluate the application of the drone swarm traversing an unknown environment.

Overall efficiency

Time and quantity are used to judge the overall efficiency of the drone swarm traversing the unknown environment. The task is completed if all the drones are located within the desired region.
\(T_{\text {end}}\) indicates the time that the drone swarm requires to complete the task.
\(N_{\text {end}}\) represents the number of drones located in the desired region within \(T_{\text {end}}\).
The task completion rate (TCR) is the ratio of the number of times of completed tasks to the total number of times of simulations.

Consensus

We use speed and direction to evaluate the consensus of the aerial swarm.
The normalized speed is expressed by
$$\begin{aligned} \Psi _{\text {vel}} = \frac{1}{T_{\text {end}}}\frac{1}{Nv_0}\int _{0}^{T_{\text {end}}} \left| \sum \limits _{i=1}^{N}{\varvec{v}}_i(t)dt\right| \end{aligned}$$
(6)
Note that \(\Psi _{\text {vel}}\) characterizes the degree to which the flocking speed tends to the preferred migration speed \(v_0\), and \(\Psi _{\text {vel}}=1\) only if the flocking speed is equal to \(v_0\).
The velocity correlation of the swarm can be evaluated by the following form:
$$\begin{aligned} \Psi _{\text {corr}} \!=\! \frac{1}{T_{\text {end}}}\frac{1}{N}\!\!\int _{0}^{T_{\text {end}}}\!\!\sum \limits _{i=1}^{N}\!\left( \!\frac{1}{N_i^{clu}\!-\!1}\!\!\sum \limits _{j\in C_i }^{N}\!\!\frac{{\varvec{v}}_i(t) \cdot {\varvec{v}}_j(t)}{\vert {\varvec{v}}_i(t)|| {\varvec{v}}_j(t)\vert }\!\right) dt\nonumber \\ \end{aligned}$$
(7)
where \(N_i^{clu}\) is the number of drones in the cluster containing the i-th drone. \(C_i\) is the set where the cluster excluding the i-th drone. Note that, the value of the velocity correlation, \(\Psi _{\text {corr}}\), are between \(-1\) (complete disorder) and 1 (perfect order).

Safety

The safety assessment of drones is divided into four parts: distance between drone and obstacle, distance between drones, minimum distance, and average distance between drones.
The ability of drones to avoid collisions with obstacles (or boundaries) is characterized by \(\Psi _{\text {obs}}\), which is governed by
$$\begin{aligned} \Psi _{\text {obs}} = \frac{1}{T_{\text {end}}}\frac{1}{N}\int _{0}^{T_{\text {end}}}\sum \limits _{i=1}^{N}{\mathcal {H}}(f({\varvec{p}} _i,t))dt \end{aligned}$$
(8)
where \(f(\cdot )=1\) when the position of the drone is inside the obstacle or outside wall, otherwise it is 0. On the other hand, \(\Psi _{\text {obs}}\) is equal to 0 when there is no collision, and it is equal to 1 when all agents collide. Next, \({\mathcal {H}}(x)\) is a heaviside step function, as follows:
$$\begin{aligned} {\mathcal {H}}(x) = \left\{ \begin{aligned}&1 ,&x \ge 0 \\&0 ,&x <0 \end{aligned} \right. \end{aligned}$$
(9)
\(\Psi _{\text {coll}}\) represents the ability of drones to avoid collision with each other, which can be evaluated by
$$\begin{aligned} \Psi _{\text {coll}} = \frac{1}{T_{\text {end}}}\frac{1}{N(N-1)}\int _{0}^{T_{\text {end}}}\!\sum \limits _{i=1}^{N}\!\sum \limits _{j\le i}{\mathcal {H}}(d_{ij}(t)-R)dt\nonumber \\ \end{aligned}$$
(10)
where \(d_{ij}=\vert {\varvec{p}}_i-{\varvec{p}}_j\vert \) is the distance between the i-th and j-th drones. If no collision is registered, \(\Psi _{\text {coll}}=0\), whereas \(\Psi _{\text {coll}}=1\), when all pairs of drones are colliding.
\(r_{\text {min}}\) and \(r_{\text {ave}}\) are the minimum distance and average distance between all drones, respectively.
Inspired by [14], we use genetic algorithm to find optimal parameter space that maximize \(\Psi _{\text {vel}}\) and \(\Psi _{\text {corr}}\), while minimizing \(\Psi _{\text {coll}}\) and \(\Psi _{\text {obs}}\).

Proposed flocking algorithm for drone swarm

In this section, the desired dynamics of agents are determined by neighbors \({\mathcal {N}}_i(t)\) and local environment. A flocking algorithm for the drone swarm traversing unknown environments is proposed first, including non-AGDS and AGDS. Then, each term in the flocking algorithm is introduced in detail.

General flocking law

This paper mainly solves the problem of the drone swarm traversing an unknown environment as a group under the condition that all agents cannot be guaranteed to have knowledge of the desired region. As shown in Fig. 1, the mathematical expression of the proposed flocking algorithm is described below:
$$\begin{aligned} {\varvec{v}}_{i\_\text {d}}= {\varvec{v}}_{i\_\text {sac}} + {\varvec{v}}_{i\_\text {v}} + (1-\lambda _i){\varvec{v}}_{i\_\text {w}} + \lambda _i{\varvec{v}}_{i\_\text {t}} \end{aligned}$$
(11)
where \({\varvec{v}}_{i\_\text {sac}}\) is the SAC interaction term that includes separation \(({\varvec{v}}_{i\_\text {s}})\), alignment \(({\varvec{v}}_{i\_\text {a}})\) and cohesion \(({\varvec{v}}_{i\_\text {c}})\), \({\varvec{v}}_{i\_\text {v}}\) is the visual perception term for obstacle avoidance, \(\lambda _i\) is the informed coefficient used to represent the role, \({\varvec{v}}_{i\_\text {w}}\) is the will propagation term and \({\varvec{v}}_{i\_\text {t}}\) is the task information term.
We define these drones equipped with task information as informed agents. The identities of informed agents are hidden, and feedback information needs to be obtained from neighbors. Therefore, the status of all agents is equal. It also means that the loss of part of the informed agents does not affect the information interaction within the flock. Of course, the informed agents can also be dynamically added, or the previous ordinary agents can be designated as informed agents.

Separation-alignment-cohesion interaction

We are inspired by a well-known model that allows the drone swarm to roam in confined environments [14]. The model includes the rule of separation to prevent collisions between drones and alignment to reduce velocity oscillations. In addition to this, we also include a cohesion rule to prevent drone swarms from splitting up. Therefore, we integrate these three basic requirements into one, as follows:
$$\begin{aligned} {\varvec{v}}_{i\_\text {sac}} = {\varvec{v}}_{i\_\text {s}} + {\varvec{v}}_{i\_\text {a}} + {\varvec{v}}_{i\_\text {c}} \end{aligned}$$
(12)
where \({\varvec{v}}_{i\_\text {s}}\), \({\varvec{v}}_{i\_\text {a}}\) and \({\varvec{v}}_{i\_\text {c}}\) define the requirements of separation, alignment, and cohesion, respectively. The three requirements are illustrated in Fig. 2.

Separation rule

Each drone needs to avoid collisions with other drones, which is one of the highest priority rules. Therefore, the separation among drones can be defined by
$$\begin{aligned} {\varvec{v}}_{i\_\text {s}}=p_{\text {sep}}\sum _{j=1}^{N}a_{ij}(r_{\text {sep}}-d_{ij})\cdot \frac{{\varvec{p}}_i-{\varvec{p}}_j}{d_{ij}}\cdot {\mathcal {H}}(r_{\text {sep}}-d_{ij})\nonumber \\ \end{aligned}$$
(13)
where \(p_{\text {sep}}\) is the linear gain of the collision avoidance effect, and \(r_{\text {sep}}\) is the radius of the separation zone.

Alignment rule

In actual flocking, the velocity alignment rule is essential. It can reduce oscillation and prevent dangerous situations by eliminating the excessively high-speed difference between nearby neighbors. The alignment rule is defined by
$$\begin{aligned} {\varvec{v}}_{i\_\text {a}}\!=\!p_{\text {ali}}\!\sum _{j=1}^{N}\!a_{ij}\left( v_{ij}\!-\!\Delta v^{\text {max}}_{ij}\right) \!\cdot \!\frac{{\varvec{v}}_i\!-\!{\varvec{v}}_j}{v_{ij}}\!\cdot \!{\mathcal {H}}\left( v_{ij}\!\!-\!\!\Delta v^{\text {max}}_{ij}\right) ,\nonumber \\ \end{aligned}$$
(14)
where \(p_{\text {ali}}\) is a linear gain of the velocity alignment effect, \(\Delta v^{\text {max}}_{ij}\) is the maximum allowable velocity difference in Ref. [14] and \(v_{ij}=\vert {\varvec{v}}_i-{\varvec{v}}_j\vert \) is the magnitude of the velocity difference between the i-th and j-th drones.

Cohesion rule

When the drone swarm traverses environments with obstacles, the drone swarm can easily be divided into multiple sub-swarms. If no individuals in some subgroups have the knowledge of the desired region, it will lead to drones in this subgroup hardly reaching the desired region. The cohesion rule is
$$\begin{aligned} \begin{aligned} {\varvec{v}}_{i\_\text {c}}=&p_{\text {coh}}\sum _{j=1}^{N}{a}_{ij}\left( d_{ij}\!-\!r_{\text {coh}}\right) \!\cdot \! \frac{{\varvec{p}}_j\!-\!{\varvec{p}}_i}{d_{ij}}\!\cdot \!{\mathcal {H}}\left( d_{ij}\!-\!r_{\text {coh}}\right) \end{aligned}\nonumber \\ \end{aligned}$$
(15)
where \(p_{\text {coh}}\) is the linear gain of cohesion and \(r_{\text {coh}}\) is the distance at which the cohesion rule starts to work. Note that \(r_{\text {coh}}\ge r_{\text {sep}}\).

Visual perception

In our model, we regard each drone as a disk with radius R. The vision projection field of the drone is a circle with a radius of \(r_{\text {vis}}\). The field projection of the drone i is divided into L parts, where l represents the label, \(l = 1,2,...,L\). Thus, line of sight \(\phi _{il}\) is abstracted as the connection between the center and the circumference (see Fig. 3a). Suppose the drone has a 360-degree field of view, but walls and obstacles will block its line of sight [17]. The angle between the two lines of sight is \(\Delta \Phi = 2\pi /L\). Obviously, the closer the robot is to walls and obstacles, the larger the obscured field of view. Therefore, we define the projection field function as follows (see Fig. 3b):
$$\begin{aligned} {V(\phi _i)} = {\left\{ \begin{array}{ll} 1,&{}{\text { if the line of sight is blocked}}\\ {0,}&{}{\text {otherwise.}} \end{array}\right. } \end{aligned}$$
(16)
As shown in Fig. 3c, the velocity \({\varvec{v}}_i\) direction of the drone i is aligned with the \({{\textbf {x}}}_v\) axis, while its counterclockwise normal direction is the \({{\textbf {y}}}_v\) axis. We define the upper and lower bounds of the occupied angle of the field of view projection field as \(\phi _{il}^{\text {upp}}\) and \(\phi _{il}^{\text {low}}\), respectively. Thus, the motion of the drone i based on visual mechanism is formulated as follows:
$$\begin{aligned} {\varvec{v}}_{i\_\text {v}}=\varvec{{\mathcal {R}}}\!\sum _{l=1}^{L}\left( \! \!-\!\sin \frac{\phi _{il}^{\text {upp}}\!-\!\phi _{il}^{\text {low}}}{2}\! \left[ \begin{matrix} p_{\text {fb}} \cos \frac{\phi _{il}^{\text {upp}}+\phi _{il}^{\text {low}}}{2}\\ p_{\text {lr}} \sin \frac{\phi _{il}^{\text {upp}}+\phi _{il}^{\text {low}}}{2} \end{matrix} \right] \! \right) \nonumber \\ \end{aligned}$$
(17)
where \(p_{\text {fb}}\) and \(p_{\text {lr}}\) are linear coefficients that characterize shifting and steering, respectively. Note that the rotation matrix to transform a vector from the body coordinate system \({{\textbf {O}}} _v- {} {{\textbf {x}}} _v{{\textbf {y}}} _v\) to the world coordinate system \({{\textbf {O}}}-{{\textbf {x}}}{{\textbf {y}}} \) is given by
$$\begin{aligned} \varvec{{\mathcal {R}}}=\left[ \begin{matrix} \cos \psi _i &{}\quad -\sin \psi _i \\ \sin \psi _i &{}\quad \cos \psi _i \end{matrix} \right] \end{aligned}$$
(18)
where \(\psi _i\) is the heading of i-th the drone expressed in the coordinate system \({{\textbf {O}}}-{{\textbf {x}}}{{\textbf {y}}} \).
The motion of the drone can be decomposed into radial motion and normal motion. The former can characterize the acceleration and deceleration behavior, and the latter can describe the left and right steering behavior [27]. Therefore, the drone generates avoidance behavior by sensing the size and position occupied by the projection of obstacles or walls in its field of view. As shown in Fig. 4, the vision term \({\varvec{v}}_{i\_v}\) can produce a short-range repulsion, where the first term of the matrix in Eq. (17) can have a repulsion in the front and back directions (Fig. 4a), and the second term of the matrix in Eq. (17) can produce a repulsion in the left and right directions (Fig. 4b).

Will propagation

Due to the limited communication and sensing capabilities of a single drone, it is difficult for the aerial swarm to quickly avoid obstacles when they encounter obstacles. Surprisingly, as the starling encounters natural enemies or obstacles, it will quickly spread the valuable information among the flocks through calls and other means to quickly escape or avoid obstacles [17]. Therefore, in the real environment, it is essential to transmit valuable information quickly between robots. For example, when a robot in front of the flock perceives an obstacle, it can quickly transmit this information to other robots so that the flock can quickly respond to obstacle avoidance instructions. To tackle such issues, a will propagation term \({\varvec{v}}_{i\_\text {w}}\) is introduced and expressed as
$$\begin{aligned} {\varvec{v}}_{i\_\text {w}}=v_0{\mathcal {W}}, \end{aligned}$$
(19)
where \({\mathcal {W}}\) is a unit vector, which indicates the direction of the will propagation, given by [32]
$$\begin{aligned} {\mathcal {W}}={\mathcal {N}}((1-\Omega _i){\mathcal {R}}[{\varvec{v}}_i;s_i^{WiSt}\delta t]+\Omega _iv_0\hat{{\varvec{v}}}_i^{WiSt}), \end{aligned}$$
(20)
where \(\hat{{\varvec{v}}}_i^{WiSt}\) is desired velocity of the WiSt model, \({\mathcal {N}}(\cdot )\) is the normalization operator, \({\mathcal {R}}[{\varvec{v}};\theta ]\) is the operator that rotates vector \({\varvec{v}}\) at angle \(\theta \) on the \({{\textbf {x}}}-{{\textbf {y}}}\) plane and \(s_i^{WiSt}\) is the intrinsic angular momentum. Note that \(\Omega _i\) can be used to distinguish between strong and weak will. Further discussion of Eq. (20) may refer to Ref. [32].

Task information with adaptive goal-directed strategy

Collective motions based on SAC self-organizing rules often show an aimless roaming phenomenon. Thus, we design a task information term to ensure that the drone can reach the desired region. The primary function of the term is to give knowledge of the desired region to the drones. The task information can be represented by the preferred direction \({\varvec{v}}_{\text {pre}}\), that is, a unit velocity vector pointing to the desired region. Therefore, the task information term is defined as
$$\begin{aligned} {{\varvec{v}}}_{i\_t} = v_0{\varvec{v}}_{\text {pre}} \end{aligned}$$
(21)
where \(v_0\) is the preferred speed. Note the non-adaptive goal-directed strategy (non-AGDS) represented by Eq. (21).
However, drones operating in unknown environments often encounter serious conflicts between task information and obstacle avoidance. This leads to a situation where the drone is caught in a dilemma. In nature, migrating or foraging animals adjust their behavior based on real-time information about the environment. Therefore, we propose an adaptive goal-directed strategy (AGDS) based on visual perception.
Firstly, the angle \(\Delta \phi _{\text {safe}}\) suitable for safe passage of drone is determined as follows:
$$\begin{aligned} \Delta \Phi _{\text {safe}} = 2\arctan \left( \frac{R}{r_{\text {vis}}}\right) \end{aligned}$$
(22)
We can then calculate the number \(n_{s}\) of the visual field occupied by the safe angle \(\Delta \Phi _{\text {safe}}\) as given by
$$\begin{aligned} n_{s} = \frac{\Delta \Phi _{\text {safe}}}{\Delta \Phi } \end{aligned}$$
(23)
The upper and lower bounds of the safe angle \(\Delta \Phi _{\text {safe}}\) in the projection field are \(\phi ^{\text {low}}_{\text {safe}}\) and \(\phi ^{\text {upp}}_{\text {safe}}\), respectively. Meanwhile, we determine the number \(n_{\theta }\) of areas in the visual field occupied by the angle \(\theta \) between the velocity \({\varvec{v}}_i\) and the goal-directed velocity \({\varvec{v}}_{\text {pre}}\). Next, we transform the original projection field \(V(\phi )\) (see Fig. 3a) starting with the velocity \({\varvec{v}}_i\) into the projection field \(V(\phi _{\text {task}})\) (see Fig. 5) starting with the goal-directed velocity \({\varvec{v}}_{\text {pre}}\) by rotating \(n_{\theta }\) regions equally.
According to whether there are obstacles in the line of sight of the task direction, there are two situations to deal with:
The first situation without obstacles is given by
$$\begin{aligned} \sum _{\phi ^{\text {low}}_{\text {safe}}}^{\phi ^{\text {upp}}_{\text {safe}}}V(\phi _{\text {safe}})=0, \end{aligned}$$
(24)
which implies that the line of sight is not blocked within the safe angle \(\Delta \Phi _{\text {safe}}\) range, then the task information term of the drone can be shown in Eq. (21).
Another situation with obstacles is expressed as
$$\begin{aligned} \sum _{\phi ^{\text {low}}_{\text {safe}}}^{\phi ^{\text {upp}}_{\text {safe}}}V(\phi _{\text {safe}})\ne 0, \end{aligned}$$
(25)
which implies that there is a conflict between preferred direction and obstacle avoidance. The goal orientation can be adaptively corrected to alleviate such a conflict. We define a cross-correlation function to represent the relationship between the passable direction and the projection field as follows:
$$\begin{aligned} V_c(\phi _{\text {task}})= & {} \frac{1}{\Delta \Phi _\text {safe}}\nonumber \\{} & {} \sum _{k=-n_{s}/2}^{n_{s}/2} {\varvec{{\mathcal {M}}}(k\Delta \Phi )V(\phi _{\text {task}}+k\Delta \Phi )\Delta \Phi } \end{aligned}$$
(26)
where \(\varvec{{\mathcal {M}}}(k\Delta \Phi )\in {\mathbb {R}}^{n_{s}}\) is a weighting matrix indicating the degree of security of the safe region \(\Delta \Phi _\text {safe}\) whose elements are all set as 1. Therefore, the value of \(V_c(\phi _{\text {task}})\) is in the range [0, 1]. Any region that satisfies \(V_c(\phi _{\text {task}})=0\) is a passable direction.
\(V_c(\phi _{\text {task}})\) is divided into two parts (\(V_c^{\text {L}}\) and \(V_c^{\text {R}}\)) with the initial preferred direction \({\varvec{v}}_{\text {pre}}\) as the origin, and the passable directions on both sides are obtained:
$$\begin{aligned} \left\{ \begin{aligned} \phi _{im}^{\text {L}}=\arg \min (V_c^\text {L})\\ \phi _{im}^{\text {R}}=\arg \min (V_c^\text {R}) \end{aligned} \right. \end{aligned}$$
(27)
Convert the above passable directions in Eq. (27) to the projection field \(V(\phi )\) with velocity \({\varvec{v}}_i\) as the origin and calculate the sight line of the passable directions as follows:
$$\begin{aligned} \phi _{im}=\mathop {\arg \min }\limits _{\phi _{im}^{\text {task}}}[\vert \phi _{im}^{\text {task}}-n_{\theta }\vert ] \end{aligned}$$
(28)
where \(\phi _{im}^{\text {task}}\in [\phi _{im}^{\text {L}},\phi _{im}^{\text {R}}]\). Thus, the corresponding angle of the sight line \(\phi _{im}\) in the projection field \(V(\phi )\) is \(\Phi _{i\_\text {m}}\). The angle \(\Phi _{i\_\text {m}}\) is converted to the world coordinate system \({{\textbf {O}}}-{{\textbf {x}}}{{\textbf {y}}} \) as follows:
$$\begin{aligned} \Phi _i^{\text {task}} = \psi _i + \Phi _{i\_\text {m}} \end{aligned}$$
(29)
Finally, the task information with adaptive goal-directed strategy (AGDS) is defined as follows:
$$\begin{aligned} {{\varvec{v}}}_{i\_\text {t}} = v_0\left[ \begin{matrix} \cos \Phi _i^{\text {task}} \\ \sin \Phi _i^{\text {task}} \end{matrix} \right] \end{aligned}$$
(30)

Validation proposed flocking algorithms: typical scenario

In this section, a typical scenario is designed to verify the proposed non-AGDS and AGDS flocking algorithms. Simulations and experiments are given in the following section to show the effectiveness of the proposed method, where a video about simulations and experiments is available on supplemental movie1.

Simulation validation

Simulation setup and parameters

A typical scenario of six drones with \(r_{\text {cm}}\) = 2.0m, \(r_{\text {vis}}\) = 0.4 m, \(r_{\text {sep}}\) = 0.4 m, \(r_{\text {coh}}\) = 0.8 m and \(v_0\) = 0.3 m/s as shown in Fig. 6 is designed. In this typical scene of 4 m \(\times \) 6.5 m, there is an obstacle region of 2.6 m \(\times \) 1.2 m in front of the desired region. The initial positions of the drones are fixed in the start region. The parameters of the algorithm are \(p_{\text {sep}}\) = 1.00, \(p_{\text {ali}}\) = 1.12, \(p_{\text {coh}}\) = 0.15, \(p_{\text {fb}} = p_{\text {lr}}\) = 0.31.

Simulation results and comparisons

The task information terms of the flocking algorithm with non-AGDS and AGDS are represented by Eq. (21) and Eq. (30), respectively. Figure 7a shows the simulation results of the proposed flocking algorithm with non-AGDS. As the initial position of the informed agent is 3# or 4#, the drone swarm successfully reaches the desired region. As shown in Fig. 6, 3# and 4# are the closest positions to the passable area, which means that the informed agent can quickly escape from the effect of obstacles to reduce the conflict between task information and obstacle avoidance. In stark contrast, 1# and 6#, which are farthest from the passable area, are the initial positions that are most likely to fall into the conflicting region. And 2# and 5# are the initial positions of intermediate cases between these two extreme cases. Figure 7b shows the simulation results of the proposed flocking algorithm with AGDS. The initial position of the informed agent does not cause the drone swarm to fail to reach the desired region. In addition, TCR of the method is significantly higher than that of the flocking algorithm with non-AGDS. When task information conflicts with obstacle avoidance, AGDS enables the task information term to be adaptively modified according to the obstacle information in the field of vision of the informed agent. This greatly alleviates the conflict between task information and obstacle avoidance, and prevents the informed agent from constantly oscillating near obstacles, which is very important for robots.
Figure 8a shows the task completion time of the non-AGDS flocking algorithm when the informed agent is in different positions in simulation. Compared with 4# position, 3# is farther from the obstacle and less prone to conflict between task information and obstacle avoidance. Therefore, when the initial position of the informed agent is 3#, the time \(T_{\text {end}}\) it takes for the drone swarm to reach the desired region is shorter, only 43.4s. Compared with the flocking algorithm with non-AGDS, the flocking algorithm with AGDS reduces the time it takes for a swarm of drones to traverse an unknown environment by more than half. In Fig. 8b, the drone swarm can successfully reach the desired region within 30s, and the shortest time is only 22.8s.

Experimental validation

Experimental setup and parameters

We establish the physical environment shown in Fig. 9. We use 6 DJI Tello drones with commercial autopilots for physical experiments. Based on this DJI autopilot, the velocity of the Tello drone can track the velocity command (see Eq. (11)) generated by the flocking algorithm in a reasonable time. A motion capture system called OptiTrack is installed indoors. The optical motion capture cameras obtain the ground truth of the position and velocity of each drone by sensing the markers on the drone with infrared rays. The ground control station is connected to OptiTrack and these drones by a local network. The information transfer of the experimental system and the configuration of the drone are given in detail in Fig. 9a. The drone swarm takes off from the start region (green area), traverses the task area (gray area), and finally lands in the desired region (red area). Note that the drone swarm may face various adverse situations in the task area, such as obstacles and walls. The parameters of the algorithm are \(p_{\text {sep}}\) = 5 .00, \(p_{\text {ali}}\) = 1.12, \(p_{\text {coh}}\) = 0.15, \(p_{\text {fb}}=p_{\text {lr}}\) = 0.24.

Experimental results and comparisons

In indoor experiments, we set the experimental time threshold to 80s. Besides, due to the limitation of site area, we set the desired region as a closed region, which will lead to an uncertain deviation in the task completion time of each experiment. Figure 10a shows the experimental results of the proposed flocking algorithm with non-AGDS. In the experiment, we verify the conclusions from the simulation. Only when the initial position of the informed agent is 3# or 4#, the drone swarm can quickly reach the desired region. When the initial position of the informed agent is 1# or 6#, it is very easy to cause conflicts between task information and obstacle avoidance, resulting in the difficulty of the drone swarm to reach the desired region. Figure 10b shows the experimental results of the proposed flocking algorithm with AGDS. In this set of experiments, the drone swarms all reach the desired region under the action of the adaptive strategy.
Figure 11a shows the task completion time of the non-AGDS flocking algorithm when the informed agent is in different positions in experiment. When the initial position of the informed agent is respectively 3# and 4#, the drone swarm takes 47.77s and 51.43s to reach the desired region. Figure 11b shows the task completion time of the AGDS flocking algorithm when the informed agent is in different positions in experiment. It can be seen from the experimental results that the drone swarm arrives at the desired area within the experimental time threshold.

Further batch analysis on flocking performance

To assess the reproducibility of the results, we performed 15 stochastic simulations for each of the six initial positions of the informed agent and for the two strategies, and we report here the aggregated performance results (Fig. 12). The initial position of the informed agent has a great influence on the flocking algorithm with non-AGDS. When the initial position of the informed agent is 3#, both \(T_{\text {end}}=32.66\pm 7.52s\) and \(N_{\text {end}}=6\pm 0\) are significantly better than other initial positions. In contrast, the flocking algorithm with AGDS performs better on \(T_{\text {end}}\) and \(N_{\text {end}}\) (1#, \(T_{\text {end}}=30.05\pm 1.98s\); 2#, \(T_{\text {end}}=30.45\pm 3.00\); 3#, \(T_{\text {end}}=30.18\pm 3.64s\); 4#, \(T_{\text {end}}=23.53\pm 3.09s\); 5#, \(T_{\text {end}}=21.74\pm 0.15s\); 6#, \(T_{\text {end}}=25.87\pm 2.78s\); \(1\#\sim 6\)#, \(N_{\text {end}}=6.00\pm 0.00s\)).
Table 1
Aggregate performance of the informed agent at different initial positions in the canyon-like environment
Metrics
1#
2#
3#
4#
5#
6#
\(T_{\text {end}}(s)\)
60.57 ± 4.91
59.07 ± 7.37
58.02 ± 5.21
59.68 ± 5.16
58.49 ± 4.82
58.37 ± 3.53
\(N_{\text {end}}\)
6.00 ± 0.00
6.00 ± 0.00
6.00 ± 0.00
6.00 ± 0.00
6.00 ± 0.00
6.00 ± 0.00
\(\Psi _{\text {vel}}\)
0.98 ± 0.01
0.97 ± 0.02
0.96 ± 0.01
0.98 ± 0.01
0.99 ± 0.03
0.98 ± 0.01
\(\Psi _{\text {corr}}\)
0.94 ± 0.02
0.91 ± 0.03
0.88 ± 0.05
0.94 ± 0.03
0.81 ± 0.04
0.95 ± 0.01
\(\Psi _{\text {obs}}\)
0.00 ± 0.00
0.00 ± 0.00
0.00 ± 0.00
0.00 ± 0.00
0.00 ± 0.00
0.00 ± 0.00
\(\Psi _{\text {coll}}\)
0.00 ± 0.00
0.00 ± 0.00
0.00 ± 0.00
0.00 ± 0.00
0.00 ± 0.00
0.00 ± 0.00
\({r_{\text {min}}(m)}\)
0.41 ± 0.01
0.41 ± 0.01
0.41 ± 0.02
0.42 ± 0.01
0.42 ± 0.02
0.42 ± 0.01
\({r_{\text {ave}}(m)}\)
0.78 ± 0.03
0.80 ± 0.01
0.88 ± 0.26
0.84 ± 0.10
0.93 ± 0.19
0.81 ± 0.03
Regardless of the initial position of the informed agent, the drone swarm successfully reaches the desired region. This means that we can increase or decrease the informed agent during the flight, which will not cause the conflict between task information and obstacle avoidance. The aerial swarm deployed with flocking algorithm with AGDS also performed very well in tracking the preferred swarm speed (\(\Psi _{\text {vel}}=0.98\pm 0.01, 1.00\pm 0.02, 1.01\pm 0.01, 0.98\pm 0.01, 0.98\pm 0.03, 0.97\pm 0.01\)). In addition to this, the drone swarm has better consistency \(\Psi _{\text {corr}}\) during flight. In terms of safety, the minimum distance \(r_{\text {min}}\) and the average distance \(r_{\text {ave}}\) between drones of both strategies are in the safe range (\(>2R\)).

Extended simulation validation: complicated scenarios

In the real world, the drone swarm often operates in complex and dangerous environments. For example, exploring pathways and transporting supplies in post-disaster canyons. In such environment, there is no way to ensure that every drone has information about the desired region. To extend the proposed flocking algorithm with AGDS to practical transportation applications, we need to verify it in a more complicated scenario. Due to the limited coverage of indoor motion capture systems, we design a canyon-like environment in the simulation. The canyon-like environment is a 16m-long irregular flight area, and there are many obstacles in the area that conflicted with the task information.
Figure 13a shows the flight trajectories of the drone swarm when traversing the complex and irregular environment. As shown in Fig. 13b, it only takes 58.4s for the drone swarm to traverse the complex and irregular environment. The drone swarm maintains high normalized swarm speed \(\Psi _{\text {vel}}\) and velocity correlation \(\Psi _{\text {corr}}\) throughout the flight (\(\Psi _{\text {vel}}=0.94\pm 0.06\) and \(\Psi _{\text {corr}}=0.93\pm 0.09\)). In addition, there are no collisions between the drones (\(\Psi _{\text {coll}}=\Psi _{\text {obs}}=0\)), and the distance between drones is kept within a safe range (\(r_{\text {min}}=0.40\pm 0.01\) and \(r_{\text {ave}}=0.77\pm 0.12\)).
To assess the reproducibility of the results in the canyon-like environment, we performed 15 stochastic simulations for each of the six initial positions of the informed agent, and we report here the aggregated performance results. From Table 1, it is evident that, the proposed flocking algorithm with AGDS is independent of the initial position of the informed agent. The drone swarm takes about 59.03 ± 5.23s to travel the complex canyon-like environment. The drone swarm also performs well at velocity correlation and tracking preferred migration speed \(v_0\) (\(\Psi _{\text {corr}}=0.90\pm 0.03\) and \(\Psi _{\text {vel}}=0.98\pm 0.01\)). In addition, there are no collisions among the drones and a safe distance is maintained between the drones (\(r_{\text {min}}=0.42\pm 0.01 > 2R\)).

Conclusions

The work presented in this paper proposed and developed an adaptive goal-directed swarm flocking algorithm for specific missions flying through an unknown environment. Simulations and experiments of six drones have been conducted for validation on the flocking algorithm. The proposed AGDS algorithm possess greater feasibility and efficiency within a typically regular scenario and an extended canyon-like environment.
Next, this work employed the bio-inspired flocking algorithm into practical applications of traversing through unknown environments, rather than aimless collective flight in confined environments [14]. The flocking swarm is composed of informed agents and uniformed agents that operate collectively to accomplish an overall missions. Only the informed agents which are a small minority and know the desired region to go, but the remaining uninformed drones have no knowledge of the destination.The informed and uninformed agents follow identical self-organization rules and obtain feedback from neighbors within their communicable range.
The initial position of the informed agent practically affects the efficiency of the drone swarm traversing the unknown environment. The uninformed agent takes the drone group away from the conflict region through alignment and cohesion operators. On the other hand, the informed agent alleviates the conflict between task information and self-organizing rules through AGDS, which makes the proposed flocking algorithm successful regardless of the initial position.
Our future works will concentrate on a few more informed agents in swarm flocking and their influence on the traversing success ratio. Furthermore, the safety and risk analyses are required to extend the developed flocking algorithm to the Unmanned Aircraft System Traffic Management (UTM), especially in urban-like environments.

Acknowledgements

This work was supported by the National Natural Science Foundation of China under Grant 61973327. This collaborative research, findings presented, and interactions among teams are also relevant to the ATMRI, NTU, Singapore.

Declarations

Conflicts of interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Ethics approval

Not applicable.
Not applicable.
Not applicable.
Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://​creativecommons.​org/​licenses/​by/​4.​0/​.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Fußnoten
1
The simulations and experiments of this paper can be viewed in the supplemental video: https://​youtu.​be/​b-le3CI5LX0.​
 
Literatur
1.
Zurück zum Zitat Chen J, Sun J, Wang G (2022) From unmanned systems to autonomous intelligent systems. Engineering 12:16–19CrossRef Chen J, Sun J, Wang G (2022) From unmanned systems to autonomous intelligent systems. Engineering 12:16–19CrossRef
2.
Zurück zum Zitat Kunze L, Hawes N, Duckett T, Hanheide M, Krajník T (2018) Artificial intelligence for long-term robot autonomy: a survey. IEEE Robot Autom Lett 3(4):4023–4030CrossRef Kunze L, Hawes N, Duckett T, Hanheide M, Krajník T (2018) Artificial intelligence for long-term robot autonomy: a survey. IEEE Robot Autom Lett 3(4):4023–4030CrossRef
3.
Zurück zum Zitat Yan C, Xiang X, Wang C (2020) Towards real-time path planning through deep reinforcement learning for a uav in dynamic environments. J Intell Robot Syst 98(2):297–309CrossRef Yan C, Xiang X, Wang C (2020) Towards real-time path planning through deep reinforcement learning for a uav in dynamic environments. J Intell Robot Syst 98(2):297–309CrossRef
4.
Zurück zum Zitat Yu W, Shaobo W, Xinting H (2021) Multi-constrained cooperative path planning of multiple drones for persistent surveillance in urban environments. Complex Intell Syst 7(3):1633–1647CrossRef Yu W, Shaobo W, Xinting H (2021) Multi-constrained cooperative path planning of multiple drones for persistent surveillance in urban environments. Complex Intell Syst 7(3):1633–1647CrossRef
5.
Zurück zum Zitat Wang J, Jiang C, Han Z, Ren Y, Maunder R, Hanzo Lajos (2017) Taking drones to the next level: cooperative distributed unmanned-aerial-vehicular networks for small and mini drones. IEEE Veh Technol Mag 12(3):73–82CrossRef Wang J, Jiang C, Han Z, Ren Y, Maunder R, Hanzo Lajos (2017) Taking drones to the next level: cooperative distributed unmanned-aerial-vehicular networks for small and mini drones. IEEE Veh Technol Mag 12(3):73–82CrossRef
6.
Zurück zum Zitat Geng N, Chen Z, Nguyen QA, Gong D (2021) Particle swarm optimization algorithm for the optimization of rescue task allocation with uncertain time constraints. Complex Intell Syst 7(2):873–890CrossRef Geng N, Chen Z, Nguyen QA, Gong D (2021) Particle swarm optimization algorithm for the optimization of rescue task allocation with uncertain time constraints. Complex Intell Syst 7(2):873–890CrossRef
7.
Zurück zum Zitat Wu Y, Low KH (2022) Route coordination of uav fleet to track a ground moving target in search and lock (sal) task over urban airspace. IEEE Internet Things J Wu Y, Low KH (2022) Route coordination of uav fleet to track a ground moving target in search and lock (sal) task over urban airspace. IEEE Internet Things J
8.
Zurück zum Zitat McGuire KN, De Wagter C, Tuyls K, Kappen HJ, de Croon GCHE (2019) Minimal navigation solution for a swarm of tiny flying robots to explore an unknown environment. Sci Robot 4(35):eaaw9710CrossRef McGuire KN, De Wagter C, Tuyls K, Kappen HJ, de Croon GCHE (2019) Minimal navigation solution for a swarm of tiny flying robots to explore an unknown environment. Sci Robot 4(35):eaaw9710CrossRef
9.
Zurück zum Zitat Pang B, Low KH, Lv C (2022) Adaptive conflict resolution for multi-uav 4d routes optimization using stochastic fractal search algorithm. Transp Res Emerg Technol Part C 139:103666CrossRef Pang B, Low KH, Lv C (2022) Adaptive conflict resolution for multi-uav 4d routes optimization using stochastic fractal search algorithm. Transp Res Emerg Technol Part C 139:103666CrossRef
10.
Zurück zum Zitat Vicsek T, Zafeiris A (2012) Collective motion. Phys Rep 517(3–4):71–140CrossRef Vicsek T, Zafeiris A (2012) Collective motion. Phys Rep 517(3–4):71–140CrossRef
11.
Zurück zum Zitat Reynolds CW (1987) Flocks, herds and schools: a distributed behavioral model. In: Proc 14th Annu Conf Comput Graph Interactive Techn, pp 25–34 Reynolds CW (1987) Flocks, herds and schools: a distributed behavioral model. In: Proc 14th Annu Conf Comput Graph Interactive Techn, pp 25–34
12.
Zurück zum Zitat Wang Z, Yun X, Jiang C, Zhang Y (2019) Self-organizing control for satellite clusters using artificial potential function in terms of relative orbital elements. Aerosp Sci Technol 84:799–811CrossRef Wang Z, Yun X, Jiang C, Zhang Y (2019) Self-organizing control for satellite clusters using artificial potential function in terms of relative orbital elements. Aerosp Sci Technol 84:799–811CrossRef
13.
Zurück zum Zitat Yan C, Wang C, Xiang X, Lan Z, Jiang Y (2021) Deep reinforcement learning of collision-free flocking policies for multiple fixed-wing uavs using local situation maps. IEEE Trans Ind Informat 18(2):1260–1270CrossRef Yan C, Wang C, Xiang X, Lan Z, Jiang Y (2021) Deep reinforcement learning of collision-free flocking policies for multiple fixed-wing uavs using local situation maps. IEEE Trans Ind Informat 18(2):1260–1270CrossRef
14.
Zurück zum Zitat Vásárhelyi G, Virágh C, Somorjai G, Nepusz T, Eiben AE, Vicsek T (2018) Optimized flocking of autonomous drones in confined environments. Sci Robot 3(20):eaat3536CrossRef Vásárhelyi G, Virágh C, Somorjai G, Nepusz T, Eiben AE, Vicsek T (2018) Optimized flocking of autonomous drones in confined environments. Sci Robot 3(20):eaat3536CrossRef
15.
Zurück zum Zitat Dorigo M, Theraulaz G, Trianni V (2020) Reflections on the future of swarm robotics. Sci Robot 5(49):eabe4385CrossRef Dorigo M, Theraulaz G, Trianni V (2020) Reflections on the future of swarm robotics. Sci Robot 5(49):eabe4385CrossRef
16.
Zurück zum Zitat Strandburg-Peshkin A, Twomey CR, Bode NWF, Kao AB, Katz Y, Ioannou Christos C, Rosenthal Sara B, Torney Colin J, Hai Shan Wu, Levin Simon A et al (2013) Visual sensory networks and effective information transfer in animal groups. Curr Biol 23(17):R709–R711CrossRef Strandburg-Peshkin A, Twomey CR, Bode NWF, Kao AB, Katz Y, Ioannou Christos C, Rosenthal Sara B, Torney Colin J, Hai Shan Wu, Levin Simon A et al (2013) Visual sensory networks and effective information transfer in animal groups. Curr Biol 23(17):R709–R711CrossRef
17.
Zurück zum Zitat Pearce DJG, Miller AM, Rowlands G, Turner MS (2014) Role of projection in the control of bird flocks. Proc Natl Acad Sci 111(29):10422–10426CrossRef Pearce DJG, Miller AM, Rowlands G, Turner MS (2014) Role of projection in the control of bird flocks. Proc Natl Acad Sci 111(29):10422–10426CrossRef
18.
Zurück zum Zitat Strandburg-Peshkin A, Farine DR, Couzin ID, Crofoot MC (2015) Shared decision-making drives collective movement in wild baboons. Science 348(6241):1358–1361CrossRef Strandburg-Peshkin A, Farine DR, Couzin ID, Crofoot MC (2015) Shared decision-making drives collective movement in wild baboons. Science 348(6241):1358–1361CrossRef
19.
Zurück zum Zitat Couzin ID, Krause J, Franks NR, Levin SA (2005) Effective leadership and decision-making in animal groups on the move. Nature 433(7025):513–516CrossRef Couzin ID, Krause J, Franks NR, Levin SA (2005) Effective leadership and decision-making in animal groups on the move. Nature 433(7025):513–516CrossRef
20.
Zurück zum Zitat Vicsek T, Czirók A, Ben-Jacob E, Cohen I, Shochet O (1995) Novel type of phase transition in a system of self-driven particles. Phys Rev Lett 75(6):1226MathSciNetCrossRef Vicsek T, Czirók A, Ben-Jacob E, Cohen I, Shochet O (1995) Novel type of phase transition in a system of self-driven particles. Phys Rev Lett 75(6):1226MathSciNetCrossRef
21.
Zurück zum Zitat Couzin ID, Krause J, James R, Ruxton GD, Franks NR (2002) Collective memory and spatial sorting in animal groups. J Theor Biol 218(1):1–11MathSciNetCrossRef Couzin ID, Krause J, James R, Ruxton GD, Franks NR (2002) Collective memory and spatial sorting in animal groups. J Theor Biol 218(1):1–11MathSciNetCrossRef
22.
Zurück zum Zitat Levine H, Rappel W-J, Cohen I (2000) Self-organization in systems of self-propelled particles. Phys Rev E 63(1):017101CrossRef Levine H, Rappel W-J, Cohen I (2000) Self-organization in systems of self-propelled particles. Phys Rev E 63(1):017101CrossRef
23.
Zurück zum Zitat Jiehong W, Yuanzhe Y, Ma J, Jinsong W, Han G, Shi J, Gao L (2021) Autonomous cooperative flocking for heterogeneous unmanned aerial vehicle group. IEEE Trans Veh Technol 70(12):12477–12490CrossRef Jiehong W, Yuanzhe Y, Ma J, Jinsong W, Han G, Shi J, Gao L (2021) Autonomous cooperative flocking for heterogeneous unmanned aerial vehicle group. IEEE Trans Veh Technol 70(12):12477–12490CrossRef
24.
Zurück zum Zitat Ballerini M, Cabibbo N, Candelier R, Cavagna A, Cisbani E, Giardina I, Lecomte V, Orlandi A, Parisi G, Procaccini A et al (2008) Interaction ruling animal collective behavior depends on topological rather than metric distance: evidence from a field study. Proc Natl Acad Sci 105(4):1232–1237CrossRef Ballerini M, Cabibbo N, Candelier R, Cavagna A, Cisbani E, Giardina I, Lecomte V, Orlandi A, Parisi G, Procaccini A et al (2008) Interaction ruling animal collective behavior depends on topological rather than metric distance: evidence from a field study. Proc Natl Acad Sci 105(4):1232–1237CrossRef
25.
Zurück zum Zitat Nagy M, Ákos Z, Biro D, Vicsek T (2010) Hierarchical group dynamics in pigeon flocks. Nature 464(7290):890–893 Nagy M, Ákos Z, Biro D, Vicsek T (2010) Hierarchical group dynamics in pigeon flocks. Nature 464(7290):890–893
26.
Zurück zum Zitat Cavagna A, Cimarelli A, Giardina I, Parisi G, Santagati R, Stefanini F, Viale M (2010) Scale-free correlations in starling flocks. Proc Natl Acad Sci 107(26):11865–11870CrossRef Cavagna A, Cimarelli A, Giardina I, Parisi G, Santagati R, Stefanini F, Viale M (2010) Scale-free correlations in starling flocks. Proc Natl Acad Sci 107(26):11865–11870CrossRef
27.
Zurück zum Zitat Bastien R, Romanczuk P (2020) A model of collective behavior based purely on vision. Sci Adv 6(6):eaay0792CrossRef Bastien R, Romanczuk P (2020) A model of collective behavior based purely on vision. Sci Adv 6(6):eaay0792CrossRef
28.
Zurück zum Zitat Slavkov I, Carrillo-Zapata D, Carranza N, Diego X, Jansson F, Kaandorp J, Hauert S, Sharpe J (2018) Morphogenesis in robot swarms. Sci Robot 3(25):eaau9178CrossRef Slavkov I, Carrillo-Zapata D, Carranza N, Diego X, Jansson F, Kaandorp J, Hauert S, Sharpe J (2018) Morphogenesis in robot swarms. Sci Robot 3(25):eaau9178CrossRef
29.
Zurück zum Zitat Berlinger F, Gauci M, Nagpal R (2021) Implicit coordination for 3d underwater collective behaviors in a fish-inspired robot swarm. Sci Robot 6(50):eabd8668CrossRef Berlinger F, Gauci M, Nagpal R (2021) Implicit coordination for 3d underwater collective behaviors in a fish-inspired robot swarm. Sci Robot 6(50):eabd8668CrossRef
30.
Zurück zum Zitat Werfel J, Petersen K, Nagpal R (2014) Designing collective behavior in a termite-inspired robot construction team. Science 343(6172):754–758CrossRef Werfel J, Petersen K, Nagpal R (2014) Designing collective behavior in a termite-inspired robot construction team. Science 343(6172):754–758CrossRef
31.
Zurück zum Zitat Rubenstein M, Cornejo A, Nagpal R (2014) Programmable self-assembly in a thousand-robot swarm. Science 345(6198):795–799CrossRef Rubenstein M, Cornejo A, Nagpal R (2014) Programmable self-assembly in a thousand-robot swarm. Science 345(6198):795–799CrossRef
32.
Zurück zum Zitat Balázs B, Vásárhelyi G, Vicsek T (2020) Adaptive leadership overcomes persistence-responsivity trade-off in flocking. J R Soc. Interface 17(167):20190853CrossRef Balázs B, Vásárhelyi G, Vicsek T (2020) Adaptive leadership overcomes persistence-responsivity trade-off in flocking. J R Soc. Interface 17(167):20190853CrossRef
33.
Zurück zum Zitat Quan Q, Fu R, Li M, Wei D, Gao Y, Cai K-Y (2021) Practical distributed control for vtol uavs to pass a virtual tube. IEEE Trans Intell Veh, pp 1–1 Quan Q, Fu R, Li M, Wei D, Gao Y, Cai K-Y (2021) Practical distributed control for vtol uavs to pass a virtual tube. IEEE Trans Intell Veh, pp 1–1
34.
Zurück zum Zitat Liu B, Zhang H-T, Meng H, Dongfei F, Housheng S (2020) Scanning-chain formation control for multiple unmanned surface vessels to pass through water channels. IEEE Trans Cybern 52(3):1850–1861CrossRef Liu B, Zhang H-T, Meng H, Dongfei F, Housheng S (2020) Scanning-chain formation control for multiple unmanned surface vessels to pass through water channels. IEEE Trans Cybern 52(3):1850–1861CrossRef
35.
Zurück zum Zitat Virágh C, Vásárhelyi G, Tarcai N, Szörényi T, Somorjai G, Nepusz T, Vicsek T (2014) Flocking algorithm for autonomous flying robots. Bioinspir Biomim 9(2):025012CrossRef Virágh C, Vásárhelyi G, Tarcai N, Szörényi T, Somorjai G, Nepusz T, Vicsek T (2014) Flocking algorithm for autonomous flying robots. Bioinspir Biomim 9(2):025012CrossRef
36.
Zurück zum Zitat Balázs B, Vásárhelyi G (2018) Coordinated dense aerial traffic with self-driving drones. In: Proc IEEE Int Conf Robot Automat (ICRA), IEEE, pp 6365–6372 Balázs B, Vásárhelyi G (2018) Coordinated dense aerial traffic with self-driving drones. In: Proc IEEE Int Conf Robot Automat (ICRA), IEEE, pp 6365–6372
37.
Zurück zum Zitat Soria E, Schiano F, Floreano D (2021) Predictive control of aerial swarms in cluttered environments. Nat Mach Intell 3(6):545–554CrossRef Soria E, Schiano F, Floreano D (2021) Predictive control of aerial swarms in cluttered environments. Nat Mach Intell 3(6):545–554CrossRef
38.
Zurück zum Zitat Soria E, Schiano F, Floreano D (2021) Distributed predictive drone swarms in cluttered environments. IEEE Robot Autom Lett 7(1):73–80CrossRef Soria E, Schiano F, Floreano D (2021) Distributed predictive drone swarms in cluttered environments. IEEE Robot Autom Lett 7(1):73–80CrossRef
39.
Zurück zum Zitat Schilling F, Lecoeur J, Schiano F, Floreano D (2019) Learning vision-based flight in drone swarms by imitation. IEEE Robot Autom Lett 4(4):4523–4530CrossRef Schilling F, Lecoeur J, Schiano F, Floreano D (2019) Learning vision-based flight in drone swarms by imitation. IEEE Robot Autom Lett 4(4):4523–4530CrossRef
40.
Zurück zum Zitat Quan Q, Fu R, Cai K-Y (2021) Practical control for multicopters to avoid non-cooperative moving obstacles. IEEE Trans Intell Transp Syst, pp 1–19 Quan Q, Fu R, Cai K-Y (2021) Practical control for multicopters to avoid non-cooperative moving obstacles. IEEE Trans Intell Transp Syst, pp 1–19
41.
Zurück zum Zitat Mesbahi M, Egerstedt M (2010) Graph theoretic methods in multiagent networks. Princeton University Press, PrincetonCrossRefMATH Mesbahi M, Egerstedt M (2010) Graph theoretic methods in multiagent networks. Princeton University Press, PrincetonCrossRefMATH
Metadaten
Titel
AGDS: adaptive goal-directed strategy for swarm drones flying through unknown environments
verfasst von
Fakui Wang
Jialei Huang
Kin Huat Low
Zisen Nie
Tianjiang Hu
Publikationsdatum
05.11.2022
Verlag
Springer International Publishing
Erschienen in
Complex & Intelligent Systems / Ausgabe 2/2023
Print ISSN: 2199-4536
Elektronische ISSN: 2198-6053
DOI
https://doi.org/10.1007/s40747-022-00900-9

Weitere Artikel der Ausgabe 2/2023

Complex & Intelligent Systems 2/2023 Zur Ausgabe

Premium Partner