PANG Weijian ,LI Hui ,MA Xinyi,3,* ,and ZHANG Hailin
1.Academy of Military Sciences,Beijing 100091,China;2.Beijing Aeronautical Engineering Technology Research Center,Beijing 100076,China;3.Air Forces Command College,Beijing 100097,China
Abstract: Rich semantic information in natural language increases team efficiency in human collaboration,reduces dependence on high precision data information,and improves adaptability to dynamic environment.We propose a semantic centered cloud control framework for cooperative multi-unmanned ground vehicle (UGV) system.Firstly,semantic modeling of task and environment is implemented by ontology to build a unified conceptual architecture,and secondly,a scene semantic information extraction method combining deep learning and semantic web rule language (SWRL) rules is used to realize the scene understanding and task-level cloud task cooperation.Finally,simulation results show that the framework is a feasible way to enable autonomous unmanned systems to conduct cooperative tasks.
Keywords: scene understanding,cloud control,ontology,autonomous cooperation.
Semantic information is much more used than precise information in human collaboration.For example,when someone is asked to “put the package on the first desk to the left of the classroom door”,a semantic-centered inference process is carried out.Relying on semantic information such as <target location,on,desk>,<desk,leftOf,door>,<desk,firstOf,desks>,<classroom,has,desks>,<classroom,has,door>,the target location can be identified easily and the “package delivery task” can be completed without accurate spatial information.Semantic is an important element in intelligence.Knowledge with semantic information enables autonomous unmanned systems to be context-aware,spatial-time aware,and behavior aware.Knowledge graph is considered as one of the keys to intelligent decision-making [1].Hu et al.[2] considered that effective utilization of expert knowledge is a key issue for artificial intelligence transferring from the game world to the real battlefield.Defence Advanced Research Projects Agency (DARPA) also proposed a theoretical third-generation artificial intelligence framework that integrates data and knowledge in 2018.And Tang [3] summarized this theory as cognitive graph,which is knowledge graph plus cognitive inference plus logical expression.
Like human beings,autonomous unmanned systems also need the ability to express,efficiently access,and share semantics when they are performing a certain task so that autonomous unmanned systems could collaborate across platforms,systems,or even domains.Semantic information is used to make unmanned systems,such as robots[4],unmanned submarine vehicles [5],self-driving cars[6] and others understand scenes and complete tasks autonomously.However,these researches focus on intelligent control of a single platform.It can be more challengeable when the unmanned vehicles work collaboratively.
Therefore,we propose a semantic-centered cloud control framework for autonomous unmanned systems to realize unified cognition,knowledge sharing,and autonomous cooperation among unmanned vehicles connected to a shared cloud.
The rest of our paper is organized as follows.Section 2 introduces the related works including cloud robotics and knowledge representation of unmanned systems.Section 3 focuses on the design of ontology and the realization of inference engine.Section 4 provides the mechanism of mining deep scene information and triggering autonomous task coordinating.Section 5 simulates and verifies the proposed framework.Finally,Section 6 summarizes the paper.
Cloud robotics is known as the research field that integrates cloud computation and autonomous robotic systems [7].On one hand,cloud robotics has a centralized cloud that provides massive environment information acquired by robotics connected to the cloud.On the other hand,the cloud provides a shared library of capabilities,behaviors,states of robotics,and even the experiences of other robotics.Cloud robotics is developed from networked robotics and is in line with the trend of multiagent systems that provide intelligent services.The concept of cloud robotics brings about important applications in areas such as smart city [8],semantic sensor network [9],cloud manufacture [10],and cyber-physic system [11].
However,existing cloud control frameworks focus on data sharing [12],and the control mechanism is tightly coupled with the hardware connected to the cloud,they are more similar to shared databases rather than cloud control systems.In this paper,we focus on building a common and shared cognitive cloud for autonomous unmanned systems.That is the unified representation of knowledge,causal reasoning based on knowledge,and their implementation in task coordination among unmanned vehicles connected to the cloud during task execution.
Ontology is a formal and normative representation of concepts.The clarity and normative characteristics of ontology make it suitable for knowledge sharing and reuse[13].Ontology-based way of autonomy has gained attention in recent years.Miguelanez et al.[14] first introduced semantic knowledge information into mission planning,which improved the autonomy level of autonomous underwater unmanned vehicles.Ekvall et al.and Kragic [15,16] proposed a task-level planning system based on experiential activity schema,which realized task planning by configuring task goals and constraints.Mokhtari et al.[17] proposesd a conceptional method for autonomous robots that conceptualized task experience into the activity mode and applied it to task planning.Chen [18] proposed a robot task planning method that integrates activity schema with conceptional task experience based on ontology knowledge.Using ontology as a scene modeling method has the advantages of strong expression ability,supporting cognitive inference,high scalability,and loose coupling with hardware.However,most of the existing researches focus on single unmanned vehicle control and their interactivity with human.
In this paper,we focus on ontology’s ability in supporting autonomous task coordination of multi-unmanned ground vehicles (UGV) environment.We define unified concepts,relationships,and terminology,provide a unified knowledge representation for the cloud-based unmanned system,and realize the sharing at the knowledge level.The ontology-based semantic knowledge base (KB)as the carrier of the unmanned system’s long-term and short-term episodic memory makes it possible to infer based on task knowledge to support the unmanned system planning autonomously.
In this paper,we take heterogeneous unmanned system which includes carriers,fork lifters,and wreckers undertaking delivery tasks as an example to verify our framework.In this scenario,carriers need lift coordinating when loading package and may encounter roadblock or maintenance when delivering the package.Therefore,this is a collaborative task scenario in which unmanned vehicles need to understand the scene and undertake collaborative tasks autonomously.
In the framework,a central knowledge engine serves as a cloud,and the unmanned vehicles connected to the knowledge engine serve as nodes that may access the engine at any time.The knowledge engine and all nodes connected to it constitute the cloud control system.All data acquired by sensors is organized under a unified ontology so that all entities in the environment and all details about the collaborative delivery task are described as a semantic network.Then,the rule-based inference engines extract deep scene information to understand the situation they are in and take a suitable collaborative task.Finally,each node undertakes collaborative tasks under the plan domain describe language (PDDL) [19] planner autonomously.The framework is described as Fig.1.UXV means unmanned X vehicle.
Fig.1 Semantic-centered cloud control framework
To implement the proposed framework,the following work must be done: firstly,an ontology must be defined to store information acquired by sensor under a unified conceptual architecture as the common sense of swarmed unmanned system;secondly,a mechanism of mining deep information of sensor data is needed so that the unmanned vehicles may be aware of the situation they are in and perform tasks autonomously.Ontology is used in this framework to model domain knowledge such as expert knowledge and environmental knowledge acquired by the unmanned system itself.
The formal definition of ontology is
whereCrepresents a collection of classes andPrepresents a collection of predicates that describe object property or data property of a certain instance.Arepresents a collection of axioms,describing the relationships between classes,properties,and instances related relationships,such as inclusion,equivalence,reciprocity,transitivity,and symmetry.Irepresent a collection of instances,which is an instantiation of classes.Environment,mission,and UXV are three core ontologies of cloud control ontology.As shown in Fig.2,ontology mainly expresses knowledge from the perspective of task planning.Among them,environment ontology defines environment-related knowledge,such as artificial objects,natural objects,terrain,and weather;mission ontology defines task-related domain knowledge,such as tasks,behaviors,goals,states,and plans;UXV ontology defines unmanned vehicle related software and hardware knowledge,such as different types of unmanned vehicles,loads,capabilities,physical and geometric parameters,as well as software modules,algorithms.Besides,the cloud control ontology also includes other important ontologies such as universal instruction ontology which defines cooperative task instructions between unmanned systems,communication and network ontology which defines communication knowledge,space-time concept ontology,and concept and common sense ontology.
Fig.2 Task ontology
In addition,the ontology includes the rule that expresses knowledge from a logical perspective.The rule is a kind of causal knowledge expression method.For example,the capability attributes of a certain vehicle in UXV ontology correspond to an atom action in the task ontology,these two ontologies are connected through causal knowledge in the form of rules.
Protege [20] is used as an ontology modeling tool.It provides a graphical modeling interface,and the ontology can be formalized and stored using the web ontology language (OWL) based on the resource description framework (RDF).Good adaptability to the web environment of OWL makes it much more suitable for building an ontology for the cloud control system.
Ontology defines a framework that describes the entities involved in the task scenario.However,these pieces of information are static and conceptual,thus two kinds of predicates are used to describe the semantic relationship needed for scene understanding and autonomous task coordinating.One kind of predicates describes the relationship between different entities,such as “located-at” which describe the positional relationship like “vehicle located at a certain point”.Some predicates that describe entity relationships are shown in Table 1.The other kind of predicates describes the relationship between entities and their data properties,such as “hasState” which describe the state of a certain entity like “road segment #39 has state ‘LOCKED’”.Some predicates that describe such data properties are shown in Table 2.
Table 1 Entity property
Table 2 Data property
In the semantic centered cloud control framework,the system implements task-level cooperative control by the scene that vehicles are facing.The connection between the scene and the vehicle’s behavior is a semantic relationship.This kind of semantic relationship is represented by semantic web rule language (SWRL).Beyond this,SWRL is also used to represent relationships such as explicit data acquired by the sensor and their deep meanings.The process of discovering these two kinds of semantic relationships may be defined as situation awareness and scene understanding.Scene understanding is the process of mining deep hidden pieces of information behind the representation.Situation awareness is the process of triggering an unmanned vehicle’s behaviors or tasks when they are in a certain situation.The semantic representation of task domain knowledge and its principle of driving scene understanding and situation awareness is shown in Fig.3.
Fig.3 Semantic representation of task domain knowledge
A semantic knowledge engine is an interactive access system for semantic KB based on Jena Fuseki [21].Jena Fuseki is deployed in the cloud,and Fig.4 shows its structure.Two main interfaces for human-machine interaction are implemented: one is the reasoner interface,which is mainly used for importing rules;the other is the ontology interface,which is mainly responsible for adding or modifying descriptive pieces of knowledge.Vehicles as nodes connected to the cloud may access the semantic knowledge engine based on hyper text transfer protocol (HTTP) protocol,query or update information by SPARQL protocol and RDF query language (SPARQL)[22].The inference is implemented in the form of query for the KB.These queries fall into two categories: simple query and complex query.Simple query mainly queries for explicitly defined facts,that is data properties of a single entity (such as orientation,connectivity,and state).Complex query has two more circumstances.One is querying for inference results based on axioms and predefined SWRL rules.These rules must be strictly validated to ensure correctness.The other is querying for inference results based on temporary and local rules that may not be suitable in all situations.These queries are implemented in the form of SPARQL.
Fig.4 Diagram of the knowledge engine
When the cloud control system runs,all nodes keep a connection to the cloud by Algorithm 1.This algorithm registers the vehicles when they first connect to the cloud and keeps updating the state of the vehicles and data acquired by the sensor.These data are the raw material for scene understanding.When any node makes a certain query,the knowledge engine conducts inference based on these facts and the rules predefined.
The process of scene understanding is generally divided into two main stages [23–25].In the first stage,standard target detection networks such as region-convolution neural network (R-CNN) [26],faster-R-CNN [27],and You Only Look Once (YOLO) [28].are used to identify the object and obtain the bounding box of the object according to the input image.Then,in the next stage,the object is instantiated as instances,and the object’s bounding box features,object labels,and spatial coordinates are instantiated as instance features,and use long short-term memory (LSTM) [29],gated recurrent unit (GRU) [30],TreeLSTM [31],etc.to predict the relationship between objects.Predicates are usually from specific vectors such as Word2Vec [32],and ConceptNet [33].In this study,the predicate vector is obtained from the task ontology rather than a certain data set.For the problem of relationship prediction,a hybrid scene understanding method that combines the advantages of deep learning and rule-based inference is proposed.Cloud control framework based on this method has the advantages of end-to-end learning,concise,efficient and quick adaption to different task scenarios as required.
YOLO series network has the advantages of fast and relatively high accuracy.In recent years,researchers have improved its performance by adding an additional pooling layer between the backbone network and the head.YOLO3-SPP adds the spatial pyramid pooling (SPP) [34]module to realize the multi-scale overlap of images,which improves the accuracy of the network in predicting images of different resolutions and objects of different sizes.This paper chooses YOLO3-SPP as the target detection network.
Extraction of semantic information needed by unmanned system tasks varies from that in the field of computer vision which pursues rich picture information.Spatial parameters and spatial semantic relations are crucial to the execution of tasks.Therefore,after completing basic target detection and image recognition,it is necessary to estimate and infer the spatial parameters and spatial semantic relations of entities in the scene.In the field of robotics,spatial parameter estimation relies heavily on stereo vision,depth information,or point cloud information.However,in the real mission scenario,the vehicles usually observe objects from a distance,and it is very difficult to obtain depth information and point cloud data.Therefore,this paper is based on the use of a monocular camera image to achieve spatial parameter estimation.
The monocular camera acquires a two-dimensional image,and it is difficult to estimate the distance using Euclidean distance directly.Inverse perspective mapping(IPM) is an effective method for realizing the conversion from two-dimensional pixel points to the three-dimensional actual position [35].In the conversion process,it only needs to know the camera’s position,height,angle of view,and the inherent parameters of the camera,which are available in advance.The relationship between the position in the two-dimensional image of the object and its position in the real world can be expressed as
whereuandvrepresent the number of pixels of the target recognition frame in the horizontal and vertical directions.Xw,Yw,Zwrepresent the position of the object in the world coordinate system.Kis the inherent parameter of the camera.RandTrespectively represent the rotation and translation matrices from the world coordinate system to the two-dimensional image coordinate system.
wherehis the height of the camera,fis the focal length,kuandkvare the numbers of horizontal and vertical pixels of the camera,sis the image zoom ratio,andcxandcyare used to correct the main optical axis of the image plane.
Assume that theZ-axis value of the target in world coordinates is 0,then the relationship between the world coordinate system and the image coordinate system is
Relative coordinates can be calculated by
Target detection and spatial parameter estimation obtain the basic physical dimensions and spatial parameters of environment entities,but these parameters cannot support the autonomous behavior of the cloud control system.Thus,as shown in Subsection 2.2,SWRL rules are used to mine deep semantic relationships behind the data,figure out the situation a single node is in,and take proper actions that are beneficial to achieving the swarm task goal.
SWRL is described as antecedent→consequent.
Both antecedent and consequent consist of zero or more atoms.SWRL as production rules may infer new information based on the information already known,so it is an efficient way to infer information that is hidden behind the data but is important to the cloud system’s autonomy behavior.For example,when a UGV encounters a fallen tree on road,the cloud control system may infer whether the road is blocked or not by the physical size of the obstacle and width of the road,and the road’s state then is the decision basis for cloud task coordinating,as shown in the following:
Tree(?tree)^ on(?tree,?road)^ Road(?road)^ hasWidth(?tree,?wk)^ hasWidth(?road,?wd)^ swrlb:divide(?wd,2)^ swrlb: greaterThan(?wk,?wd) →hasState(?road,BLOCKED)
Furthermore,based on the information inferenced,the cloud control system may recognize the situation it confronts,and perform further inference about the actions the system can take.For example,when the road is blocked,the cloud control system should dispatch another wrecker for obstacle cleaning task.The rule for triggering the obstacle cleaning task is shown as follows:
Carrier(?u)^ Wrecker(?w)^ Road(?r)^ Tree(?tree)^ at(?u,?r)^ hasState(?r,BLOCKED)^ on(?tree,?road)^ hasMass(?t,?m)^ hasMaxLift(?w,?l)^ swrlb:lessThan(?m,?l)→hasCandidateTask(?w,OBSTACLE_CLEANNING)
SWRL rules used for scene semantic information extraction are shown in Table 3.
Table 3 SWRL rules in scene understanding and situation awareness
On the whole,SWRL has the following functions,which are helpful for situation awareness:
(i) Link various attributes of environmental entities according to their inherent logic.
(ii) Link the physical attributes of environmental entities to their semantic descriptions.
(iii) Link the properties of environmental entities with the behaviors that can be imposed on them.
According to the assumption of the cloud control system[12],all nodes connected to the cloud should be intelligent enough to undertake the coordinative task.In this paper,we use a PDDL planner as the individual planner.The detailed cloud control system is shown in Fig.5.
Fig.5 Diagram of cloud control system
Knowledge of task is defined according to the PDDL language specifications and integrated into task ontology,so that knowledge inference may be integrated with task planning algorithms,enabling high-level logic inference and planning algorithms such as fast down (FD),and forward-chaining partial-order planning (POPF).
After inference,the problem file is transmitted to the individual PDDL planner of relevant vehicles.The planner completes plan generation,plan distribution,and action schedule.In the process of task execution,when the environment changes,for example,the task cannot be completed,or new task requirements are generated.The problem file would be regenerated and the planner is called to replan the task.
The test scenario is shown in Fig.6.Two delivery UGVs,a forklift,and a wrecker are in the simulation.The UGVs are equipped with a camera to detect targets.The UGVs use the A* based route navigation method for path planning.The simulation platform is Ubuntu 18.04,ROS Melodic,Gazebo 7.The semantic ontology model of the task,primary entities of environment,and SWRL rules have been built in advance and integrated into the knowledge engine based on Jena Fuseki.The Jena Fuseki server serves as the cloud control center,and both UGVs connect to the server are nodes of the cloud control system.The cloud control system assigns delivery UGV to conduct delivery tasks when a package arrives,and dispatchs the wrecker for obstacle cleaning task or forklift for loading tasks if needed.All the coordinating tasks are dispatched autonomously.
Fig.6 Simulated experiment scenario
To make the experiment easier,this scenario is designed as a coordinated delivery task in a closed and limited environment.The control center generates a delivery task when a certain package arrives.The delivery UGV that can perform the task goes to the warehouse to obtain the package and perform the delivery task.During the delivery,the UGV may encounter a fallen tree on the road which blocks the road,so the delivery task is aborted.The UGV acquires the sensor data,and synchronizes it to the knowledge engine.Cloud knowledge engine perceives the “blocked road” situation and generates the “Obstacle cleanning” task,and assigns it to a wrecker UGV.All the UGVs work coordinatively for faster delivery.
Summary of the plot of the experimental scene is shown in Table 4.
Table 4 Plot summary of the experimental scenario
To verify the target detection,a database is constructed based on the experimentally scenes.The categories of objects to be recognized are houses,roads,road signs,trees,and cars.In the constructed Gazebo 3D environment,the pictures of each kind of object under different angles are collected and the images are labeled with the Labelme[36].The network training results are shown in Fig.7.
Fig.7 Training results of YOLO network
The ontology model and inference function test is used to find whether the ontology model and rules properly meet the requirements of scene modeling.The test was carried out using the Drools plug-in in protege.After the construction of the ontology model,some virtual facts are added to the ontology for the test.
Four scenarios are tested: (i) Infer entity state based on sensor data;(ii) Infer the lift cooperative task when delivery UGV arrives at the warehouse;(iii) Infer road blocking state and obstacle type according to sensor data;(iv)Infer the cooperative task of cleaning obstacles when the delivery UGV encounters road closure.The test results are shown in Fig.8.
Fig.8 Logic test result of knowledge engine
According Subsection to 3.3,knowledge queries fall into three categories: simple query,SWRL based inference,and SPARQL based inference [37].An example of SPARQL based inference statement and result is shown in Fig.9(a).Response time of different queries is shown in Fig.9(b).It can be seen that the query type is the main factor affecting the time consuming,and the overall query time is relatively low,which can meet the real-time requirements of task planning.
Fig.9 Responsiveness of KB
When the delivery UGV encounters a fallen tree obstacle,the unmanned vehicle classifies the tree obstacle,estimates its size,location,etc.These data are updated to the cloud;then,when the UGV queries for the candidate task,the cloud knowledge engine inferences road state,tree mass.When the road is blocked,classifies tree is an instance of Obstacle class.
This process is scene understanding and its procedure is shown in Fig.10.
Fig.10 Extracting semantic knowledge of the scene
The cloud KB first inferences the deep scene information according to sensor data such as entity category,size,and location.Then,if the scene meets the situation predefined as SWRL rules,a cooperative task order may be obtained by querying for “hasCandidateTask” predicate,as shown in Fig.10(b).
The cloud KB provides an information-sharing mechanism that inspires swarm intelligence.When delivery UGV finds the road is blocked,this information is synchronized to the cloud,the follow-up delivery task would consider this condition when planning.Thus,four delivery tasks are designed to verify the advantage of the cloud control framework,task1-task4 corresponding to the delivery package for address #1-#4 in Fig.6(a).Task planning time and task execution time with and without cloud information sharing are shown in Fig.11.The scenario is that packages for address #1-#4 arrive sequentially,when performing task 1,the delivery UGV finds the road is blocked and the cloud control system dispatches a wrecker to clean the obstacle and the delivery UGV replanning its path,the blocking situation persists before task 4 starts.It can be seen that during the execution of task 2 and task 3,due to the realization of information sharing within the cloud system,the follow-up UGVs can avoid the blocked road.However,in the absence of information sharing,the UGVs could not be aware of the situation of the road segment being blocked until they arrive at the blocked road segment,which cost extra task planning time and execution time.The information-sharing mechanism provided by the knowledge engine optimizes the task performance of all vehicles as a whole and shows the feature of preliminary swarm intelligence.
Fig.11 Comparison of performance of task planning with and without knowledge sharing
In this paper,we propose a semantic-centered cloud control framework for heterogeneous cooperative autonomous unmanned systems.The semantic knowledge is represented by ontology,and a semantic knowledge engine is designed to express the domain knowledge which is applied to the cloud task coordination.Then a hybrid scene understanding method combining deep learning with rulebased knowledge reasoning is proposed,which extracts the knowledge required for task execution,and triggers autonomous behavior of the cooperative cloud system.In addition,natural language processing attributes are added for each scene to lay the foundation for the next step of realizing natural language-based task collaboration and human-machine interaction.Finally,simulation experiments verify the feasibility of the framework.
The semantic centered cloud control system is realized according to the knowledge-based method,and works as a knowledge parse system.These make it easier to be applied into other fields such as agriculture,industrial manufacture,or supplies delivery and medical evacuation in the military.This research shows a feasible way to realize the cognitive ability of autonomous unmanned systems at task-level.However,there is still much to do to deal with large-scale graph problems if the scenario gets more complicated.
Journal of Systems Engineering and Electronics2022年4期