IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 1 Graph Learning: A Survey Feng Xia, Senior Member, IEEE, Ke Sun, Shuo Yu, Member, IEEE, Abdul Aziz, Liangtian Wan, Member, IEEE, Shirui Pan, and Huan Liu, Fellow, IEEE Abstract —Graphs are widely used as a popular representation of the network structure of connected data. Graph data can be found in a broad spectrum of application domains such as social systems, ecosystems, biological networks, knowledge graphs, and information systems. With the continuous penetra- tion of artificial intelligence technologies, graph learning (i.e., machine learning on graphs) is gaining attention from both researchers and practitioners. Graph learning proves effective for many tasks, such as classification, link prediction, and matching. Generally, graph learning methods extract relevant features of graphs by taking advantage of machine learning algorithms. In this survey, we present a comprehensive overview on the state-of-the-art of graph learning. Special attention is paid to four categories of existing graph learning methods, including graph signal processing, matrix factorization, random walk, and deep learning. Major models and algorithms under these categories are reviewed respectively. We examine graph learning applications in areas such as text, images, science, knowledge graphs, and combinatorial optimization. In addition, we discuss several promising research directions in this field. Index Terms —Graph learning, graph data, machine learning, deep learning, graph neural networks, network representation learning, network embedding. IMPACT STATEMENT Real-world intelligent systems generally rely on machine learning algorithms handling data of various types. Despite their ubiquity, graph data have imposed unprecedented chal- lenges to machine learning due to their inherent complexity. Unlike text, audio and images, graph data are embedded in an irregular domain, making some essential operations of existing machine learning algorithms inapplicable. Many graph learning models and algorithms have been developed to tackle these challenges. This paper presents a systematic review of the state-of-the-art graph learning approaches as well as their potential applications. The paper serves mul- tiple purposes. First, it acts as a quick reference to graph learning for researchers and practitioners in different areas such as social computing, information retrieval, computer vision, bioinformatics, economics, and e-commence. Second, it presents insights into open areas of research in the field. Third, it aims to stimulate new research ideas and more interests in graph learning. I. I NTRODUCTION F. Xia is with School of Engineering, IT and Physical Sciences, Federation University Australia, Ballarat, VIC 3353, Australia K. Sun, S. Yu, A. Aziz, and L. Wan are with School of Software, Dalian University of Technology, Dalian 116620, China. S. Pan is with Faculty of Information Technology, Monash University, Melbourne, VIC 3800, Australia. H. Liu is with School of Computing, Informatics, and Decision Systems Engineering, Arizona State University, Tempe, AZ 85281, USA. Corresponding author: Feng Xia; e-mail: f.xia@ieee.orgGRAPHS, also referred to as networks, can be extracted from various real-world relations among abundant enti- ties. Some common graphs have been widely used to formulate different relationships, such as social networks, biological networks, patent networks, traffic networks, citation networks, and communication networks [1]–[3]. A graph is often defined by two sets, i.e., vertex set and edge set. Vertices represent en- tities in graph, whereas edges represent relationships between those entities. Graph learning has attracted considerable atten- tion because of its wide applications in the real world, such as data mining and knowledge discovery. Graph learning meth- ods have gained increasing popularity for capturing complex relationships, as graphs exploit essential and relevant relations among vertices [4], [5]. For example, in microblog networks, the spread trajectory of rumors can be tracked by detecting information cascades. In biological networks, new treatments for difficult diseases can be discovered by inferring protein interactions. In traffic networks, human mobility patterns can be predicted by analyzing the co-occurrence phenomenon with different timestamps [6]. Efficient analysis of these networks massively depends on the way how networks are represented. A. What is Graph Learning? Generally speaking, graph learning refers to machine learn- ing on graphs. Graph learning methods map the features of a graph to feature vectors with the same dimensions in the embedding space. A graph learning model or algorithm directly converts the graph data into the output of the graph learning architecture without projecting the graph into a low dimensional space. Most graph learning methods are based on or generalized from deep learning techniques, because deep learning techniques can encode and represent graph data into vectors. The output vectors of graph learning are in continuous space. The target of graph learning is to extract the desired features of a graph. Thus the representation of a graph can be easily used by downstream tasks such as node classification and link prediction without an explicit embedding process. Consequently, graph learning is a more powerful and meaningful technique for graph analysis. In this survey paper, we try to examine machine learning methods on graphs in a comprehensive manner. As shown in Fig. 1, we focus on existing methods that fall into the following four categories: graph signal processing (GSP) based methods, matrix factorization based methods, random walk based methods, and deep learning based methods. Roughly speaking, GSP deals with sampling and recovery of graph, and learning topology structure from data. Matrix factorization can be divided into graph Laplacian matrix factorization and vertex proximity matrix factorization. Random walk based methodsarXiv:2105.00696v1 [cs.LG] 3 May 2021IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 2 Fig. 1: The categorization of graph learning. include structure-based random walk, structure and node in- formation based random walk, random walk in heterogeneous networks, and random walk in time-varying networks. Deep learning based methods include graph convolutional networks, graph attention networks, graph auto-encoder, graph generative networks, and graph spatial-temporal networks. Basically, the model architectures of these methods/techniques differ from each other. This paper presents an extensive review of the state-of-the-art graph learning techniques. Traditionally, researchers adopt an adjacency matrix to represent a graph, which can only capture the relationship between two adjacent vertices. However, many complex and irregular structures cannot be captured by this simple rep- resentation. When we analyze large-scale networks, tradi- tional methods are computationally expensive and hard to be implemented in real-world applications. Therefore, effective representation of these networks is a paramount problem to solve [4]. Network Representation Learning (NRL) proposed in recent years can learn latent features of network vertices with low dimensional representation [7]–[9]. When the new representation has been learned, previous machine learning methods can be employed for analyzing the graph data as well as discovering relationships hidden in the data. When complex networks are embedded into a latent, low dimensional space, the structural information and vertex at- tributes can be preserved [4]. Thus the vertices of networks can be represented by low dimensional vectors. These vectors can be regarded as the features of input in previous machine learn- ing methods. Graph learning methods pave the way for graph analysis in the new representation space, and many graph analytical tasks, such as link prediction, recommendation and classification, can be solved efficiently [10], [11]. Graphical network representation sheds light on various aspects of social life, such as communication patterns, community structure, and information diffusion [12], [13]. According to the at- tributes of vertices, edges and subgraph, graph learning tasks can be divided into three categories, which are vertices based, edges based, and subgraph based, respectively. The relation- ships among vertices in a graph can be exploited for, e.g.,classification, risk identification, clustering, and community detection [14]. By judging the presence of edges between two vertices in graphs, we can perform recommendation and knowledge reasoning, for instance. Based on the classification of subgraphs [15], the graph can be used for, e.g., polymer classification, 3D visual classification, etc. For GSP, it is sig- nificant to design suitable graph sampling methods to preserve the features of the original graph, which aims at recovering the original graph efficiently [16]. Graph recovery methods can be used for constructing the original graph in the presence of incomplete data [17]. Afterwards, graph learning can be exploited to learn the topology structure from graph data. In summary, graph learning can be used to tackle the following challenges, which are difficult to solve by using traditional graph analysis methods [18]. 1)Irregular domains: Data collected by traditional sen- sors have a clear grid structure. However, graphs lie in an irregular domain (i.e., non-Euclidean space). In contrast to regular domain (i.e., Euclidean space), data in non- Euclidean space are not ordered regularly. Distance is hence difficult to be defined. As a result, basic methods based on traditional machine learning and signal pro- cessing cannot be directly generalized to graphs. 2)Heterogeneous networks: In many cases, networks involved in the traditional graph analysis algorithms are homogeneous. The appropriate modeling methods only consider the direct connection of the network and strip other irrelevant information, which significantly simplifies the processing. However, it is prone to cause information loss. In the real world, the edges among vertices and the types of vertices are usually diverse, such as in the academic network shown in Fig. 2. Thus it isn’t easy to discover potential value from heterogeneous information networks with abundant vertices and edges. 3)Distributed algorithms: In big social networks, there are often millions of vertices and edges [19]. Centralized algorithms cannot handle this since the computational complexity of these algorithms would significantly in- crease with the growth of vertex number. The design ofIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 3 distributed algorithms for dealing with big networks is a critical problem yet to be solved [20]. One major benefit of distributed algorithms is that the algorithms can be executed in multiple CPUs or GPUs simultaneously, and hence the running time can be reduced significantly. B. Related Surveys There are several surveys that are partially related to the scope of this paper. Unlike these surveys, we aim to provide a comprehensive overview of graph learning methods, with a focus on four specific categories. In particular, graph signal processing is introduced as one approach for graph learning, which is not covered by other surveys. Goyal and Ferrara [21] summarized graph embedding meth- ods, such as matrix factorization, random walk and their applications in graph analysis. Cai et al. [22] reviewed graph embedding methods based on problem settings and embedding techniques. Zhang et al. [4] summarized NRL methods based on two categories, i.e., unsupervised NRL and semi-supervised NRL, and discussed their applications. Nickel et al. [23] introduced knowledge extraction methods from two aspects: latent feature models and graph based models. Akoglu et al. [24] reviewed state-of-the-art techniques for event detec- tion in data represented as graphs, and their applications in the real world. Zhang et al. [18] summarized deep learning based methods for graphs, such as graph neural networks (GNNs), graph convolutional networks (GCNs) and graph auto-encoders (GAEs). Wu et al. [25] reviewed state-of-the- art GNN methods and discussed their applications in dif- ferent fields. Ortega et al. [26] introduced GSP techniques for representation, sampling and learning, and discussed their applications. Huang et al. [27] examined the applications of GSP in functional brain imaging and addressed the problem of how to perform brain network analysis from signal processing perspective. In summary, none of the existing surveys provides a com- prehensive overview of graph learning. They only cover some parts of graph learning, such as network embedding and deep learning based network representation. The NRL and/or GNN based surveys do not cover the GSP techniques. In contrast, we review GSP techniques in the context of graph learning, as it is an important approach for GNNs. Specifically, this survey paper integrates state-of-the-art machine learning techniques for graph data, gives a general description of graph learning, and discusses its applications in various domains. C. Contributions and Organization The contributions of this paper can be summarized as follows. A comprehensive overview of state-of-the-art graph learning methods: we present an integral introduction to graph learning methods, including, e.g., technical sketches, application scenarios, and potential research directions. Taxonomy of graph learning: we give a technical clas- sification of mainstream graph learning methods from the perspective of theoretical models. Technical descriptions Fig. 2: Heterogeneous academic network [28]. are provided wherever appropriate to improve understand- ing of the taxonomy. Insights into future directions in graph learning: Besides qualitative analysis of existing methods, we shed light on potential research directions in the field of graph learning through summarizing several open issues and relevant challenges. The rest of this paper is organized as follows. An overview of graph learning approaches containing graph signal pro- cessing based methods, matrix factorization based methods, random walk based methods, and deep learning based methods is provided in Section II. The applications of graph learning are examined in Section III. Some future directions as well as challenges are discussed in Section IV. We conclude the survey in Section V. II. G RAPH LEARNING MODELS AND ALGORITHMS The feature vectors that represent various categorical at- tributes are viewed as the input in previous machine learning methods. However, the mapping from the input feature vectors to the output prediction results need to be handled by graph learning [21]. Deep learning has been regarded as one of the most successful techniques in artificial intelligence [29], [30]. Extracting complex patterns by exploiting deep learning from a massive amount of irregular data has been found very useful in various fields, such as pattern recognition and image pro- cessing. Consequently, how to utilize deep learning techniques to extract patterns from complex graphs has attracted lots of attention. Deep learning on graphs, such as GNNs, GCNs, and GAEs, has been recognized as a powerful technique for graph analysis [18]. Besides, GSP has also been proposed to deal with graph analysis [26]. One of the most typical scenarios is that a set of values reside on a set of vertices, and these vertices are connected by edges [31]. Graph signals can be adopted to model various phenomena in real world. For example, in social networks, users in Facebook can be viewed as vertices, and their friendships can be modeled as edges. The number of followers of each vertex is marked in this social network. Based on this assumption, many techniques (e.g., convolution, filter, wavelet, etc.) in classical signal processing can be employed for GSP with suitable modifications [26]. In this section, we review graph learning models and algo- rithms under four categories as mentioned before, namely GSPIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 4 based methods, matrix factorization based methods, random walk based methods, and deep learning based methods. In Table I, we list the abbreviations used in this paper. TABLE I: Definitions of abbreviations Abbreviation Definition PCA Principal component analysis NRL Network representation learning LSTM Long short-term memory (networks) GSP Graph signal processing GNN Graph neural network GMRF Gauss markov random field GCN Graph convolutional network GAT Graph attention network GAN Generative adversarial network GAE Graph auto-encoder ASP Algebraic signal processing RNN Recurrent neural network CNN Convolutional neural network A. Graph Signal Processing Signal processing is a traditional subject that processes signals defined in regular data domain. In recent years, re- searchers extend concepts of traditional signal processing into graphs. Classical signal processing techniques and tools such as Fourier transform and filtering can be used to analyze graphs. In general, graphs are a kind of irregular data, which are hard to handle directly. As a complement to learning methods based on structures and models, GSP provides a new perspective of spectral analysis of graphs. Derived from signal processing, GSP can give an explanation of graph property consisting of connectivity, similarity, etc. Fig. 3 gives a simple example of graph signals at a certain time point, which is defined as observed values. In a graph, the above mentioned observed values can be regarded as graph signals. Each node is then mapped to the real number field in GSP. The main task of GSP is to expand signal processing approaches to mine implicit information in graphs. Fig. 3: The measurements of PM2.5 from different sensors on July 5, 2014 (data source: https://www.epa.gov/).1) Representation on Graphs: A meaningful representation of graphs has contributed a lot to the rapid growth of graph learning. There are two main models of GSP, i.e., adjacency matrix based GSP [31] and Laplacian based GSP [32]. An adjacency matrix based GSP comes from algebraic signal processing (ASP) [33], which interprets linear signal process- ing from algebraic theory. Linear signal processing contains signals, filters, signal transformation, etc. It can be applied in both continuous and discrete time domains. The basic assumption of linear algebra is extended to the algebra space in ASP. By selecting signal model appropriately, ASP can obtain different instances in linear signal processing. In adjacency matrix based GSP, the signal model is generated from a shift. Similar to traditional signal processing, a shift in GSP is a filter in graph domain [31], [34], [35]. GSP usually defines graph signal models using adjacency matrices as shifts. Signals of a graph are normally defined at vertices. Laplacian based GSP originates from spectral graph theory. High dimensional data are transferred into a low dimensional space generated by a part of the Laplacian basis [36]. Some researchers exploited sensor networks [37] to achieve dis- tributed processing of graph signals. Other researchers solved the problem globally under the assumption that the graph is smooth. Unlike adjacency matrix based GSP, Laplacian matrix is symmetric with real and non-negative edge weights, which is used to index undirected graphs. Although the models use different matrices as basic shifts, most of the notions in GSP are derived from signal processing. Notions with different definitions in these models may have similar meanings. All of them correspond to concepts in signal processing. Signals in GSP are values defined on graphs, and they are usually written as a vector, s= [s0;s1;:::;sN1]2 CN: N is the number of vertices, and each element in the vector represents the value on a vertex. Some studies [26] allow complex-value signals, even though most applications are based on real-value signals. In the context of adjacency matrix based GSP, a graph can be represented as a triple G(V;E;W), whereVis the vertex set,Eis the edge set and Wis the adjacency matrix. With the definition of graphs, we can also define degree matrix Dii=di, whereDis a diagonal matrix, and diis the degree of vertexi. Graph Laplacian is defined as L=DW, and normalized Laplacian is defined as Lnorm =D1=2LD1=2. Filters in signal processing can be seen as a function that amplifies or reduces relevant frequencies, eliminating irrele- vant ones. Matrix multiplication in linear space equals to scale changing, which is identical with filter operation in frequency domain. It is obvious that we can use matrix multiplication as a filter in GSP, which is written as sout=Hsin, whereH stands for a filter. Shift is an important concept to describe variation in sig- nal, and time-invariant signals are used frequently [31]. In fact, there are different choices of shifts in GSP. Adjacency matrix based GSP uses Aas shift. Laplacian based GSP uses L[32], and some researchers also use other matrices [38]. By following time invariance in traditional signal processing, shift invariance is defined in GSP. If filters are commutative with shift, they are shift-invariant, which can be written asIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 5 AH =HA . It is proved that shift-invariant filter can be represented by the shift. The properties of shift are vital, and they determine the fashion of other definitions such as Fourier transform and frequency. In adjacency matrix based GSP, eigenvalue decomposition of shiftAisA=VV1.Vis the matrix of eigenvectors [v0;v1;:::;vN1]and =2 640 ... N13 75 is a diagonal matrix of eigenvalues. The Fourier transform matrix is the inverse of V, i.e.,F=V1. Frequency of shift is defined as total variation, which states the difference after shift TVG=jjvk1 maxAvkjj1; where1 maxis a normalized factor of matrix. It means that the frequencies of eigenvalue far away from the largest eigenval- ues on complex plane are large. A large frequency means that signals are changed with a large scale after shift filtering. The differences between minimum and maximum can be seen in Fig. 4. Generally, the total variation tends to be relatively low with larger frequency, and vice versa. Eigenvectors of larger eigenvalues can be used to construct low-frequency filters, which capture fundamental characteristics, and smaller ones can be employed to capture the variation among neighbor nodes. For topology learning problems, we can distinguish the corresponding solutions depending on known information. When topology information is partly known, we can use the known information to infer the whole graph. In case the topology information is unknown while we still can observe the signals on the graph, the topology structure has to be inferred from the signals. The former one is often solved as a sampling and recovery problem, and blind topology inference is also known as graph topology (or structure) learning. 2) Sampling and Recovery: Sampling is not a new concept defined in GSP. In conventional signal processing, we normally need to reconstruct original signals with the least samples and retain all information of original signals for a sampling problem. Few samples lead to the lack of information and more samples need more space to store. The well-known Nyquist- Shannon sampling theorem gives the sufficient condition of perfect recovery of signals in time domain. Researchers have migrated the sampling theories into GSP to study the sampling problem on graphs. As the volume of data is large in some real-world applications such as sensor networks and social networks, sampling less and recover- ing better are vital for GSP. In fact, most algorithms and frameworks solving sampling problems require that the graph models correlations within signals observed on it [39]. The sampling problem can be defined as reconstructing signals from samples on a subset of vertices, and signals in it are usually band-limited. Nyquist-Shannon sampling theorem was extended to graph signals in [40]. Based on the normalized Laplacian matrix, sampling theorem and cut-off frequency are (a) The maximum frequency (b) The minimum frequency Fig. 4: Illustration of difference between minimum and max- imum frequencies. defined for GSP. Moreover, the authors provided a method for computing cut-off frequency from a given sampling set and a method for choosing sampling set for a given bandwidth. It should be noted that the sampling theorem proposed therein is merely applied to undirected graph. As Laplacian matrix represents undirected graphs only, sampling theory for directed graph adopts adjacent matrix. An optimal operator with a guarantee for perfect recovery was proposed in [35], and it is robust to noise for general graphs. One of the explicit distinctions between classical signal processing and GSP is that signals of the former fall in regular domain while the latter falls in irregular domain. For sampling and recovery problems, classical signal processing samples successive signals and can recover successive signals from samplings. GSP samples a discrete sequence, and recovers the original sequences from samplings. By following this order, the solution is generally separated into two parts, i.e., findingIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 6 sampling vertex sets and reconstructing original signals based on various models. When the size of the dataset is small, we can handle the signal and shift directly. However, for a large-scale dataset, some algorithms require matrix decomposition to obtain fre- quencies and save eigenvalues in the procedure, which are almost impossible to realize. As a simple technique applicable to large-scale datasets, a random method can also be used in sampling. Puy et al. [41] proposed two sample strategies: a non-adaptive one depending on a parameter and an adaptive random sampling strategy. By relaxing the optimized con- straint, they extended random sampling to large scale graphs. Another common strategy is greedy sampling. For example, Shomorony and Avestimehr [42] proposed an efficient method based on linear algebraic conditions that can exactly compute cut-off frequency. Chamon and Ribeiro [43] provided near- optimal guarantee for greedy sampling, which guarantees the performance of greedy sampling in the worst cases. All of the sampling strategies mentioned above can be categorized as selecting sampling, where signals are observed on a subset of vertices [43]. Besides selecting sampling, there exists a type of sampling called aggregation sampling [44], which uses observations taken at a single vertex as input, containing a sequential applications of graph shift operator. Similar to classical signal processing, the reconstruction task on graphs can also be interpreted as data interpolation problem [45]. By projecting the samples on a proper signal space, researchers obtain interpolated signals. Least squares reconstruction is an available method in practice. Gadde and Ortega [46] defined a generative model for signal recovery derived from a pairwise Gaussian random field (GRF) and a covariance matrix on graphs. Under sampling theorem, the reconstruction of graph signals can be viewed as the maximum posterior inference of GRF with low-rank approximation. Wang et al. [47] aimed at achieving the distributed reconstruc- tion of time-varying band limited signal, where the distributed least squares reconstruction (DLSR) was proposed to recover the signals iteratively. DLSR can track time-varying signals and achieve perfect reconstruction. Di Lorenzo et al. [48] proposed a linear mean squares (LMS) strategy for adaptive estimation. LMS enables online reconstruction and tracking from the observation on a subset of vertices. It also allows the subset to vary over time. Moreover, a sparse online estimation was proposed to solve the problems with unknown bandwidth. Another common technique for recovering original signals is smoothness. Smoothness is used for inferring missing values in graph signals with low frequencies. Wang et al. [17] defined the concept of local set. Based on the definition of graph signals, two iterative methods were proposed to recover band limited signals on graphs. Besides, Romero et al. [49] advocated kernel regression as a framework for GSP modeling and reconstruction. For parameter selection in estimators, two multi-kernel methods were proposed to solve a single optimization problem as well. In addition, some researchers investigated different recovery problems with compressed sensing [50]. In addition, there exists some researches on sampling of different kinds of signals such as smooth graph signals, piece-wise constant signals and piece-wise smooth signals [51]. Chen et al. [51] gave a uniform framework to analyze graph signals. The reconstruction of a known graph signal was stud- ied in [52], where the signal is sparse, which means only a few vertices are non-zeros. Three kinds of reconstruction schemes corresponding to various seeding patterns were examined. By analyzing single simultaneous injection, single successive value injection, and multiple successive simultaneous injec- tions, the conditions for perfect reconstruction on any vertices were derived. 3) Learning Topology Structure from Data: In most appli- cation scenes, graphs are constructed according to connections of entity correlations. For example, in sensor networks, the cor- relations between sensors are often consistent with geographic distance. Edges in social networks are defined as relations such as friends or colleagues [53]. In biochemical networks, edges are generated by interactions. Although GSP is an efficient framework for solving problems on graphs such as sampling, reconstruction, and detection, there lacks a step to extract relations from datasets. Connections exist in many datasets without explicit records. Fortunately, they can be inferred in many ways. As a result, researchers want to learn complete graphs from datasets. The problem of learning graph from a dataset is stated as estimating graph Laplacian, or graph topology [54]. Generally, they require the graph to satisfy some properties, such as sparsity and smoothness. Smoothness is a widespread assumption in networks generated from datasets. Therefore, it is usually used to constrain observed signals and provide a rational guarantee for graph signals. Researchers have applied it to graph topology learning. The intuition behind smoothness based algorithms is that most signals on graph are stationary, and the result filtered by shift tends to be the lowest frequency. Dong et al. [55] adopted a factor analysis model for graph signals, and also imposed a Gaussian prior on latent variables to obtain a Principal Component Analysis (PCA) like represen- tation. Kalofolias [56] formulated the objective as a weighted l1problem and designed a general framework to solve it. Gauss Markov Random Field (GMRF) is also a widely used theory for graph topology learning in GSP [54], [57], [58]. The models of GRMF based graph topology learning select graphs that are more likely to generate signals which are similar to the ones generated by GMRF. Egilmez et al. [54] formulated the problem as a maximum posterior parameter estimation of GMRF, and the graph Laplacian is a precision matrix. Pavez and Ortega [57] also formulated the problem as a precision matrix estimation, and the rows and columns are updated iteratively by optimizing a quadratic problem. Both of them restrict the result matrix, which should be Laplacian. In [58], Pavez et al. chose a two steps framework to find the structure of the underlying graph. First, a graph topology inference step is employed to select a proper topology. Then, a generalized graph Laplacian is estimated. An error bound of Laplacian estimation is computed. In the next step, the error bound can be utilized to obtain a matrix in a specific form as the precision matrix estimation. It is one of the first work that suggests adjusting the model to obtain a graph satisfying the requirement of various problems.IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 7 Diffusion is also a relevant model that can be exploited to solve the topology interfering problem [39], [59]–[61]. Diffusion refers to that the node continuously influences its neighborhoods. In graphs, nodes with larger values will have higher influence on their neighborhood nodes. Using a few components to represent signals will help to find the main factors of signal formation. The models of diffusion are often under the assumption of independent identically-distributed signals. Pasdeloup et al. [59] gave the concept of valid graphs to explain signals and assumed that the signals are observed after diffusion. Segarra et al. [60] agreed that there exists a diffusion process in the shift, and the signals can be observed. The signals in [61] were explained as a linear combination of a few components. For time series recorded in data, researchers tried to construct time-sequential networks. For instance, Mei and Moura [62] proposed a methodology to estimate graphs, which considers both time and space dependencies and models them by auto-regressive process. Segarra et al. [63] proposed a method that can be seen as an extension of graph learning. The aim of the paper was to solve the problem of joint identification of a graph filter and its input signal. For recovery methods, a well-known partial inference prob- lem is recommendation [45], [64], [65]. The typical algorithm used in recommendation is collaborative filtering (CF) [66]. Given the observed ratings in a matrix, the objective of CF is to estimate the full rating matrix. Huang et al. [65] demonstrated that collaborative filtering could be viewed as a specific band- stop graph filter on networks representing correlations between users and items. Furthermore, linear latent factor methods can also be modeled as band limited interpretation problem. 4) Discussion: GSP algorithms have strict limitations on experimental data, thus leading to less real-world applications. Moreover, GSP algorithms require the input data to be exactly the whole graph, which means that part of graph data cannot be the input. Therefore, the computational complexity of this kind of methods could be significantly high. In comparison with other kinds of graph learning methods, the scalability of GSP algorithms is relatively poor. B. Matrix Factorization Based Methods Matrix factorization is a method of simplifying a matrix into its components. These components have a lower dimension and could be used to represent the original information of a network, such as relationships among nodes. Matrix fac- torization based graph learning methods adopt a matrix to represent graph characteristics like vertex pairwise similarity, and the vertex embedding can be achieved by factorizing this matrix [67]. Early graph learning approaches usually utilized matrix factorization based methods to solve the graph embed- ding problem. The input of matrix factorization is the non- relational high dimensional data feature represented as a graph. The output of matrix factorization is a set of vertex embedding. If the input data lies in a low dimensional manifold, the graph learning for embedding can be treated as a dimension-reduced problem that preserves the structure information. There are mainly two types of matrix factorization based graph learning.One is graph Laplacian matrix factorization, and the other is vertex proximity matrix factorization. 1) Graph Laplacian Matrix Factorization: The preserved graph characteristics can be expressed as pairwise vertex similarities. Generally, there are two kinds of graph Laplacian matrix factorization, i.e., transductive and inductive matrix factorization. The former only embeds the vertices contained in the training set, and the latter can embed the vertices that are not contained in the training set. The general framework has been designed in [68], and the graph Laplacian matrix factor- ization based graph learning methods have been summarized in [69]. The Euclidean distance between two feature vectors is directly adopted in the initial Metric Multidimensional Scaling (MDS) [70] to find the optimal embedding. The neighborhoods of vertices are not considered in the MDS, i.e., any pair of training instances are considered as connected. The data feature is extracted by constructing a knearest neighbor graph, and the subsequent studies [67], [71]–[73] tackle this issue. The topksimilar neighbors of each vertex are connected with itself. A similar matrix is calculated by exploiting different methods, and thus the graph characteristics can be preserved as much as possible. Recently, researchers have designed more sophisticated models. The performance of earlier matrix factorization model Locality Preserving Projection (LPP) can be improved by introducing an anchor taking advantage of Anchorgraph-based Locality Preserving Projection (AgLPP) [74], [75]. The graph structure can be captured by using a local regression model and a global regression process based on Local and Global Regressive Mapping (LGRM) [76]. The global geometry can be preserved by using local spline regression [77]. More information can be preserved by exploiting the auxil- iary information. An adjacency graph and a labelled graph were constructed in [78]. The objective function of LPP preserves the local geometric structure of the datasets [67]. An adjacency graph and a relational feedback graph were con- structed in [79] as well. The graph Laplacian regularization, k-means and PCA were considered in RF-Semi-NMF-PCA si- multaneously [80]. Other works, e.g., [81], adopt semi-definite programming to learn the adjacency graph that maximizes the pairwise distances. 2) Vertex Proximity Matrix Factorization: Apart from solv- ing the above generalized eigenvalue problem, another ap- proach of matrix factorization is to factorize vertex proximity matrix directly. In general, matrix factorization can be used to learn the graph structure from non-relational data, and it is applicable to learn homogeneous graphs. Based on matrix factorization, vertex proximity can be approximated in a low dimensional space. The objective of preserving vertex proximity is to minimize the error. The Singular Value Decomposition (SVD) of vertex proximity matrix was adopted in [82]. There are some other approaches such as regularized Gaussian matrix factorization [83], low- rank matrix factorization [84], for solving SVD. 3) Discussion: Matrix factorization algorithms operate on an interaction matrix to decompose several lower dimension matrices. The process brings some drawbacks. For example, the algorithms require a large memory when the decomposedIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 8 matrices become large. In addition, matrix factorization al- gorithms are not applicable to supervised or semi-supervised tasks with the training process. C. Random Walk Based Methods Random walk is a convenient and effective way to sample networks [85], [86]. This method can generate sequences of nodes meanwhile preserving original relations between nodes. Based on network structure, NRL can generate feature vectors of vertices so that downstream tasks can mine network information in a low dimensional space. An example of NRL is shown in Fig. 5. The image in Euclidean space is shown in Fig. 5(a), and the corresponding graph in non-Euclidean space is shown in Fig. 5(b). As one of the most successful NRL algorithms, random walks play an important role in dimensionality reduction. (a) Image in Euclidean space (b) Graph in non-Euclidean space Fig. 5: An example of NRL mapping an image from Euclidean space into non-Euclidean space. 1) Structure Based Random Walks: Graph-structured data have various data types and structures. The information en- coded in a graph is related to graph structure and vertex attributes, which are the two key factors affecting the reason- ing of networks. In real-world applications, many networks only have structural information, but lack vertex attribute information. How to identify network structure information effectively, such as important vertices and invisible links,attracts the interest of network scientists [87]. Graph data have high dimensional characteristics. Traditional network analysis methods cannot be used for analyzing graph data in a continuous space. In recent years, various NRL methods have been proposed, which preserve rich structural information of networks. Deep- Walk [88] and Node2vec [7] are two representative methods for generating network representation of basic network topol- ogy information. These methods use random walk models to generate random sequences on networks. By treating the vertices as words and the generated random sequences of vertices as word sequences (sentences), the models can learn the embedding representation of the vertices by inputting these sequences into the Word2vec model [89]–[91]. The principle of the learning model is to maximize the co-occurrence prob- ability of vertices such as Word2vec. In addition, Node2vec shows that network has complex structural characteristics, and different network structure samplings can obtain different results. The sampling mode of DeepWalk is not enough to capture the diversity of connection patterns in networks. Node2vec designs a random walk sampling strategy, which can sample the networks with the preference of breadth-first sampling and depth-first sampling by adjusting the parameters. The NRL algorithms mentioned above focused on the first- order proximity information of vertices. Tang et al. [92] proposed a method called LINE for large-scale network embedding. LINE can maintain the first and second order approximations. The first-order neighbor refers to the one- hop neighbor between two vertices, and the second-order neighbor is the neighbor with two hops. LINE is not a deep learning based model, but it is often compared with these edge modeling based methods. It has been proved that the network structure information plays an important role in various network analysis tasks. In addition to this structural information, network attributes in the original network space are also critical in modeling the formation and evolution of the network [93]. 2) Structure and Vertex Information Based Random Walks: In addition to network topology, many types of networks also have rich vertex information, such as vertex content or label in networks. Yang et al. [84] proposed an algorithm called TADW. The model is based on DeepWalk and considers the text information of vertices. The MMDW [94] is another model based on DeepWalk, which is a kind of semi-supervised network embedding algorithm, by leveraging labelling infor- mation of vertices to enhance the performance. Focusing on the structural identity of nodes, Ribeiro et al. [95] formulated a framework named Struc2vec. The frame- work considers nodes with similar local structure rather than neighborhood and labels of nodes. With hierarchy to evaluate structural similarity, the framework constrains structural sim- ilarity more stringently. The experiments indicate that Deep- Walk and Node2vec are worse than Struc2vec which considers structural identity. There are some other NRL models, such as Planetoid [96], which learn network representation using the feature of network structure and vertex attribute informa- tion. It is well known that vertex attributes provide effective information for improving network representation and helpIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 9 to learn embedded vector space. In the case of relatively sparse network topology, vertex attribute information can be used as supplementary information to improve the accuracy of representation. In practice, how to use vertex information effectively and how to apply this information to network vertex embedding are the main challenges in NRL. Researchers not only investigate random walk based NRL on vertices but also on graphs. Adhikari et al. [97] proposed an unsupervised scalable algorithm, Sub2Vec, to learn arbitrary subgraph. To be more specific, they proposed a method to measure the similarities between subgraphs without disturbing local proximity. Narayanan et al. [98] proposed graph2vec, which is a neural embedding framework. Modeling on neural document embedding models, graph2vec takes a graph as a document and the subgraph around words as vertices. By migrating the model to graphs, the performance of graph2vec significantly exceeds other substructure representation learning algorithms. Generally, random walk can be regarded as a Markov process. The next state of the process is only related to last state, which is known as Markov chain. Inspired by vertex- reinforced random walks, Benson et al. [99] presented spacey random walk, a non-Markovian stochastic process. As a spe- cific type of a more general class of vertex-reinforced random walks, it takes the view that the probability of time remained on each vertex relates to the long term behavior of dynamical systems. They proved that dynamical systems can converge to a stationary distribution under sufficient conditions. Recently, with the development of Generative Adversarial Network (GAN), researchers combined random walks with the GAN method [100], [101]. Existing research on NRL can be divided into generative models and discriminative models. GraphGAN [100] integrated these two kinds of models and played a game-theoretical minimax game. With the process of the game, the performance of the two models can be strengthened. Random walk is used as a generator in the game. NetGAN [101] is a generative model that can model network in real applications. The method takes the distribution of biased random walk as input, and can produce graphs with known patterns. It preserves important topology properties and does not need to define them in model definition. 3) Random Walks in Heterogeneous Networks: In reality, most networks contain more than one type of vertex, and hence networks are heterogeneous. Different from homoge- neous NRL, heterogenous NRL should well reserve various relationships among different vertices [102]. Considering the ubiquitous existence of heterogeneous networks, many ef- forts have been made to learn network representations of heterogeneous networks. Compared to homogeneous NRL, the proximity among entities in heterogeneous NRL is more than a simple measure of distance or closeness. The semantics among vertices and links should be considered. Some typical scenarios include knowledge graphs and social networks. Knowledge graph is a popular research domain in recent years. A vital part in knowledge base population is relational inference. The central problem of relational inference is infer- ring unknown knowledge from the existing facts in knowledge bases [103]. There are three types of common relationalinference method in general: statistical relational learning (SRL), latent factor models (LFM) and random walk models (RWM). Relational learning methods based on statistics lack generality and scalability. As a result, latent factor model based graph embedding and relational paths based random walk have been adopted more widely. In a knowledge graph, there exist various vertices and various types of relationships among different vertices. For example, in a scholar related knowledge graph [2], [28], the types of vertices include scholar, paper, publication venue, institution, etc. The types of relationships include coauthor, citation, publication, etc. The key idea of knowledge graph embedding is to embed vertices and their relationships into a low dimensional vector space, while the inherent structure of the knowledge graph can be reserved [104]. For relational paths based random walk, the path ranking algorithm (PRA) is a path finding method using random walks to generate relational features on graph data [105]. Random walks in PRA are with restart, and combine features with a logistic regression. However, PRA cannot predict connection between two vertices if there does not exist a path between them. Gardner et al. [106], [107] introduced two ways to improve the performance of PRA. One method enables more efficient processing to incorporate new corpus into knowledge base, while the other method uses vector space to reduce the sparsity of surface forms. To resolve cascade errors in knowledge construction, Wang and Cohen [108] proposed a joint information extraction and knowledge base based model with a recursive random walk. Using latent context of the text, the model obtains additional improvement. Liu et al. [109] developed a new random walk based learning algorithm named Hierarchical Random-walk inference (HiRi). It is a two-tier scheme: the upper tier recognizes relational sequence pattern, and the lower tier captures information from subgraphs of knowledge bases. Another widely-investigated type of heterogeneous net- works is social networks, such as online social networks and location based social networks. Social networks are heteroge- neous in nature because of the different types of vertices and relations. There are two main ways to embed heterogeneous social networks, including meta path-based approaches and random walk-based approaches. A meta path in heterogeneous networks is defined as a sequence of vertex types encoding significant composite re- lations among various types of vertices. Aiming to employ the rich information in social networks by exploiting various types of relationships among vertices, Fu et al. [110] proposed HIN2Vec, which is a representation learning framework based on meta-paths. HIN2Vec is a neural network model and the meta-paths are well embedded based on two independent phases, i.e., training data preparation and representation learn- ing. Experimental results on various social network datasets show that HIN2Vec model is able to automatically learn vertex vector in heterogeneous networks to support a variety of applications. Metapath2vec [111] was designed by formalizing meta-path based random walks to construct the neighborhoods of a vertex in heterogeneous networks. It takes the advantage of a heterogeneous skip-gram model to perform vertex em-IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 10 bedding. Meta path based methods require either prior knowledge for optimal meta-path selection or extended computations for path length selection. To overcome these challenges, random walk based approaches have been proposed. Hussein et al. [112] proposed the JUST model, which is a heterogeneous graph embedding approach using random walks with jump and stay strategies so that the aforementioned bias can be overcomed ef- fectively. Another method which does not require prior knowl- edge for meta-path definition is MPDRL [113], meta-path discovery with reinforcement earning. This method employs the reinforcement learning algorithm to perform multi-hop rea- soning to generate path instances and then further summarizes the important meta-paths using the Lowest Common Ancestor principle. Shi et al. [114] proposed the HERec model, which utilizes the heterogeneous information network embedding for providing accurate recommendations in social networks. HERec is designed based on a random walk based approach for generating meaningful vertex sequences for heterogeneous network embedding. HERec can effectively adopt the auxiliary information in heterogeneous information networks. Other typical heterogeneous social network embedding approaches include, e.g., PTE [115] and SHNE [116]. 4) Random Walks in Time-varying Networks: Network is evolving over time, which means that new vertices may emerge and new relations may appear. Therefore, it is significant to capture the temporal behaviour of networks in network analysis. Many efforts have been made to learn time-varying network embedding (e.g., dynamic networks or temporal net- works) [117]. In contrast to static network embedding, time- varying NRL should consider the network dynamics, which means that old relationships may become invalid and new links may appear. The key of time-varying NRL is to find a suitable way to incorporate the time characteristic into embedding via reason- able updating approaches. Nguyen et al. [118] proposed the CTDNE model for continuous dynamic network embedding based on random walk with ”chronological” paths which can only move forward as time goes on. Their model is more suitable for time-dependent network representation that can capture the important temporal characteristics of continuous- time dynamic networks. Results on various datasets show that CTDNE outperforms static NRL approaches. Zuo et al. [119] proposed the HTNE model which is a temporal NRL approach based on the Hawkes process. HTNE can well integrate the Hawkes process into network embedding so that the influence of historical neighbors on the current neighbors can be accurately captured. For unseen vertices in a dynamical network, Graph- SAGE [120] was presented to efficiently generate embed- dings for new vertices in network. In contrast to methods that training embedding for every vertex in the network, GraphSAGE designs a function to generate embedding for a vertex with features of the neighborhoods locally. After sampling neighbors of a vertex, GraphSAGE uses different aggregators to update the embedding of the vertex. However, current graph neural methods are proficient of only learning local neighborhood information and cannot directly explorethe higher-order proximity and the community structure of graphs. 5) Discussion: As mentioned before, random walk is a fundamental way to sample networks. The sequences of nodes could preserve the information of network structure. However, there are some disadvantages of this method. For example, random walk relies on random strategies, which creates some uncertain relations of nodes. To reduce this uncertainty, it needs to increase the number of samples, which will signifi- cantly increase the complexity of algorithms. Some random walk variants could preserve local and global information of networks, but they might not be effective in adjusting parameters to adapt to different types of networks. D. Deep Learning on Graphs Deep learning is one of the hottest areas over the past few years. Nevertheless, it is an attractive and challenging task to extend the existing neural network models, such as Recurrent Neural Networks (RNNs) or Convolutional Neural Networks (CNNs), to graph data. Gori et al. [121] proposed a GNN model based on recursive neural network. In this model, a transfer function is implemented, which maps the graph or its vertices to an m-dimensional Euclidean space. In recent years, lots of GNN models have been proposed. 1) Graph Convolutional Networks: GCN works on the ba- sis of grid structure domain and graph structure domain [122]. Time Domain and Spectral Methods . Convolution is one of a common operation in deep learning. However, since graph lacks a grid structure, standard convolution over images or text cannot be directly applied to graphs. Bruna et al. [122] extended the CNN algorithm from image processing to the graph using the graph Laplacian matrix, dubbed as spectral graph CNN. The main idea is similar to Fourier basis for signal processing. Based on [122], Henaff et al. [123] defined kernels to reduced the learning parameters by analogizing the local connection of CNNs on the image. Defferrard et al. [124] provided two ways for generalizing CNNs to graph structure data based on graph theory. One method is to reduce the parameters by using polynomial kernel, and this method can be accelerated by using Chebyshev polynomial approx- imation. The other method is the special pooling method, which is pooling on the binary tree constructed from vertices. An improved version of [124] was introduced by Kipf and Welling [125]. The proposed method is a semi-supervised learning method for graphs. The algorithm employs an excel- lent and straightforward neural network followed by a layer- by-layer propagation rule, which is based on the first-order approximation of spectral convolution on the graph and can be directly acted on the graph. There are some other time domain based methods. Based on the mixture model of CNNs, for instance, Monti et al. [126] generalized the CNN to non-Euclidean space. Zhou and Li [127] proposed a new CNN graph modeling framework, which designs two modules for graph structure data: K- order convolution operator and adaptive filtering module. In addition, the high-order adaptive graph convolution network (HA-GCN) framework proposed in [127] is a general ar- chitecture that is suitable for many applications of verticesIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 11 and graph centers. Manessi et al. [128] proposed a dynamic graph convolution network algorithm for dynamic graphs. The core idea of the algorithm is to combine the expansion of graph convolution with the improved Long Short Term- Memory networks (LSTM) algorithm, and then train and learn the downstream recursive unit by using graph structure data and vertex features. The spectral based NRL methods have many applications, such as vertex classification [125], traffic forecasting [129], [130], and action recognition [131]. Space Domain and Spatial Methods . Spectral graph theory provides a convolution method on graphs, but many NRL methods directly use convolution operation on graphs in space domain. Niepert et al. [132] applied graph labeling procedures such as Weisfeiler-Lehman kernel on graphs to generate unique order of vertices. The generated sub-graphs can be fed to the traditional CNN operation in space domain. Duvenaud et al. [133] designed Neural fingerprints (FP), which is a spatial method using the first-order neighbors similar to the GCN algorithm. Atwood and Towsley [134] proposed an- other convolution method, called diffusion-convolutional neu- ral network, which incorporates transfer probability matrix and replaces the characteristic basis of convolution with diffusion basis. Gilmer et al. [135] reformulated existing models into a single common framework, and exploited this framework to discover new variations. Allamanis et al. [136] represented the structure of code from syntactic and semantic, and utilized the GNN method to recognize program structures. Zhuang and Ma [137] designed dual graph convolution networks (DGCN), which use diffusion basis and adjacency basis. DGCN uses two convolutions: one is the characteristic form of polynomial filter, and the other is to replace the adjacency matrix with the PPMI (Positive Pointwise Mutual Information) of the transition probability [89]. Dai et al. [138] proposed the SSE algorithm, which uses asynchronous ran- dom to learn vertex representation so as to improve learning efficiency. In this model, a recursive method is adopted to learn vertex latent representation and the sampled batch data are utilized to update parameters. The recursive function of SSE is calculated from the weighted average of historical state and new state. Zhu et al. [139] proposed a graph smoothing splines neural network which exploits non-smoothing node features and global topological knowledge such as centrality for graph classification. Gao et al. [140] proposed a large scale graph convolution network (LGCN) based on vertex feature information. In order to adapt to the scene of large scale graphs, they proposed a sub-graph training strategy, which first trained the sampled sub-graph in a small batch. Based on a deep generative graph model, a novel method called DeepNC for inferring the missing parts of a network was proposed in [141]. A brief history of deep learning on graphs is shown in Fig. 6. GNN has attracted lots of attention since 2015, and it is widely studied and used in various fields. 2) Graph Attention Networks: In sequence-based tasks, attention mechanism has been regarded as a standard [142]. GNNs achieve lots of benefits from the expanded model capacity of attention mechanisms. GATs are a kind of spatial- based GCNs [143]. It takes the attention mechanism into con-sideration when determining the weights of vertex’s neighbors. Likewise, Gated Attention Networks (GAANs) also introduced the multi-head attention mechanism for updating the hidden state of some vertices [144]. Unlike GATs, GAANs employ a self-attention mechanism which can compute different weights for different heads. Some other models such as graph at- tention model (GAM) were proposed for solving different problems [145]. Take GAM as an example, the purpose of GAM is to handle graph classification. Therefore, GAM is set to process informative parts by visiting a sequence of signifi- cant vertices adaptively. The model of GAM contains LSTM network, and some parameters contain historical information, policies, and other information generated from exploration of the graph. Attention Walks (AWs) are another kind of learning model based on GNN and random walks [146]. In contrast to DeepWalk, AWs use differentiable attention weights when factorizing the co-occurrence matrix [88]. 3) Graph Auto-Encoders: GAE uses GNN structure to embed network vertices into low dimensional vectors. One of the most general solutions is to employ a multi-layer perception as the encoder for inputs [147]. Therein the decoder reconstructs neighborhood statistics of the vertex. PPMI or the first and the second nearest neighborhood can be taken into statistics [148], [149]. Deep neural networks for graph representations (DNGR) employ PPMI. Structural deep net- work embedding (SDNE) employs stacked auto-encoder to maintain both the first-order and the second-order proximity. Auto-encoder [150] is a traditional deep learning model, which can be classified as a self-supervised model [151]. Deep recursive network embedding (DRNE) reconstructs some vertices’ hidden state rather than the entire graph [152]. It has been found that if we regard GCN as an encoder, and combine GCN with GAN or LSTM with GAN, then we can design the auto-encoder for graphs. Generally speaking, DNGR and SDNE embed vertices by the given structure features, while other methods such as DRNE learn both topology structure and content features [148], [149]. Variational graph auto- encoder [153] is another successful approach that employs GCN as an encoder and a link prediction layer as a decoder. Its successor, adversarially regularized variational graph auto- encoder [154], adds a regularization process with an adversar- ial training approach to learn a more robust embedding. 4) Graph Generative Networks: The purpose of graph generative networks is to generate graphs according to the given observed set of graphs. Many previous methods of graph generative networks have their own application domains. For example, in natural language processing, the semantic graph or the knowledge graph is generated based on the given sentences. Some general methods have been proposed recently. One kind of them considers the generation process as the for- mation of vertices and edges. Another kind is to employ gener- ative adversarial training. Some GCNs based graph generative networks such as molecular generative adversarial networks (MolGAN) integrate GNN with reinforcement learning [155]. Deep generative models of graphs (DGMG) achieves a hidden representation of existing graphs by utilizing spatial-based GCNs [156]. There are some knowledge graph embedding algorithms based on GAN and Zero-Shot Learning [157]. VyasIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 12 Fig. 6: A brief history of algorithms of deep learning on graphs. et al. [158] proposed a Generalized Zero-Shot learning model, which can find unseen semantic in knowledge graphs. 5) Graph Spatial-Temporal Networks: Graph spatial- temporal networks simultaneously capture the spatial and tem- poral dependence of graphs. The global structure is included in the spatial-temporal graphs, and the input of each vertex varies with the change of time. For example, in traffic networks, each sensor records the traffic speed of a road continuously as a vertex, in which the edge of the traffic networks is determined by the distance between the sensor pairs [129]. The goal of a spatial-temporal network can be to predict future vertex values or labels, or to predict spatial-temporal graph labels. Recent studies in this direction have discussed the use of GCNs, the combination of GCNs with RNN or CNN, and recursive structures for graph structures [130], [131], [159]. 6) Discussion: In this context, the task of graph learning can be seen as optimizing the objective function by using gradient descent algorithms. Therefore the performance of deep learning based NRL models is influenced by gradient descent algorithms. They may encounter challenges like local optimal solutions and the vanishing gradient problem. III. A PPLICATIONS Many problems can be solved by graph learning methods, including supervised, semi-supervised, unsupervised, and re- inforcement learning. Some researchers classify the applica- tions of graph learning into three categories, i.e., structural scenarios, non-structural scenarios, and other application sce- narios [18]. Structural scenarios refer to the situation where data are performed in explicit relational structures, such as physical systems, molecular structures, and knowledge graphs. Non-structural scenarios refer to the situation where data are with unclear relational structures, such as images and texts. Other application scenarios include, e.g., integrating models and combinatorial optimization problems. Table II lists the neural components and applications of various graph learning methods. A. Datasets and Open-source Libraries There are several datasets and benchmarks used to evaluate the performance of graph learning approaches for various tasks such as link prediction, node classification, and graph visual- ization. For instance, datasets like Cora1(citation network), 1https://relational.fit.cvut.cz/dataset/CORAPubmed2(citation network), BlogCatalog3(social network), Wikipedia4(language network) and PPI5(biological network) include nodes, edges, labels or attributes of nodes. Some research institutions developed graph learning libraries, which include common and classical graph learning algorithms. For example, OpenKE6is a Python library for knowledge graph embedding based on PyTorch. The open-source framework has the implementations of RESCAL, HolE, DistMult, ComplEx, etc. CogDL7is a graph representation learning framework, which can be used for node classification, link prediction, graph classification, etc. B. Text Many data are in textual form coming from various re- sources like web pages, emails, documents (technical and corporate), books, digital libraries and customer complains, letters, patents, etc. Textual data are not well structured for obtaining any meaningful information as text often contains rich context information. There exist abundant applications around text, including text classification, sequence labeling, sentiment classification, etc. Text classification is one of the most classical problems in natural language processing. Popular algorithms proposed to handle this problem include GCNs [120], [125], GATs [143], Text GCNs [160], and Sentence LSTM [161]. Sentence LSTM has also been applied to sequence labeling, text generation, multi-hop reading com- prehension, etc [161]. Syntactic GCN was proposed to solve semantic role labeling and neural machine translation [162]. Gated Graph Neural Networks (GGNNs) can also be used to address neural machine translation and text generation [163]. For relational extraction, Tree LSTM, graph LSTM, and GCN are better solutions [164]. C. Images Graph learning applications pertaining to images include social relationship understanding, image classification, visual question answering, object detection, region classification, and semantic segmentation, etc. For social relationship understand- ing, for instance, graph reasoning model (GRM) is widely 2https://catalog.data.gov/dataset/pubmed 3http://networkrepository.com/soc-BlogCatalog.php 4https://en.wikipedia.org/wiki/Wikipedia:Database download 5https://openwetware.org/wiki/Protein-protein interaction databases 6https://github.com/thunlp/OpenKE 7https://github.com/THUDM/cogdl/IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 13 TABLE II: Summary of graph learning methods and their applications Categories Algorithms Neural Component Applications Time Domain and Spectral MethodsSNLCN [122] Graph Neural Network Classification DCN [123] Spectral Network Classification ChebNet [124] Convolution Network Classification GCN [125] Spectral Network Classification HA-GCN [127] GCN Classification Dynamic GCN [128] GCN, LSTM Classification DCRNN [129] Diffusion Convolution Network Traffic Forecasting ST-GCN [131] GCN Action Recognition Space Domain and Spatial MethodsPATCHY-SAN [132] Convolutional Network Runtime Analysis, Feature Visualization, Graph Classification Neural FP [133] Sub-graph Classification DCNN [134] DCNN Classification DGCN [137] Graph-Structure-Based Convolution, PPMI-Based Convolution.Classification SSE [138] Vertex Classification LGCN [140] Convolutional Neural Network Vertex Classification STGCN [130] Gated Sequential Convolution Traffic Forecasting Deep Learning Model Based MethodsGATs [143] Attention Neural NetworkClassification GAAN [144] Vertex Classification GAM [145] Graph Classification Aws [146] Auto-encoder Neural NetworkLink Prediction, Sensitivity Analysis, Vertex Classification SDNE [149] Classification, Link Prediction, Visualization DNGR [148] Clustering, Visualization DRNE [152] Regular Equivalence Predic- tion, Structural Role Classifica- tion, Network Visualization MolGAN [155] Generative Neural NetworkGenerative Model DGMG [156] Molecule Generation DCRNN [129] Diffusion Convolution Network Traffic Forecasting STGCN [130] Gated Sequential Convolution ST-GCN [131] GCNs Action Recognition used [165]. Since social relationships such as friendships are the basis of social networks in real world, automatically interpreting these relationships is important for understanding human behaviors. GRM introduces GGNNs to learn a propa- gation mechanism. Image classification is a classical problem, in which GNNs have demonstrated promising performance. Visual question answering (VQA) is a learning task that in- volves both computer vision and natural language processing. A VQA system takes the form of a certain pictures and its open natural language question as input, in order to generate a natural language answer as output. Generally speaking, VQA is question-and-answer for a given picture. GGNNs have been exploited to help with VQA [166].D. Science Graph learning has been widely adopted in science. Model- ing real-world physical systems is one of the most fundamental perspectives in understanding human intelligence. Represent- ing objects as vertices and relations as edges between them is a simple but effective way to perform physics. Battaglia et al. [167] proposed interaction networks (IN) to predict and infer abundant physical systems, in which IN takes objects and relationships as input. Based on IN, the interactions can be reasoned and the effects can be applied. Therefore, physical dynamics can be predicted. Visual interaction networks (VIN) can make predictions from pixels by firstly learning a stateIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 14 code from two continuous input frames per object [168]. Other graph networks based models have been developed to address chemistry and biology problems. Calculating molecu- lar fingerprints, i.e., using feature vectors to represent molec- ular, is a central step. Researchers [169] proposed neural graph fingerprints using GCNs to calculate substructure feature vectors. Some studies focused on protein interface prediction. This is a challenging issue with significant applications in biology. Besides, GNNs can be used in biomedical engineering as well. Based on protein-protein interaction networks, Rhee et al. [170] used graph convolution and protein relation networks to classify breast cancer subtypes. E. Knowledge Graphs Various heterogeneous objects and relationships are re- garded as the basis for a knowledge graph [171]. GNNs can be applied in knowledge base completion (KBC) for solving the out-of-knowledge-base (OOKB) entity problem [172]. The OOKB entities are connected to existing entities. Therefore, the embedding of OOKB entities can be aggregated from existing entities. Such kind of algorithms achieve reasonable performance in both settings of KBC and OOKB. Likewise, GCNs can also be used to solve the problem of cross-lingual knowledge graph alignment. The main idea of the model is to embed entities from different languages into an integrated em- bedding space. Then the model aligns these entities according to their embedding similarities. Generally speaking, knowledge graph embedding can be categorized into two types: translational distance models and semantic matching models. Translational distance models aim to learn the low dimensional vector of entities in a knowledge graph by employing distance-based scoring functions. These methods calculate the plausibility as the distance between two entities after a translation measured by the relationships between them. Among current translational distance models, TransE [173] is the most influential one. TransE can model the relationship of entities by interpreting them as translations op- erating on the low dimensional embedding. Inspired by TranE, TranH [174] was proposed to overcome the disadvantages of TransE in dealing with 1-to-N, N-to-1, and N-to-N rela- tions by introducing relation-specific hyperplanes. Instead of hyperplanes, TransR [175] introduces relation-specific spaces to solve the flows in TransE. Meanwhile, various extensions of TransE have been proposed to enhance knowledge graph embeddings, such as TransD [176] and TransF [177]. On the basis of TransE, DeepPath [178] incorporates reinforcement learning methods for learning relational paths in knowledge graphs. By designing a complex reward function involving accuracy, efficiency and path diversity, the path finding process is better controlled and more flexible. Semantic matching models utilize the similarity-based scor- ing functions. They measure the plausibility among entities by matching latent semantics of entities and relations in low dimensional vector space. Typical models of this type include RESCAL [179], DistMult [180], ANALOGY [181], etc.F . Combinatorial Optimization Classical problems such as traveling salesman problem (TSP) and minimum spanning tree (MST) have been solved by using different heuristic solutions. Recently, deep neural networks have been applied to these problems. Some solutions make further use of GNNs thanks to their structures. Bello et al. [182] first proposed such kind of methods to solve TSP. Their method mainly contains two steps, i.e., a parameterized reward pointer network and a strategy gradient module for training. Khalil et al. [183] improved this work with GNN and achieved better performance by two main procedures. First, they used structure2vec to achieve vertex embedding and then input them into Q-learning module for decision-making. This work also proves the embedding ability of GNN. Nowak et al. [184] focused on the secondary assignment problem, i.e., measuring the similarity of two graphs. The GNN model learns each graph’s vertex embedding and uses the attention mechanism to match the two graphs. Other studies use GNNs directly as the classifiers, which can perform the intensive prediction on graphs with two sides. The rest of the model facilitates diverse choices and effective training. IV. O PEN ISSUES In this section, we briefly summarize several future research directions and open issues for graph learning. Dynamic Graph Learning : For the purpose of graph learn- ing, most existing algorithms are suitable for static networks without specific constraints. However, dynamic networks such as traffic networks vary over time. Therefore, they are hard to deal with. Dynamic graph learning algorithms have rarely been studied in the literature. It is of significant importance that dynamic graph learning algorithms are designed to maintain good performance, especially in the case of dynamic graphs. Generative Graph Learning : Inspired by the generative adversarial networks, generative graph learning algorithms can unify the generative and discriminative models by playing a game-theoretical min-max game. This generative graph learn- ing method can be used for link prediction, network evolution, and recommendation by boosting the performance of genera- tive and discriminative models alternately and iteratively. Fair Graph Learning : Most graph learning algorithms rely on deep neural networks, and the resulting vectors may have captured undesired sensitive information. The bias existing in the network is reinforced, and hence it is of significant importance to integrate the fair metrics into the graph learning algorithms to address the inherent bias issue. Interpretability of Graph Learning : The models of graph learning are generally complex by incorporating both graph structure and feature information. The interpretability of graph learning (based) algorithms remains unsolved since the struc- tures of graph learning algorithms are still a black box. For example, drug discovery can be achieved by graph learning al- gorithms. However, it is unknown how this drug is discovered as well as the reason behind this discovery. The interpretability behind graph learning needs to be further studied.IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 15 V. C ONCLUSION This survey gives a general description of graph learning, and provides a comprehensive review of the state-of-the-art graph learning methods. We examined existing graph learning methods under four categories: graph signal processing based methods, matrix factorization based methods, random walk based methods, and deep learning based methods. The ap- plications of graph learning methods mainly under these four categories in areas such as text, images, science, knowledge graphs, and combinatorial optimization are outlined. We also discuss some future research directions in the field of graph learning. Graph learning is currently a hot area which is grow- ing at an unprecedented speed. We do hope that this survey will help researchers and practitioners with their research and development in graph learning and related areas. ACKNOWLEDGMENTS The authors would like to thank Prof. Hussein Abbass at University of New South Wales, Yuchen Sun, Jiaying Liu, Hao Ren at Dalian University of Technology, and anonymous reviewers for their valuable comments and suggestions. REFERENCES [1] S. Fortunato, C. T. Bergstrom, K. B ¨orner, J. A. Evans, D. Helbing, S. Milojevi ´c, A. M. Petersen, F. Radicchi, R. Sinatra, B. Uzzi et al. , “Science of science,” Science , vol. 359, no. 6379, 2018, eaao0185. [2] J. Liu, J. Ren, W. Zheng, L. Chi, I. Lee, and F. Xia, “Web of scholars: A scholar knowledge graph,” in the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR) , 2020, pp. 2153–2156. [3] J. Liu, F. Xia, L. Wang, B. Xu, X. Kong, H. Tong, and I. King, “Shifu2: A network representation learning based model for advisor- advisee relationship mining,” IEEE Transactions on Knowledge and Data Engineering , vol. 33, no. 4, pp. 1763–1777, 2021. [4] D. Zhang, J. Yin, X. Zhu, and C. Zhang, “Network representation learning: A survey,” IEEE Transactions on Big Data , vol. 6, no. 1, pp. 3–28, 2020. [5] K. Sun, J. Liu, S. Yu, B. Xu, and F. Xia, “Graph force learning,” inIEEE International Conference on Big Data (BigData) , 2020, pp. 2987–2994. [6] F. Xia, J. Wang, X. Kong, D. Zhang, and Z. Wang, “Ranking station im- portance with human mobility patterns using subway network datasets,” IEEE Transactions on Intelligent Transportation Systems , vol. 21, no. 7, pp. 2840–2852, 2020. [7] A. Grover and J. Leskovec, “node2vec: Scalable feature learning for networks,” in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining . ACM, 2016, pp. 855–864. [8] K. Sun, L. Wang, B. Xu, W. Zhao, S. W. Teng, and F. Xia, “Net- work representation learning: From traditional feature learning to deep learning,” IEEE Access , vol. 8, no. 1, pp. 205 600–205 617, 2020. [9] S. Yu, F. Xia, J. Xu, Z. Chen, and I. Lee, “Offer: A motif dimensional framework for network representation learning,” in The 29th ACM International Conference on Information and Knowledge Management (CIKM) , 2020, pp. 3349–3352. [10] Y . Bengio, A. Courville, and P. Vincent, “Representation learning: A review and new perspectives,” IEEE Transactions on Pattern Analysis and Machine Intelligence , vol. 35, no. 8, pp. 1798–1828, 2013. [11] T. Guo, F. Xia, S. Zhen, X. Bai, D. Zhang, Z. Liu, and J. Tang, “Graduate employment prediction with bias,” in Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI) , 2020, pp. 670–677. [12] F. Xia, A. M. Ahmed, L. T. Yang, J. Ma, and J. J. Rodrigues, “Ex- ploiting social relationship to enable efficient replica allocation in ad- hoc social networks,” IEEE Transactions on Parallel and Distributed Systems , vol. 25, no. 12, pp. 3167–3176, 2014. [13] J. Zhang, W. Wang, F. Xia, Y .-R. Lin, and H. Tong, “Data-driven computational social science: A survey,” Big Data Research , vol. 21, p. 100145, 2020.[14] F. Xia, A. M. Ahmed, L. T. Yang, and Z. Luo, “Community-based event dissemination with optimal load balancing,” IEEE Transactions on Computers , vol. 64, no. 7, pp. 1857–1869, 2014. [15] F. Xia, H. Wei, S. Yu, D. Zhang, and B. Xu, “A survey of measures for network motifs,” IEEE Access , vol. 7, no. 1, pp. 106 576–106 587, 2019. [16] J. Leskovec and C. Faloutsos, “Sampling from large graphs,” in Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining . ACM, 2006, pp. 631–636. [17] X. Wang, P. Liu, and Y . Gu, “Local-set-based graph signal reconstruc- tion,” IEEE Transactions on Signal Processing , vol. 63, no. 9, pp. 2432–2444, 2015. [18] Z. Zhang, P. Cui, and W. Zhu, “Deep learning on graphs: A survey,” IEEE Transactions on Knowledge and Data Engineering , 2020. [19] J. Xu, S. Yu, K. Sun, J. Ren, I. Lee, S. Pan, and F. Xia, “Multivariate relations aggregation learning in social networks,” in ACM/IEEE Joint Conference on Digital Libraries (JCDL) , 2020, pp. 77–86. [20] H. D. Bedru, S. Yu, X. Xiao, D. Zhang, L. Wan, H. Guo, and F. Xia, “Big networks: A survey,” Computer Science Review , vol. 37, p. 100247, 2020. [21] P. Goyal and E. Ferrara, “Graph embedding techniques, applications, and performance: A survey,” Knowledge-Based Systems , vol. 151, pp. 78–94, 2018. [22] H. Cai, V . W. Zheng, and K. C.-C. Chang, “A comprehensive survey of graph embedding: Problems, techniques, and applications,” IEEE Transactions on Knowledge and Data Engineering , vol. 30, no. 9, pp. 1616–1637, 2018. [23] M. Nickel, K. Murphy, V . Tresp, and E. Gabrilovich, “A review of relational machine learning for knowledge graphs,” Proceedings of the IEEE , vol. 104, no. 1, pp. 11–33, 2016. [24] L. Akoglu, H. Tong, and D. Koutra, “Graph based anomaly detection and description: a survey,” Data Mining and Knowledge Discovery , vol. 29, no. 3, pp. 626–688, 2015. [25] Z. Wu, S. Pan, F. Chen, G. Long, C. Zhang, and P. S. Yu, “A comprehensive survey on graph neural networks,” IEEE Transactions on Neural Networks and Learning Systems , vol. 32, no. 1, pp. 4–24, 2021. [26] A. Ortega, P. Frossard, J. Kova ˇcevi´c, J. M. Moura, and P. Van- dergheynst, “Graph signal processing: Overview, challenges, and ap- plications,” Proceedings of the IEEE , vol. 106, no. 5, pp. 808–828, 2018. [27] W. Huang, T. A. Bolton, J. D. Medaglia, D. S. Bassett, A. Ribeiro, and D. Van De Ville, “A graph signal processing perspective on functional brain imaging,” Proceedings of the IEEE , vol. 106, no. 5, pp. 868–885, 2018. [28] F. Xia, W. Wang, T. M. Bekele, and H. Liu, “Big scholarly data: A survey,” IEEE Transactions on Big Data , vol. 3, no. 1, pp. 18–35, 2017. [29] Y . LeCun, Y . Bengio, and G. Hinton, “Deep learning,” Nature , vol. 521, no. 7553, p. 436, 2015. [30] J. Liu, X. Kong, F. Xia, X. Bai, L. Wang, Q. Qing, and I. Lee, “Artificial intelligence in the 21st century,” IEEE Access , vol. 6, pp. 34 403– 34 421, 2018. [31] A. Sandryhaila and J. M. Moura, “Discrete signal processing on graphs,” IEEE Transactions on Signal Processing , vol. 61, no. 7, pp. 1644–1656, 2013. [32] D. Shuman, S. Narang, P. Frossard, A. Ortega, and P. Vandergheynst, “The emerging field of signal processing on graphs: Extending high- dimensional data analysis to networks and other irregular domains,” IEEE Signal Processing Magazine , vol. 3, no. 30, pp. 83–98, 2013. [33] M. Puschel and J. M. Moura, “Algebraic signal processing theory: Foundation and 1-d time,” IEEE Transactions on Signal Processing , vol. 56, no. 8, pp. 3572–3585, 2008. [34] A. Sandryhaila and J. M. Moura, “Discrete signal processing on graphs: Graph filters,” in 2013 IEEE International Conference on Acoustics, Speech and Signal Processing . IEEE, 2013, pp. 6163–6166. [35] S. Chen, R. Varma, A. Sandryhaila, and J. Kova ˇcevi´c, “Discrete signal processing on graphs: Sampling theory,” IEEE Transactions on Signal Processing , vol. 63, no. 24, pp. 6510–6523, 2015. [36] U. V on Luxburg, “A tutorial on spectral clustering,” Statistics and Computing , vol. 17, no. 4, pp. 395–416, 2007. [37] X. Zhu and M. Rabbat, “Graph spectral compressed sensing for sensor networks,” in 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) . IEEE, 2012, pp. 2865–2868. [38] A. Gavili and X.-P. Zhang, “On the shift operator, graph frequency, and optimal filtering in graph signal processing,” IEEE Transactions on Signal Processing , vol. 65, no. 23, pp. 6303–6318, 2017.IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 16 [39] B. Pasdeloup, M. Rabbat, V . Gripon, D. Pastor, and G. Mercier, “Graph reconstruction from the observation of diffused signals,” in 2015 53rd Annual Allerton Conference on Communication, Control, and Computing (Allerton) . IEEE, 2015, pp. 1386–1390. [40] A. Anis, A. Gadde, and A. Ortega, “Towards a sampling theorem for signals on arbitrary graphs,” in 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) . IEEE, 2014, pp. 3864–3868. [41] G. Puy, N. Tremblay, R. Gribonval, and P. Vandergheynst, “Random sampling of bandlimited signals on graphs,” Applied and Computa- tional Harmonic Analysis , vol. 44, no. 2, pp. 446–475, 2018. [42] H. Shomorony and A. S. Avestimehr, “Sampling large data on graphs,” in2014 IEEE Global Conference on Signal and Information Processing (GlobalSIP) . IEEE, 2014, pp. 933–936. [43] L. F. Chamon and A. Ribeiro, “Greedy sampling of graph signals,” IEEE Transactions on Signal Processing , vol. 66, no. 1, pp. 34–47, 2018. [44] A. G. Marques, S. Segarra, G. Leus, and A. Ribeiro, “Sampling of graph signals with successive local aggregations.” IEEE Transactions Signal Processing , vol. 64, no. 7, pp. 1832–1843, 2016. [45] S. K. Narang, A. Gadde, and A. Ortega, “Signal processing techniques for interpolation in graph structured data,” in 2013 IEEE International Conference on Acoustics, Speech and Signal Processing . IEEE, 2013, pp. 5445–5449. [46] A. Gadde and A. Ortega, “A probabilistic interpretation of sampling theory of graph signals,” in 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) . IEEE, 2015, pp. 3257–3261. [47] X. Wang, M. Wang, and Y . Gu, “A distributed tracking algorithm for reconstruction of graph signals,” IEEE Journal of Selected Topics in Signal Processing , vol. 9, no. 4, pp. 728–740, 2015. [48] P. Di Lorenzo, S. Barbarossa, P. Banelli, and S. Sardellitti, “Adaptive least mean squares estimation of graph signals,” IEEE Transactions on Signal and Information Processing over Networks , vol. 2, no. 4, pp. 555–568, 2016. [49] D. Romero, M. Ma, and G. B. Giannakis, “Kernel-based reconstruction of graph signals.” IEEE Transactions Signal Processing , vol. 65, no. 3, pp. 764–778, 2017. [50] M. Nagahara, “Discrete signal reconstruction by sum of absolute values,” IEEE Signal Processing Letters , vol. 22, no. 10, pp. 1575– 1579, 2015. [51] S. Chen, R. Varma, A. Singh, and J. Kova ˇcevi´c, “Signal representations on graphs: Tools and applications,” arXiv preprint arXiv:1512.05406 , 2015. [52] S. Segarra, A. G. Marques, G. Leus, and A. Ribeiro, “Reconstruction of graph signals through percolation from seeding nodes,” IEEE Transactions on Signal Processing , vol. 64, no. 16, pp. 4363–4378, 2016. [53] F. Xia, J. Liu, J. Ren, W. Wang, and X. Kong, “Turing number: How far are you to a. m. turing award?” ACM SIGWEB Newsletter , vol. Autumn, 2020, article No.: 5. [54] H. E. Egilmez, E. Pavez, and A. Ortega, “Graph learning from data under laplacian and structural constraints,” IEEE Journal of Selected Topics in Signal Processing , vol. 11, no. 6, pp. 825–841, 2017. [55] X. Dong, D. Thanou, P. Frossard, and P. Vandergheynst, “Learning laplacian matrix in smooth graph signal representations,” IEEE Trans- actions on Signal Processing , vol. 64, no. 23, pp. 6160–6173, 2016. [56] V . Kalofolias, “How to learn a graph from smooth signals,” in Artificial Intelligence and Statistics , 2016, pp. 920–929. [57] E. Pavez and A. Ortega, “Generalized laplacian precision matrix estimation for graph signal processing,” in 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) . IEEE, 2016, pp. 6350–6354. [58] E. Pavez, H. E. Egilmez, and A. Ortega, “Learning graphs with monotone topology properties and multiple connected components,” IEEE Transactions on Signal Processing , vol. 66, no. 9, pp. 2399– 2413, 2018. [59] B. Pasdeloup, V . Gripon, G. Mercier, D. Pastor, and M. G. Rabbat, “Characterization and inference of graph diffusion processes from observations of stationary signals,” IEEE Transactions on Signal and Information Processing over Networks , vol. 4, no. 3, pp. 481–496, 2018. [60] S. Segarra, A. G. Marques, G. Mateos, and A. Ribeiro, “Network topology inference from spectral templates,” IEEE Transactions on Signal and Information Processing over Networks , vol. 3, no. 3, pp. 467–483, 2017.[61] D. Thanou, X. Dong, D. Kressner, and P. Frossard, “Learning heat diffusion graphs,” IEEE Transactions on Signal and Information Pro- cessing over Networks , vol. 3, no. 3, pp. 484–499, 2017. [62] J. Mei and J. M. Moura, “Signal processing on graphs: Causal modeling ofunstructured data,” IEEE Transactions on Signal Processing , vol. 65, no. 8, pp. 2077–2092, 2016. [63] S. Segarra, G. Mateos, A. G. Marques, and A. Ribeiro, “Blind iden- tification of graph filters,” IEEE Transactions on Signal Processing , vol. 65, no. 5, pp. 1146–1159, 2017. [64] F. Xia, N. Y . Asabere, A. M. Ahmed, J. Li, and X. Kong, “Mobile multimedia recommendation in smart communities: A survey,” IEEE Access , vol. 1, no. 1, pp. 606–624, 2013. [65] W. Huang, A. G. Marques, and A. R. Ribeiro, “Rating prediction via graph signal processing,” IEEE Transactions on Signal Processing , vol. 66, no. 19, pp. 5066–5081, 2018. [66] F. Xia, H. Liu, I. Lee, and L. Cao, “Scientific article recommendation: Exploiting common author relations and historical preferences,” IEEE Transactions on Big Data , vol. 2, no. 2, pp. 101–112, 2016. [67] X. He and P. Niyogi, “Locality preserving projections,” in Advances in Neural Information Processing Systems , 2004, pp. 153–160. [68] M. Chen, I. W. Tsang, M. Tan, and T. J. Cham, “A unified feature selection framework for graph embedding on high dimensional data,” IEEE Transactions on Knowledge and Data Engineering , vol. 27, no. 6, pp. 1465–1477, 2014. [69] S. Yan, D. Xu, B. Zhang, H.-J. Zhang, Q. Yang, and S. Lin, “Graph embedding and extensions: A general framework for dimensionality reduction,” IEEE Transactions on Pattern Analysis & Machine Intelli- gence , no. 1, pp. 40–51, 2007. [70] I. Borg and P. Groenen, “Modern multidimensional scaling: Theory and applications,” Journal of Educational Measurement , vol. 40, no. 3, pp. 277–280, 2003. [71] M. Balasubramanian and E. L. Schwartz, “The isomap algorithm and topological stability,” Science , vol. 295, no. 5552, pp. 7–7, 2002. [72] W. N. Anderson Jr and T. D. Morley, “Eigenvalues of the laplacian of a graph,” Linear and Multilinear Algebra , vol. 18, no. 2, pp. 141–145, 1985. [73] S. T. Roweis and L. K. Saul, “Nonlinear dimensionality reduction by locally linear embedding,” Science , vol. 290, no. 5500, pp. 2323–2326, 2000. [74] R. Jiang, W. Fu, L. Wen, S. Hao, and R. Hong, “Dimensionality reduc- tion on anchorgraph with an efficient locality preserving projection,” Neurocomputing , vol. 187, pp. 109–118, 2016. [75] L. Wan, Y . Yuan, F. Xia, and H. Liu, “To your surprise: Identifying serendipitous collaborators,” IEEE Transactions on Big Data , 2019. [76] Y . Yang, F. Nie, S. Xiang, Y . Zhuang, and W. Wang, “Local and global regressive mapping for manifold learning with out-of-sample extrapo- lation,” in Twenty-Fourth AAAI Conference on Artificial Intelligence , 2010, pp. 649–654. [77] S. Xiang, F. Nie, C. Zhang, and C. Zhang, “Nonlinear dimensionality reduction with local spline embedding,” IEEE Transactions on Knowl- edge and Data Engineering , vol. 21, no. 9, pp. 1285–1298, 2008. [78] D. Cai, X. He, and J. Han, “Spectral regression: A unified subspace learning framework for content-based image retrieval,” in Proceedings of the 15th ACM international conference on Multimedia . ACM, 2007, pp. 403–412. [79] X. He, W.-Y . Ma, and H.-J. Zhang, “Learning an image manifold for retrieval,” in Proceedings of the 12th annual ACM international conference on Multimedia . ACM, 2004, pp. 17–23. [80] K. Allab, L. Labiod, and M. Nadif, “A semi-nmf-pca unified framework for data clustering,” IEEE Transactions on Knowledge and Data Engineering , vol. 29, no. 1, pp. 2–16, 2017. [81] L. Vandenberghe and S. Boyd, “Semidefinite programming,” SIAM Review , vol. 38, no. 1, pp. 49–95, 1996. [82] G. H. Golub and C. Reinsch, “Singular value decomposition and least squares solutions,” Numerische Mathematik , vol. 14, no. 5, pp. 403– 420, 1970. [83] A. Ahmed, N. Shervashidze, S. Narayanamurthy, V . Josifovski, and A. J. Smola, “Distributed large-scale natural graph factorization,” in Proceedings of the 22nd International Conference on World Wide Web . ACM, 2013, pp. 37–48. [84] C. Yang, Z. Liu, D. Zhao, M. Sun, and E. Y . Chang, “Network representation learning with rich text information,” in International Joint Conference on Artificial Intelligence , 2015, pp. 2111–2117. [85] F. Xia, J. Liu, H. Nie, Y . Fu, L. Wan, and X. Kong, “Random walks: A review of algorithms and applications,” IEEE Transactions on Emerging Topics in Computational Intelligence , vol. 4, no. 2, pp. 95–107, 2019.IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 17 [86] F. Xia, Z. Chen, W. Wang, J. Li, and L. T. Yang, “Mvcwalker: Random walk-based most valuable collaborators recommendation exploiting academic factors,” IEEE Transactions on Emerging Topics in Com- puting , vol. 2, no. 3, pp. 364–375, 2014. [87] M. A. Al-Garadi, K. D. Varathan, S. D. Ravana, E. Ahmed, G. Mujtaba, M. U. S. Khan, and S. U. Khan, “Analysis of online social network connections for identification of influential users: Survey and open research issues,” ACM Computing Surveys (CSUR) , vol. 51, no. 1, pp. 1–37, 2018. [88] B. Perozzi, R. Al-Rfou, and S. Skiena, “Deepwalk: Online learning of social representations,” in Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining . ACM, 2014, pp. 701–710. [89] O. Levy and Y . Goldberg, “Neural word embedding as implicit matrix factorization,” in Advances in Neural Information Processing Systems , 2014, pp. 2177–2185. [90] X. Rong, “word2vec parameter learning explained,” arXiv preprint arXiv:1411.2738 , 2014. [91] Y . Goldberg and O. Levy, “word2vec explained: Deriving mikolov et al.’s negative-sampling word-embedding method,” arXiv preprint arXiv:1402.3722 , 2014. [92] J. Tang, M. Qu, M. Wang, M. Zhang, J. Yan, and Q. Mei, “Line: Large-scale information network embedding,” in Proceedings of the 24th International Conference on World Wide Web , 2015, pp. 1067– 1077. [93] W. Wang, J. Liu, Z. Yang, X. Kong, and F. Xia, “Sustainable collabora- tor recommendation based on conference closure,” IEEE Transactions on Computational Social Systems , vol. 6, no. 2, pp. 311–322, 2019. [94] C. Tu, W. Zhang, Z. Liu, M. Sun et al. , “Max-margin deepwalk: Discriminative learning of network representation.” in International Joint Conference on Artificial Intelligence , 2016, pp. 3889–3895. [95] L. F. Ribeiro, P. H. Saverese, and D. R. Figueiredo, “struc2vec: Learning node representations from structural identity,” in Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining . ACM, 2017, pp. 385–394. [96] Z. Yang, W. W. Cohen, and R. Salakhutdinov, “Revisiting semi- supervised learning with graph embeddings,” in Proceedings of The 33rd International Conference on Machine Learning , 2016, pp. 40–48. [97] B. Adhikari, Y . Zhang, N. Ramakrishnan, and B. A. Prakash, “Dis- tributed representations of subgraphs,” in 2017 IEEE International Conference on Data Mining Workshops (ICDMW) . IEEE, 2017, pp. 111–117. [98] A. Narayanan, M. Chandramohan, R. Venkatesan, L. Chen, Y . Liu, and S. Jaiswal, “graph2vec: Learning distributed representations of graphs,” arXiv preprint arXiv:1707.05005 , 2017. [99] A. R. Benson, D. F. Gleich, and L.-H. Lim, “The spacey random walk: A stochastic process for higher-order data,” SIAM Review , vol. 59, no. 2, pp. 321–345, 2017. [100] H. Wang, J. Wang, J. Wang, M. Zhao, W. Zhang, F. Zhang, X. Xie, and M. Guo, “Graphgan: Graph representation learning with generative adversarial nets,” in Thirty-Second AAAI Conference on Artificial Intelligence , 2018, pp. 2508–2515. [101] A. Bojchevski, O. Shchur, D. Z ¨ugner, and S. G ¨unnemann, “Netgan: Generating graphs via random walks,” Proceedings of the 35th Inter- national Conference on Machine Learning (ICML 2018) , pp. 609–618, 2018. [102] C. Shi, Y . Li, J. Zhang, Y . Sun, and S. Y . Philip, “A survey of heterogeneous information network analysis,” IEEE Transactions on Knowledge and Data Engineering , vol. 29, no. 1, pp. 17–37, 2017. [103] N. Lao and W. W. Cohen, “Relational retrieval using a combination of path-constrained random walks,” Machine learning , vol. 81, no. 1, pp. 53–67, 2010. [104] Q. Wang, Z. Mao, B. Wang, and L. Guo, “Knowledge graph embed- ding: A survey of approaches and applications,” IEEE Transactions on Knowledge and Data Engineering , vol. 29, no. 12, pp. 2724–2743, 2017. [105] N. Lao, T. Mitchell, and W. W. Cohen, “Random walk inference and learning in a large scale knowledge base,” in Proceedings of the Conference on Empirical Methods in Natural Language Processing . Association for Computational Linguistics, 2011, pp. 529–539. [106] M. Gardner, P. P. Talukdar, B. Kisiel, and T. Mitchell, “Improving learning and inference in a large knowledge-base using latent syntactic cues,” in Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing , 2013, pp. 833–838. [107] M. Gardner, P. Talukdar, J. Krishnamurthy, and T. Mitchell, “Incorpo- rating vector space similarity in random walk inference over knowledgebases,” in Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) , 2014, pp. 397–406. [108] W. Y . Wang and W. W. Cohen, “Joint information extraction and rea- soning: A scalable statistical relational learning approach,” in Proceed- ings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers) , 2015, pp. 355–364. [109] Q. Liu, L. Jiang, M. Han, Y . Liu, and Z. Qin, “Hierarchical random walk inference in knowledge graphs,” in Proceedings of the 39th International ACM SIGIR conference on Research and Development in Information Retrieval . ACM, 2016, pp. 445–454. [110] T.-y. Fu, W.-C. Lee, and Z. Lei, “Hin2vec: Explore meta-paths in heterogeneous information networks for representation learning,” in Proceedings of the 2017 ACM on Conference on Information and Knowledge Management . ACM, 2017, pp. 1797–1806. [111] Y . Dong, N. V . Chawla, and A. Swami, “metapath2vec: Scalable representation learning for heterogeneous networks,” in Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining , 2017, pp. 135–144. [112] R. Hussein, D. Yang, and P. Cudr ´e-Mauroux, “Are meta-paths neces- sary?: Revisiting heterogeneous graph embeddings,” in Proceedings of the 27th ACM International Conference on Information and Knowledge Management . ACM, 2018, pp. 437–446. [113] G. Wan, B. Du, S. Pan, and G. Haffari, “Reinforcement learning based meta-path discovery in large-scale heterogeneous information networks,” in AAAI Conference on Artificial Intelligence . AAAI, apr 2020. [114] C. Shi, B. Hu, W. X. Zhao, and S. Y . Philip, “Heterogeneous informa- tion network embedding for recommendation,” IEEE Transactions on Knowledge and Data Engineering , vol. 31, no. 2, pp. 357–370, 2019. [115] J. Tang, M. Qu, and Q. Mei, “Pte: Predictive text embedding through large-scale heterogeneous text networks,” in Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining . ACM, 2015, pp. 1165–1174. [116] C. Zhang, A. Swami, and N. V . Chawla, “Shne: Representation learning for semantic-associated heterogeneous networks,” in Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining . ACM, 2019, pp. 690–698. [117] M. Hou, J. Ren, D. Zhang, X. Kong, D. Zhang, and F. Xia, “Network embedding: Taxonomies, frameworks and applications,” Computer Sci- ence Review , vol. 38, p. 100296, 2020. [118] G. H. Nguyen, J. B. Lee, R. A. Rossi, N. K. Ahmed, E. Koh, and S. Kim, “Continuous-time dynamic network embeddings,” in Companion Proceedings of the The Web Conference , 2018, pp. 969– 976. [119] Y . Zuo, G. Liu, H. Lin, J. Guo, X. Hu, and J. Wu, “Embedding temporal network via neighborhood formation,” in Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining . ACM, 2018, pp. 2857–2866. [120] W. Hamilton, Z. Ying, and J. Leskovec, “Inductive representation learn- ing on large graphs,” in Advances in Neural Information Processing Systems , 2017, pp. 1024–1034. [121] M. Gori, G. Monfardini, and F. Scarselli, “A new model for learning in graph domains,” in IEEE International Joint Conference on Neural Networks , vol. 2. IEEE, 2005, pp. 729–734. [122] J. Bruna, W. Zaremba, A. Szlam, and Y . LeCun, “Spectral net- works and locally connected networks on graphs,” arXiv preprint arXiv:1312.6203 , 2013. [123] M. Henaff, J. Bruna, and Y . LeCun, “Deep convolutional networks on graph-structured data,” Advances in Neural Information Processing Systems , pp. 1–9, 2015. [124] M. Defferrard, X. Bresson, and P. Vandergheynst, “Convolutional neural networks on graphs with fast localized spectral filtering,” in Advances in Neural Information Processing Systems , 2016, pp. 3844– 3852. [125] T. N. Kipf and M. Welling, “Semi-supervised classification with graph convolutional networks,” International Conference on Learning Representations , 2017. [126] F. Monti, D. Boscaini, J. Masci, E. Rodola, J. Svoboda, and M. M. Bronstein, “Geometric deep learning on graphs and manifolds using mixture model cnns,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition , 2017, pp. 5115–5124. [127] Z. Zhou and X. Li, “Graph convolution: a high-order and adaptive approach,” arXiv preprint arXiv:1706.09916 , 2017. [128] F. Manessi, A. Rozza, and M. Manzo, “Dynamic graph convolutional networks,” arXiv preprint arXiv:1704.06199 , 2017.IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 18 [129] Y . Li, R. Yu, C. Shahabi, and Y . Liu, “Diffusion convolutional re- current neural network: Data-driven traffic forecasting,” International Conference on Learning Representations , 2017. [130] B. Yu, H. Yin, and Z. Zhu, “Spatio-temporal graph convolutional networks: A deep learning framework for traffic forecasting,” Proceed- ings of the Twenty-Seventh International Joint Conference on Artificial Intelligence , pp. 3634–3640, 2017. [131] S. Yan, Y . Xiong, and D. Lin, “Spatial temporal graph convolutional networks for skeleton-based action recognition,” in Thirty-Second AAAI Conference on Artificial Intelligence , 2018, pp. 3634–3640. [132] M. Niepert, M. Ahmed, and K. Kutzkov, “Learning convolutional neural networks for graphs,” in International Conference on Machine Learning , 2016, pp. 2014–2023. [133] D. K. Duvenaud, D. Maclaurin, J. Iparraguirre, R. Bombarell, T. Hirzel, A. Aspuru-Guzik, and R. P. Adams, “Convolutional networks on graphs for learning molecular fingerprints,” in Advances in Neural Information Processing Systems , 2015, pp. 2224–2232. [134] J. Atwood and D. Towsley, “Diffusion-convolutional neural networks,” inAdvances in Neural Information Processing Systems , 2016, pp. 1993–2001. [135] J. Gilmer, S. S. Schoenholz, P. F. Riley, O. Vinyals, and G. E. Dahl, “Neural message passing for quantum chemistry,” in Proceedings of the 34th International Conference on Machine Learning-Volume 70 . JMLR. org, 2017, pp. 1263–1272. [136] M. Allamanis, M. Brockschmidt, and M. Khademi, “Learning to represent programs with graphs,” International Conference on Learning Representations , 2018. [137] C. Zhuang and Q. Ma, “Dual graph convolutional networks for graph- based semi-supervised classification,” in Proceedings of the Web Con- ference , 2018, pp. 499–508. [138] H. Dai, Z. Kozareva, B. Dai, A. Smola, and L. Song, “Learning steady- states of iterative algorithms over graphs,” in International Conference on Machine Learning , 2018, pp. 1114–1122. [139] S. Zhu, L. Zhou, S. Pan, C. Zhou, G. Yan, and B. Wang, “GSSNN: Graph smoothing splines neural networks,” in AAAI Conference on Artificial Intelligence . AAAI, apr 2020. [140] H. Gao, Z. Wang, and S. Ji, “Large-scale learnable graph convolutional networks,” in Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining . ACM, 2018, pp. 1416–1424. [141] C. Tran, W.-Y . Shin, A. Spitz, and M. Gertz, “Deepnc: Deep generative network completion,” arXiv preprint arXiv:1907.07381 , 2019. [142] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” in Advances in Neural Information Processing Systems , 2017, pp. 5998– 6008. [143] P. Veli ˇckovi ´c, G. Cucurull, A. Casanova, A. Romero, P. Lio, and Y . Bengio, “Graph attention networks,” International Conference on Learning Representations , 2018. [144] J. Zhang, X. Shi, J. Xie, H. Ma, I. King, and D.-Y . Yeung, “GaAN: Gated attention networks for learning on large and spatiotemporal graphs,” Thirty-Fourth Conference on Uncertainty in Artificial Intel- ligence (UAI) , 2018. [145] J. B. Lee, R. Rossi, and X. Kong, “Graph classification using structural attention,” in Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining . ACM, 2018, pp. 1666–1674. [146] S. Abu-El-Haija, B. Perozzi, R. Al-Rfou, and A. A. Alemi, “Watch your step: Learning node embeddings via graph attention,” in Advances in Neural Information Processing Systems , 2018, pp. 9180–9190. [147] M. Hou, L. Wang, J. Liu, X. Kong, and F. Xia, “A3graph: Adversarial attributed autoencoder for graph representation,” in The 36th ACM Symposium on Applied Computing (SAC) , 2021, pp. 1697–1704. [148] S. Cao, W. Lu, and Q. Xu, “Deep neural networks for learning graph representations,” in Thirtieth AAAI Conference on Artificial Intelligence , 2016, pp. 1145–1152. [149] D. Wang, P. Cui, and W. Zhu, “Structural deep network embedding,” in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining . ACM, 2016, pp. 1225–1234. [150] Y . Qi, Y . Wang, X. Zheng, and Z. Wu, “Robust feature learning by stacked autoencoder with maximum correntropy criterion,” in 2014 IEEE International Conference on Acoustics, Speech and Signal Pro- cessing (ICASSP) . IEEE, 2014, pp. 6716–6720. [151] L. Jing and Y . Tian, “Self-supervised visual feature learning with deep neural networks: A survey,” IEEE Transactions on Pattern Analysis and Machine Intelligence , 2020.[152] K. Tu, P. Cui, X. Wang, P. S. Yu, and W. Zhu, “Deep recursive network embedding with regular equivalence,” in Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining . ACM, 2018, pp. 2357–2366. [153] T. N. Kipf and M. Welling, “Variational graph auto-encoders,” arXiv preprint arXiv:1611.07308 , 2016. [154] S. Pan, R. Hu, S.-f. Fung, G. Long, J. Jiang, and C. Zhang, “Learning graph embedding with adversarial training methods,” IEEE Transac- tions on Cybernetics , 2019. [155] M. Schlichtkrull, T. N. Kipf, P. Bloem, R. Van Den Berg, I. Titov, and M. Welling, “Modeling relational data with graph convolutional networks,” in European Semantic Web Conference . Springer, 2018, pp. 593–607. [156] Y . Li, O. Vinyals, C. Dyer, R. Pascanu, and P. Battaglia, “Learning deep generative models of graphs,” arXiv preprint arXiv:1803.03324 , 2018. [157] Y . Xian, C. H. Lampert, B. Schiele, and Z. Akata, “Zero-shot learn- ing—a comprehensive evaluation of the good, the bad and the ugly,” IEEE Transactions on Pattern Analysis and Machine Intelligence , vol. 41, no. 9, pp. 2251–2265, 2018. [158] M. R. Vyas, H. Venkateswara, and S. Panchanathan, “Leveraging seen and unseen semantic relationships for generative zero-shot learning,” inEuropean Conference on Computer Vision . Springer, 2020, pp. 70–86. [159] Z. Wu, S. Pan, G. Long, J. Jiang, and C. Zhang, “Graph wavenet for deep spatial-temporal graph modeling,” in Proceedings of the 28th International Joint Conference on Artificial Intelligence . AAAI Press, 2019, pp. 1907–1913. [160] L. Yao, C. Mao, and Y . Luo, “Graph convolutional networks for text classification,” in Proceedings of the AAAI Conference on Artificial Intelligence , vol. 33, 2019, pp. 7370–7377. [161] Y . Zhang, Q. Liu, and L. Song, “Sentence-state LSTM for text represen- tation,” The 56th Annual Meeting of the Association for Computational Linguistics , pp. 317–327, 2018. [162] D. Marcheggiani and I. Titov, “Encoding sentences with graph convolu- tional networks for semantic role labeling,” in Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing , 2017, pp. 1506–1515. [163] D. Beck, G. Haffari, and T. Cohn, “Graph-to-sequence learning using gated graph neural networks,” in Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) , 2018, pp. 273–283. [164] H. Peng, J. Li, Y . He, Y . Liu, M. Bao, L. Wang, Y . Song, and Q. Yang, “Large-scale hierarchical text classification with recursively regularized deep graph-cnn,” in Proceedings of the Web Conference , 2018, pp. 1063–1072. [165] Z. Wang, T. Chen, J. Ren, W. Yu, H. Cheng, and L. Lin, “Deep reasoning with knowledge graph for social relationship understanding,” inProceedings of the 27th International Joint Conference on Artificial Intelligence . AAAI Press, 2018, pp. 1021–1028. [166] C.-W. Lee, W. Fang, C.-K. Yeh, and Y .-C. Frank Wang, “Multi-label zero-shot learning with structured knowledge graphs,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition , 2018, pp. 1576–1585. [167] P. Battaglia, R. Pascanu, M. Lai, D. J. Rezende et al. , “Interaction networks for learning about objects, relations and physics,” in Advances in Neural Information Processing systems , 2016, pp. 4502–4510. [168] N. Watters, D. Zoran, T. Weber, P. Battaglia, R. Pascanu, and A. Tac- chetti, “Visual interaction networks: Learning a physics simulator from video,” in Advances in Neural Information Processing systems , 2017, pp. 4539–4547. [169] K. T. Butler, D. W. Davies, H. Cartwright, O. Isayev, and A. Walsh, “Machine learning for molecular and materials science,” Nature , vol. 559, no. 7715, pp. 547–555, 2018. [170] S. Rhee, S. Seo, and S. Kim, “Hybrid approach of relation network and localized graph convolutional filtering for breast cancer subtype clas- sification,” in Proceedings of the 27th International Joint Conference on Artificial Intelligence . AAAI Press, 2018, pp. 3527–3534. [171] S. Ji, S. Pan, E. Cambria, P. Marttinen, and P. S. Yu, “A survey on knowledge graphs: Representation, acquisition and applications,” arXiv preprint arXiv:2002.00388 , 2020. [172] T. Hamaguchi, H. Oiwa, M. Shimbo, and Y . Matsumoto, “Knowledge base completion with out-of-knowledge-base entities: A graph neural network approach,” Transactions of the Japanese Society for Artificial Intelligence , vol. 33, pp. 1–10, 2018.IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 19 [173] A. Bordes, N. Usunier, A. Garcia-Duran, J. Weston, and O. Yakhnenko, “Translating embeddings for modeling multi-relational data,” in Ad- vances in Neural Information Processing Systems , 2013, pp. 2787– 2795. [174] Z. Wang, J. Zhang, J. Feng, and Z. Chen, “Knowledge graph embedding by translating on hyperplanes,” in Twenty-Eighth AAAI Conference on Artificial Intelligence , 2014, pp. 1112–1119. [175] Y . Lin, Z. Liu, M. Sun, Y . Liu, and X. Zhu, “Learning entity and relation embeddings for knowledge graph completion,” in Twenty-ninth AAAI conference on artificial intelligence , 2015, pp. 2181–2187. [176] G. Ji, S. He, L. Xu, K. Liu, and J. Zhao, “Knowledge graph embedding via dynamic mapping matrix,” in Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers) , vol. 1, 2015, pp. 687–696. [177] J. Feng, M. Huang, M. Wang, M. Zhou, Y . Hao, and X. Zhu, “Knowledge graph embedding by flexible translation,” in Fifteenth In- ternational Conference on the Principles of Knowledge Representation and Reasoning , 2016, pp. 557–560. [178] Z. Huang and N. Mamoulis, “Heterogeneous information net- work embedding for meta path based proximity,” arXiv preprint arXiv:1701.05291 , 2017. [179] R. Jenatton, N. L. Roux, A. Bordes, and G. R. Obozinski, “A latent factor model for highly multi-relational data,” in Advances in Neural Information Processing Systems , 2012, pp. 3167–3175. [180] B. Yang, W.-t. Yih, X. He, J. Gao, and L. Deng, “Embedding entities and relations for learning and inference in knowledge bases,” Interna- tional Conference on Learning Representations , 2015. [181] H. Liu, Y . Wu, and Y . Yang, “Analogical inference for multi-relational embeddings,” in Proceedings of the 34th International Conference on Machine Learning-Volume 70 . JMLR. org, 2017, pp. 2168–2178. [182] I. Bello, H. Pham, Q. V . Le, M. Norouzi, and S. Bengio, “Neural combinatorial optimization with reinforcement learning,” International Conference on Learning Representations , 2017. [183] E. Khalil, H. Dai, Y . Zhang, B. Dilkina, and L. Song, “Learning combinatorial optimization algorithms over graphs,” in Advances in Neural Information Processing Systems , 2017, pp. 6348–6358. [184] A. Nowak, S. Villar, A. S. Bandeira, and J. Bruna, “Revised note on learning quadratic assignment with graph neural networks,” in 2018 IEEE Data Science Workshop (DSW) . IEEE, 2018, pp. 1–5. Feng Xia (M’07SM’12) received the B.Sc. and Ph.D. degrees from Zhejiang University, Hangzhou, China. He is currently an Associate Professor and Discipline Leader in School of Engineering, IT and Physical Sciences, Federation University Australia. Dr. Xia has published 2 books and over 300 scientific papers in international journals and conferences. His research interests include data science, computa- tional intelligence, social computing, and systems engineering. He is a Senior Member of IEEE and ACM. Ke Sun received the B.Sc. and M.Sc. degrees from Shandong Normal University, Jinan, China. He is currently Ph.D. Candidate in Software Engineering at Dalian University of Technology, Dalian, China. His research interests include deep learning, network representation learning, and knowledge graph. Shuo Yu (M’20) received the B.Sc. and M.Sc. degrees from Shenyang University of Technology, China, and the Ph.D. degree from Dalian University of Technology, Dalian, China. She is currently a Post-Doctoral Research Fellow with the School of Software, Dalian University of Technology. She has published over 30 papers in ACM/IEEE conferences, journals, and magazines. Her research interests in- clude network science, data science, and computa- tional social science. Abdul Aziz received the Bachelor’s degree in com- puter science from COMSATS Institute of Informa- tion Technology, Lahore Pakistan in 2013 and Mas- ter degree in Computer science from National Uni- versity of Computer & Emerging Sciences Karachi in 2018. He is currently a PhD student at the Alpha Lab, Dalian University of Technology, China. His research interests include big data, information retrieval, graph learning, and social computing. Liangtian Wan (M’15) received the B.S. degree and the Ph.D. degree from Harbin Engineering University, Harbin, China, in 2011 and 2015, re- spectively. From Oct. 2015 to Apr. 2017, he has been a Research Fellow at Nanyang Technological University, Singapore. He is currently an Associate Professor of School of Software, Dalian University of Technology, China. He is the author of over 70 papers. His current research interests include data science, big data and graph learning. Shirui Pan received a Ph.D. in computer science from the University of Technology Sydney (UTS), Australia. He is currently a lecturer with the Fac- ulty of Information Technology, Monash University, Australia. His research interests include data mining and machine learning. Dr Pan has published over 60 research papers in top-tier journals and conferences. Huan Liu (F’12) received the B.Eng. degree in computer science and electrical engineering from Shanghai Jiaotong University and the Ph.D. degree in computer science from the University of Southern California. He is currently a Professor of computer science and engineering at Arizona State Univer- sity. His research interests include data mining, machine learning, social computing, and artificial intelligence, investigating problems that arise in many real-world applications with high-dimensional data of disparate forms. His well-cited publications include books, book chapters, and encyclopedia entries and conference, and journal papers. He is a Fellow of IEEE, ACM, AAAI, and AAAS.