Deep matrix factorization for trust-aware recommendation in social networks
- Wan, Liangtian, Xia, Feng, Kong, Xiangjie, Hsu, Ching-Hsien, Huang, Runhe, Ma, Jianhua
- Authors: Wan, Liangtian , Xia, Feng , Kong, Xiangjie , Hsu, Ching-Hsien , Huang, Runhe , Ma, Jianhua
- Date: 2021
- Type: Text , Journal article
- Relation: IEEE Transactions on Network Science and Engineering Vol. 8, no. 1 (2021), p. 511-528
- Full Text:
- Reviewed:
- Description: Recent years have witnessed remarkable information overload in online social networks, and social network based approaches for recommender systems have been widely studied. The trust information in social networks among users is an important factor for improving recommendation performance. Many successful recommendation tasks are treated as the matrix factorization problems. However, the prediction performance of matrix factorization based methods largely depends on the matrixes initialization of users and items. To address this challenge, we develop a novel trust-aware approach based on deep learning to alleviate the initialization dependence. First, we propose two deep matrix factorization (DMF) techniques, i.e., linear DMF and non-linear DMF to extract features from the user-item rating matrix for improving the initialization accuracy. The trust relationship is integrated into the DMF model according to the preference similarity and the derivations of users on items. Second, we exploit deep marginalized Denoising Autoencoder (Deep-MDAE) to extract the latent representation in the hidden layer from the trust relationship matrix to approximate the user factor matrix factorized from the user-item rating matrix. The community regularization is integrated in the joint optimization function to take neighbours' effects into consideration. The results of DMF are applied to initialize the updating variables of Deep-MDAE in order to further improve the recommendation performance. Finally, we validate that the proposed approach outperforms state-of-the-art baselines for recommendation, especially for the cold-start users. © 2013 IEEE.
- Authors: Wan, Liangtian , Xia, Feng , Kong, Xiangjie , Hsu, Ching-Hsien , Huang, Runhe , Ma, Jianhua
- Date: 2021
- Type: Text , Journal article
- Relation: IEEE Transactions on Network Science and Engineering Vol. 8, no. 1 (2021), p. 511-528
- Full Text:
- Reviewed:
- Description: Recent years have witnessed remarkable information overload in online social networks, and social network based approaches for recommender systems have been widely studied. The trust information in social networks among users is an important factor for improving recommendation performance. Many successful recommendation tasks are treated as the matrix factorization problems. However, the prediction performance of matrix factorization based methods largely depends on the matrixes initialization of users and items. To address this challenge, we develop a novel trust-aware approach based on deep learning to alleviate the initialization dependence. First, we propose two deep matrix factorization (DMF) techniques, i.e., linear DMF and non-linear DMF to extract features from the user-item rating matrix for improving the initialization accuracy. The trust relationship is integrated into the DMF model according to the preference similarity and the derivations of users on items. Second, we exploit deep marginalized Denoising Autoencoder (Deep-MDAE) to extract the latent representation in the hidden layer from the trust relationship matrix to approximate the user factor matrix factorized from the user-item rating matrix. The community regularization is integrated in the joint optimization function to take neighbours' effects into consideration. The results of DMF are applied to initialize the updating variables of Deep-MDAE in order to further improve the recommendation performance. Finally, we validate that the proposed approach outperforms state-of-the-art baselines for recommendation, especially for the cold-start users. © 2013 IEEE.
Deep Reinforcement Learning for Vehicular Edge Computing: An Intelligent Offloading System
- Ning, Zhaolong, Dong, Peiran, Wang, Xiaojie, Rodrigues, Joel, Xia, Feng
- Authors: Ning, Zhaolong , Dong, Peiran , Wang, Xiaojie , Rodrigues, Joel , Xia, Feng
- Date: 2019
- Type: Text , Journal article
- Relation: ACM Transactions on Intelligent Systems and Technology Vol. 10, no. 6 (Dec 2019), p. 24
- Full Text:
- Reviewed:
- Description: The development of smart vehicles brings drivers and passengers a comfortable and safe environment. Various emerging applications are promising to enrich users' traveling experiences and daily life. However, how to execute computing-intensive applications on resource-constrained vehicles still faces huge challenges. In this article, we construct an intelligent offloading system for vehicular edge computing by leveraging deep reinforcement learning. First, both the communication and computation states are modelled by finite Markov chains. Moreover, the task scheduling and resource allocation strategy is formulated as a joint optimization problem to maximize users' Quality of Experience (QoE). Due to its complexity, the original problem is further divided into two sub-optimization problems. A two-sided matching scheme and a deep reinforcement learning approach are developed to schedule offloading requests and allocate network resources, respectively. Performance evaluations illustrate the effectiveness and superiority of our constructed system.
- Authors: Ning, Zhaolong , Dong, Peiran , Wang, Xiaojie , Rodrigues, Joel , Xia, Feng
- Date: 2019
- Type: Text , Journal article
- Relation: ACM Transactions on Intelligent Systems and Technology Vol. 10, no. 6 (Dec 2019), p. 24
- Full Text:
- Reviewed:
- Description: The development of smart vehicles brings drivers and passengers a comfortable and safe environment. Various emerging applications are promising to enrich users' traveling experiences and daily life. However, how to execute computing-intensive applications on resource-constrained vehicles still faces huge challenges. In this article, we construct an intelligent offloading system for vehicular edge computing by leveraging deep reinforcement learning. First, both the communication and computation states are modelled by finite Markov chains. Moreover, the task scheduling and resource allocation strategy is formulated as a joint optimization problem to maximize users' Quality of Experience (QoE). Due to its complexity, the original problem is further divided into two sub-optimization problems. A two-sided matching scheme and a deep reinforcement learning approach are developed to schedule offloading requests and allocate network resources, respectively. Performance evaluations illustrate the effectiveness and superiority of our constructed system.
Deep video anomaly detection : opportunities and challenges
- Ren, Jing, Xia, Feng, Liu, Yemeng, Lee, Ivan
- Authors: Ren, Jing , Xia, Feng , Liu, Yemeng , Lee, Ivan
- Date: 2021
- Type: Text , Conference paper
- Relation: 21st IEEE International Conference on Data Mining Workshops, ICDMW 2021, Virtual, Online 7-10 December 2021, IEEE International Conference on Data Mining Workshops, ICDMW Vol. 2021-December, p. 959-966
- Full Text:
- Reviewed:
- Description: Anomaly detection is a popular and vital task in various research contexts, which has been studied for several decades. To ensure the safety of people's lives and assets, video surveillance has been widely deployed in various public spaces, such as crossroads, elevators, hospitals, banks, and even in private homes. Deep learning has shown its capacity in a number of domains, ranging from acoustics, images, to natural language processing. However, it is non-trivial to devise intelligent video anomaly detection systems cause anomalies significantly differ from each other in different application scenarios. There are numerous advantages if such intelligent systems could be realised in our daily lives, such as saving human resources in a large degree, reducing financial burden on the government, and identifying the anomalous behaviours timely and accurately. Recently, many studies on extending deep learning models for solving anomaly detection problems have emerged, resulting in beneficial advances in deep video anomaly detection techniques. In this paper, we present a comprehensive review of deep learning-based methods to detect the video anomalies from a new perspective. Specifically, we summarise the opportunities and challenges of deep learning models on video anomaly detection tasks, respectively. We put forth several potential future research directions of intelligent video anomaly detection system in various application domains. Moreover, we summarise the characteristics and technical problems in current deep learning methods for video anomaly detection. © 2021 IEEE.
- Authors: Ren, Jing , Xia, Feng , Liu, Yemeng , Lee, Ivan
- Date: 2021
- Type: Text , Conference paper
- Relation: 21st IEEE International Conference on Data Mining Workshops, ICDMW 2021, Virtual, Online 7-10 December 2021, IEEE International Conference on Data Mining Workshops, ICDMW Vol. 2021-December, p. 959-966
- Full Text:
- Reviewed:
- Description: Anomaly detection is a popular and vital task in various research contexts, which has been studied for several decades. To ensure the safety of people's lives and assets, video surveillance has been widely deployed in various public spaces, such as crossroads, elevators, hospitals, banks, and even in private homes. Deep learning has shown its capacity in a number of domains, ranging from acoustics, images, to natural language processing. However, it is non-trivial to devise intelligent video anomaly detection systems cause anomalies significantly differ from each other in different application scenarios. There are numerous advantages if such intelligent systems could be realised in our daily lives, such as saving human resources in a large degree, reducing financial burden on the government, and identifying the anomalous behaviours timely and accurately. Recently, many studies on extending deep learning models for solving anomaly detection problems have emerged, resulting in beneficial advances in deep video anomaly detection techniques. In this paper, we present a comprehensive review of deep learning-based methods to detect the video anomalies from a new perspective. Specifically, we summarise the opportunities and challenges of deep learning models on video anomaly detection tasks, respectively. We put forth several potential future research directions of intelligent video anomaly detection system in various application domains. Moreover, we summarise the characteristics and technical problems in current deep learning methods for video anomaly detection. © 2021 IEEE.
DEFINE: friendship detection based on node enhancement
- Pan, Hanxiao, Guo, Teng, Bedru, Hayat, Qing, Qing, Zhang, Dongyu, Xia, Feng
- Authors: Pan, Hanxiao , Guo, Teng , Bedru, Hayat , Qing, Qing , Zhang, Dongyu , Xia, Feng
- Date: 2020
- Type: Text , Conference paper
- Relation: 31st Australasian Database Conference, ADC 2019 Vol. 12008 LNCS, p. 81-92
- Full Text:
- Reviewed:
- Description: Network representation learning (NRL) is a matter of importance to a variety of tasks such as link prediction. Learning low-dimensional vector representations for node enhancement based on nodes attributes and network structures can improve link prediction performance. Node attributes are important factors in forming networks, like psychological factors and appearance features affecting friendship networks. However, little to no work has detected friendship using the NRL technique, which combines students’ psychological features and perceived traits based on facial appearance. In this paper, we propose a framework named DEFINE (No enhancement based r e dship D tection) to detect students’ friend relationships, which combines with students’ psychological factors and facial perception information. To detect friend relationships accurately, DEFINE uses the NRL technique, which considers network structure and the additional attributes information for nodes. DEFINE transforms them into low-dimensional vector spaces while preserving the inherent properties of the friendship network. Experimental results on real-world friendship network datasets illustrate that DEFINE outperforms other state-of-art methods. © 2020, Springer Nature Switzerland AG.
- Description: E1
- Authors: Pan, Hanxiao , Guo, Teng , Bedru, Hayat , Qing, Qing , Zhang, Dongyu , Xia, Feng
- Date: 2020
- Type: Text , Conference paper
- Relation: 31st Australasian Database Conference, ADC 2019 Vol. 12008 LNCS, p. 81-92
- Full Text:
- Reviewed:
- Description: Network representation learning (NRL) is a matter of importance to a variety of tasks such as link prediction. Learning low-dimensional vector representations for node enhancement based on nodes attributes and network structures can improve link prediction performance. Node attributes are important factors in forming networks, like psychological factors and appearance features affecting friendship networks. However, little to no work has detected friendship using the NRL technique, which combines students’ psychological features and perceived traits based on facial appearance. In this paper, we propose a framework named DEFINE (No enhancement based r e dship D tection) to detect students’ friend relationships, which combines with students’ psychological factors and facial perception information. To detect friend relationships accurately, DEFINE uses the NRL technique, which considers network structure and the additional attributes information for nodes. DEFINE transforms them into low-dimensional vector spaces while preserving the inherent properties of the friendship network. Experimental results on real-world friendship network datasets illustrate that DEFINE outperforms other state-of-art methods. © 2020, Springer Nature Switzerland AG.
- Description: E1
Detecting outlier patterns with query-based artificially generated searching conditions
- Yu, Shuo, Xia, Feng, Sun, Yuchen, Tang, Tao, Yan, Xiaoran, Lee, Ivan
- Authors: Yu, Shuo , Xia, Feng , Sun, Yuchen , Tang, Tao , Yan, Xiaoran , Lee, Ivan
- Date: 2021
- Type: Text , Journal article
- Relation: IEEE Transactions on Computational Social Systems Vol. 8, no. 1 (2021), p. 134-147
- Full Text:
- Reviewed:
- Description: In the age of social computing, finding interesting network patterns or motifs is significant and critical for various areas, such as decision intelligence, intrusion detection, medical diagnosis, social network analysis, fake news identification, and national security. However, subgraph matching remains a computationally challenging problem, let alone identifying special motifs among them. This is especially the case in large heterogeneous real-world networks. In this article, we propose an efficient solution for discovering and ranking human behavior patterns based on network motifs by exploring a user's query in an intelligent way. Our method takes advantage of the semantics provided by a user's query, which in turn provides the mathematical constraint that is crucial for faster detection. We propose an approach to generate query conditions based on the user's query. In particular, we use meta paths between the nodes to define target patterns as well as their similarities, leading to efficient motif discovery and ranking at the same time. The proposed method is examined in a real-world academic network using different similarity measures between the nodes. The experiment result demonstrates that our method can identify interesting motifs and is robust to the choice of similarity measures. © 2014 IEEE.
- Authors: Yu, Shuo , Xia, Feng , Sun, Yuchen , Tang, Tao , Yan, Xiaoran , Lee, Ivan
- Date: 2021
- Type: Text , Journal article
- Relation: IEEE Transactions on Computational Social Systems Vol. 8, no. 1 (2021), p. 134-147
- Full Text:
- Reviewed:
- Description: In the age of social computing, finding interesting network patterns or motifs is significant and critical for various areas, such as decision intelligence, intrusion detection, medical diagnosis, social network analysis, fake news identification, and national security. However, subgraph matching remains a computationally challenging problem, let alone identifying special motifs among them. This is especially the case in large heterogeneous real-world networks. In this article, we propose an efficient solution for discovering and ranking human behavior patterns based on network motifs by exploring a user's query in an intelligent way. Our method takes advantage of the semantics provided by a user's query, which in turn provides the mathematical constraint that is crucial for faster detection. We propose an approach to generate query conditions based on the user's query. In particular, we use meta paths between the nodes to define target patterns as well as their similarities, leading to efficient motif discovery and ranking at the same time. The proposed method is examined in a real-world academic network using different similarity measures between the nodes. The experiment result demonstrates that our method can identify interesting motifs and is robust to the choice of similarity measures. © 2014 IEEE.
DINE : a framework for deep incomplete network embedding
- Hou, Ke, Liu, Jiaying, Peng, Yin, Xu, Bo, Lee, Ivan, Xia, Feng
- Authors: Hou, Ke , Liu, Jiaying , Peng, Yin , Xu, Bo , Lee, Ivan , Xia, Feng
- Date: 2019
- Type: Text , Conference paper
- Relation: 32nd Australasian Joint Conference on Artificial Intelligence, AI 2019 Vol. 11919 LNAI, p. 165-176
- Full Text:
- Reviewed:
- Description: Network representation learning (NRL) plays a vital role in a variety of tasks such as node classification and link prediction. It aims to learn low-dimensional vector representations for nodes based on network structures or node attributes. While embedding techniques on complete networks have been intensively studied, in real-world applications, it is still a challenging task to collect complete networks. To bridge the gap, in this paper, we propose a Deep Incomplete Network Embedding method, namely DINE. Specifically, we first complete the missing part including both nodes and edges in a partially observable network by using the expectation-maximization framework. To improve the embedding performance, we consider both network structures and node attributes to learn node representations. Empirically, we evaluate DINE over three networks on multi-label classification and link prediction tasks. The results demonstrate the superiority of our proposed approach compared against state-of-the-art baselines. © 2019, Springer Nature Switzerland AG.
- Description: E1
- Authors: Hou, Ke , Liu, Jiaying , Peng, Yin , Xu, Bo , Lee, Ivan , Xia, Feng
- Date: 2019
- Type: Text , Conference paper
- Relation: 32nd Australasian Joint Conference on Artificial Intelligence, AI 2019 Vol. 11919 LNAI, p. 165-176
- Full Text:
- Reviewed:
- Description: Network representation learning (NRL) plays a vital role in a variety of tasks such as node classification and link prediction. It aims to learn low-dimensional vector representations for nodes based on network structures or node attributes. While embedding techniques on complete networks have been intensively studied, in real-world applications, it is still a challenging task to collect complete networks. To bridge the gap, in this paper, we propose a Deep Incomplete Network Embedding method, namely DINE. Specifically, we first complete the missing part including both nodes and edges in a partially observable network by using the expectation-maximization framework. To improve the embedding performance, we consider both network structures and node attributes to learn node representations. Empirically, we evaluate DINE over three networks on multi-label classification and link prediction tasks. The results demonstrate the superiority of our proposed approach compared against state-of-the-art baselines. © 2019, Springer Nature Switzerland AG.
- Description: E1
Early-stage reciprocity in sustainable scientific collaboration
- Wang, Wei, Ren, Jing, Alrashoud, Mubarak, Xia, Feng, Mao, Mengyi, Tolba, Amr
- Authors: Wang, Wei , Ren, Jing , Alrashoud, Mubarak , Xia, Feng , Mao, Mengyi , Tolba, Amr
- Date: 2020
- Type: Text , Journal article
- Relation: Journal of Informetrics Vol. 14, no. 3 (2020), p.
- Full Text:
- Reviewed:
- Description: Scientific collaboration is of significant importance in tackling grand challenges and breeding innovations. Despite the increasing interest in investigating and promoting scientific collaborations, we know little about the collaboration sustainability as well as mechanisms behind it. In this paper, we set out to study the relationships between early-stage reciprocity and collaboration sustainability. By proposing and defining h-index reciprocity, we give a comprehensive statistical analysis on how reciprocity influences scientific collaboration sustainability, and find that scholars are not altruism and the key to sustainable collaboration is fairness. The unfair h-index reciprocity has an obvious negative impact on collaboration sustainability. The bigger the reciprocity difference, the less sustainable in collaboration. This work facilitates understanding sustainable collaborations and thus will benefit both individual scholar in optimizing collaboration strategies and the whole academic society in improving teamwork efficiency. © 2020 Elsevier Ltd.
- Description: The authors extend their appreciation to the International Scientific Partnership Program ISPP at King Saud University for funding this research work through ISPP-78. This work is partially supported by China Postdoctoral Science Foundation ( 2019M651115 ).
- Authors: Wang, Wei , Ren, Jing , Alrashoud, Mubarak , Xia, Feng , Mao, Mengyi , Tolba, Amr
- Date: 2020
- Type: Text , Journal article
- Relation: Journal of Informetrics Vol. 14, no. 3 (2020), p.
- Full Text:
- Reviewed:
- Description: Scientific collaboration is of significant importance in tackling grand challenges and breeding innovations. Despite the increasing interest in investigating and promoting scientific collaborations, we know little about the collaboration sustainability as well as mechanisms behind it. In this paper, we set out to study the relationships between early-stage reciprocity and collaboration sustainability. By proposing and defining h-index reciprocity, we give a comprehensive statistical analysis on how reciprocity influences scientific collaboration sustainability, and find that scholars are not altruism and the key to sustainable collaboration is fairness. The unfair h-index reciprocity has an obvious negative impact on collaboration sustainability. The bigger the reciprocity difference, the less sustainable in collaboration. This work facilitates understanding sustainable collaborations and thus will benefit both individual scholar in optimizing collaboration strategies and the whole academic society in improving teamwork efficiency. © 2020 Elsevier Ltd.
- Description: The authors extend their appreciation to the International Scientific Partnership Program ISPP at King Saud University for funding this research work through ISPP-78. This work is partially supported by China Postdoctoral Science Foundation ( 2019M651115 ).
Edge computing for Internet of Everything : a survey
- Kong, Xiangjie, Wu, Yuhan, Wang, Hui, Xia, Feng
- Authors: Kong, Xiangjie , Wu, Yuhan , Wang, Hui , Xia, Feng
- Date: 2022
- Type: Text , Journal article
- Relation: IEEE Internet of Things Journal Vol. 9, no. 23 (2022), p. 23472-23485
- Full Text:
- Reviewed:
- Description: In this era of the Internet of Everything (IoE), edge computing has emerged as the critical enabling technology to solve a series of issues caused by an increasing amount of interconnected devices and large-scale data transmission. However, the deficiencies of edge computing paradigm are gradually being magnified in the context of IoE, especially in terms of service migration, security and privacy preservation, and deployment issues of edge node. These issues can not be well addressed by conventional approaches. Thanks to the rapid development of upcoming technologies, such as artificial intelligence (AI), blockchain, and microservices, novel and more effective solutions have emerged and been applied to solve existing challenges. In addition, edge computing can be deeply integrated with technologies in other domains (e.g., AI, blockchain, 6G, and digital twin) through interdisciplinary intersection and practice, releasing the potential for mutual benefit. These promising integrations need to be further explored and researched. In addition, edge computing provides strong support in applications scenarios, such as remote working, new physical retail industries, and digital advertising, which has greatly changed the way we live, work, and study. In this article, we present an up-to-date survey of the edge computing research. In addition to introducing the definition, model, and characteristics of edge computing, we discuss a set of key issues in edge computing and novel solutions supported by emerging technologies in IoE era. Furthermore, we explore the potential and promising trends from the perspective of technology integration. Finally, new application scenarios and the final form of edge computing are discussed. © 2014 IEEE.
- Authors: Kong, Xiangjie , Wu, Yuhan , Wang, Hui , Xia, Feng
- Date: 2022
- Type: Text , Journal article
- Relation: IEEE Internet of Things Journal Vol. 9, no. 23 (2022), p. 23472-23485
- Full Text:
- Reviewed:
- Description: In this era of the Internet of Everything (IoE), edge computing has emerged as the critical enabling technology to solve a series of issues caused by an increasing amount of interconnected devices and large-scale data transmission. However, the deficiencies of edge computing paradigm are gradually being magnified in the context of IoE, especially in terms of service migration, security and privacy preservation, and deployment issues of edge node. These issues can not be well addressed by conventional approaches. Thanks to the rapid development of upcoming technologies, such as artificial intelligence (AI), blockchain, and microservices, novel and more effective solutions have emerged and been applied to solve existing challenges. In addition, edge computing can be deeply integrated with technologies in other domains (e.g., AI, blockchain, 6G, and digital twin) through interdisciplinary intersection and practice, releasing the potential for mutual benefit. These promising integrations need to be further explored and researched. In addition, edge computing provides strong support in applications scenarios, such as remote working, new physical retail industries, and digital advertising, which has greatly changed the way we live, work, and study. In this article, we present an up-to-date survey of the edge computing research. In addition to introducing the definition, model, and characteristics of edge computing, we discuss a set of key issues in edge computing and novel solutions supported by emerging technologies in IoE era. Furthermore, we explore the potential and promising trends from the perspective of technology integration. Finally, new application scenarios and the final form of edge computing are discussed. © 2014 IEEE.
Edge data based trailer inception probabilistic matrix factorization for context-aware movie recommendation
- Chen, Honglong, Li, Zhe, Wang, Zhu, Ni, Zhichen, Li, Junjian, Xu, Ge, Aziz, Abdul, Xia, Feng
- Authors: Chen, Honglong , Li, Zhe , Wang, Zhu , Ni, Zhichen , Li, Junjian , Xu, Ge , Aziz, Abdul , Xia, Feng
- Date: 2022
- Type: Text , Journal article
- Relation: World Wide Web Vol. 25, no. 5 (2022), p. 1863-1882
- Full Text:
- Reviewed:
- Description: The rapid growth of edge data generated by mobile devices and applications deployed at the edge of the network has exacerbated the problem of information overload. As an effective way to alleviate information overload, recommender system can improve the quality of various services by adding application data generated by users on edge devices, such as visual and textual information, on the basis of sparse rating data. The visual information in the movie trailer is a significant part of the movie recommender system. However, due to the complexity of visual information extraction, data sparsity cannot be remarkably alleviated by merely using the rough visual features to improve the rating prediction accuracy. Fortunately, the convolutional neural network can be used to extract the visual features precisely. Therefore, the end-to-end neural image caption (NIC) model can be utilized to obtain the textual information describing the visual features of movie trailers. This paper proposes a trailer inception probabilistic matrix factorization model called Ti-PMF, which combines NIC, recurrent convolutional neural network, and probabilistic matrix factorization models as the rating prediction model. We implement the proposed Ti-PMF model with extensive experiments on three real-world datasets to validate its effectiveness. The experimental results illustrate that the proposed Ti-PMF outperforms the existing ones. © 2021, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.
- Authors: Chen, Honglong , Li, Zhe , Wang, Zhu , Ni, Zhichen , Li, Junjian , Xu, Ge , Aziz, Abdul , Xia, Feng
- Date: 2022
- Type: Text , Journal article
- Relation: World Wide Web Vol. 25, no. 5 (2022), p. 1863-1882
- Full Text:
- Reviewed:
- Description: The rapid growth of edge data generated by mobile devices and applications deployed at the edge of the network has exacerbated the problem of information overload. As an effective way to alleviate information overload, recommender system can improve the quality of various services by adding application data generated by users on edge devices, such as visual and textual information, on the basis of sparse rating data. The visual information in the movie trailer is a significant part of the movie recommender system. However, due to the complexity of visual information extraction, data sparsity cannot be remarkably alleviated by merely using the rough visual features to improve the rating prediction accuracy. Fortunately, the convolutional neural network can be used to extract the visual features precisely. Therefore, the end-to-end neural image caption (NIC) model can be utilized to obtain the textual information describing the visual features of movie trailers. This paper proposes a trailer inception probabilistic matrix factorization model called Ti-PMF, which combines NIC, recurrent convolutional neural network, and probabilistic matrix factorization models as the rating prediction model. We implement the proposed Ti-PMF model with extensive experiments on three real-world datasets to validate its effectiveness. The experimental results illustrate that the proposed Ti-PMF outperforms the existing ones. © 2021, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.
Educational anomaly analytics : features, methods, and challenges
- Guo, Teng, Bai, Xiaomei, Tian, Xue, Firmin, Sally, Xia, Feng
- Authors: Guo, Teng , Bai, Xiaomei , Tian, Xue , Firmin, Sally , Xia, Feng
- Date: 2022
- Type: Text , Journal article , Review
- Relation: Frontiers in Big Data Vol. 4, no. (2022), p.
- Full Text:
- Reviewed:
- Description: Anomalies in education affect the personal careers of students and universities' retention rates. Understanding the laws behind educational anomalies promotes the development of individual students and improves the overall quality of education. However, the inaccessibility of educational data hinders the development of the field. Previous research in this field used questionnaires, which are time- and cost-consuming and hardly applicable to large-scale student cohorts. With the popularity of educational management systems and the rise of online education during the prevalence of COVID-19, a large amount of educational data is available online and offline, providing an unprecedented opportunity to explore educational anomalies from a data-driven perspective. As an emerging field, educational anomaly analytics rapidly attracts scholars from a variety of fields, including education, psychology, sociology, and computer science. This paper intends to provide a comprehensive review of data-driven analytics of educational anomalies from a methodological standpoint. We focus on the following five types of research that received the most attention: course failure prediction, dropout prediction, mental health problems detection, prediction of difficulty in graduation, and prediction of difficulty in employment. Then, we discuss the challenges of current related research. This study aims to provide references for educational policymaking while promoting the development of educational anomaly analytics as a growing field. Copyright © 2022 Guo, Bai, Tian, Firmin and Xia.
- Authors: Guo, Teng , Bai, Xiaomei , Tian, Xue , Firmin, Sally , Xia, Feng
- Date: 2022
- Type: Text , Journal article , Review
- Relation: Frontiers in Big Data Vol. 4, no. (2022), p.
- Full Text:
- Reviewed:
- Description: Anomalies in education affect the personal careers of students and universities' retention rates. Understanding the laws behind educational anomalies promotes the development of individual students and improves the overall quality of education. However, the inaccessibility of educational data hinders the development of the field. Previous research in this field used questionnaires, which are time- and cost-consuming and hardly applicable to large-scale student cohorts. With the popularity of educational management systems and the rise of online education during the prevalence of COVID-19, a large amount of educational data is available online and offline, providing an unprecedented opportunity to explore educational anomalies from a data-driven perspective. As an emerging field, educational anomaly analytics rapidly attracts scholars from a variety of fields, including education, psychology, sociology, and computer science. This paper intends to provide a comprehensive review of data-driven analytics of educational anomalies from a methodological standpoint. We focus on the following five types of research that received the most attention: course failure prediction, dropout prediction, mental health problems detection, prediction of difficulty in graduation, and prediction of difficulty in employment. Then, we discuss the challenges of current related research. This study aims to provide references for educational policymaking while promoting the development of educational anomaly analytics as a growing field. Copyright © 2022 Guo, Bai, Tian, Firmin and Xia.
Educational big data : predictions, applications and challenges
- Bai, Xiaomei, Zhang, Fuli, Li, Jinzhou, Guo, Teng, Xia, Feng
- Authors: Bai, Xiaomei , Zhang, Fuli , Li, Jinzhou , Guo, Teng , Xia, Feng
- Date: 2021
- Type: Text , Journal article , Review
- Relation: Big Data Research Vol. 26, no. (2021), p.
- Full Text:
- Reviewed:
- Description: Educational big data is becoming a strategic educational asset, exceptionally significant in advancing educational reform. The term educational big data stems from the rapidly growing educational data development, including students' inherent attributes, learning behavior, and psychological state. Educational big data has many applications that can be used for educational administration, teaching innovation, and research management. The representative examples of such applications are student academic performance prediction, employment recommendation, and financial support for low-income students. Different empirical studies have shown that it is possible to predict student performance in the courses during the next term. Predictive research for the higher education stage has become an attractive area of study since it allowed us to predict student behavior. In this survey, we will review predictive research, its applications, and its challenges. We first introduce the significance and background of educational big data. Second, we review the students' academic performance prediction research, such as factors influencing students' academic performance, predicting models, evaluating indices. Third, we introduce the applications of educational big data such as prediction, recommendation, and evaluation. Finally, we investigate challenging research issues in this area. This discussion aims to provide a comprehensive overview of educational big data. © 2021 Elsevier Inc. **Please note that there are multiple authors for this article therefore only the name of the first 5 including Federation University Australia affiliate “Feng Xia” is provided in this record**
- Authors: Bai, Xiaomei , Zhang, Fuli , Li, Jinzhou , Guo, Teng , Xia, Feng
- Date: 2021
- Type: Text , Journal article , Review
- Relation: Big Data Research Vol. 26, no. (2021), p.
- Full Text:
- Reviewed:
- Description: Educational big data is becoming a strategic educational asset, exceptionally significant in advancing educational reform. The term educational big data stems from the rapidly growing educational data development, including students' inherent attributes, learning behavior, and psychological state. Educational big data has many applications that can be used for educational administration, teaching innovation, and research management. The representative examples of such applications are student academic performance prediction, employment recommendation, and financial support for low-income students. Different empirical studies have shown that it is possible to predict student performance in the courses during the next term. Predictive research for the higher education stage has become an attractive area of study since it allowed us to predict student behavior. In this survey, we will review predictive research, its applications, and its challenges. We first introduce the significance and background of educational big data. Second, we review the students' academic performance prediction research, such as factors influencing students' academic performance, predicting models, evaluating indices. Third, we introduce the applications of educational big data such as prediction, recommendation, and evaluation. Finally, we investigate challenging research issues in this area. This discussion aims to provide a comprehensive overview of educational big data. © 2021 Elsevier Inc. **Please note that there are multiple authors for this article therefore only the name of the first 5 including Federation University Australia affiliate “Feng Xia” is provided in this record**
Efficient anomaly recognition using surveillance videos
- Saleem, Gulshan, Bajwa, Usama, Raza, Rana, Alqahtani, Fayez, Tolba, Amr, Xia, Feng
- Authors: Saleem, Gulshan , Bajwa, Usama , Raza, Rana , Alqahtani, Fayez , Tolba, Amr , Xia, Feng
- Date: 2022
- Type: Text , Journal article
- Relation: PeerJ Computer Science Vol. 8, no. (2022), p.
- Full Text:
- Reviewed:
- Description: Smart surveillance is a difficult task that is gaining popularity due to its direct link to human safety. Today, many indoor and outdoor surveillance systems are in use at public places and smart cities. Because these systems are expensive to deploy, these are out of reach for the vast majority of the public and private sectors. Due to the lack of a precise definition of an anomaly, automated surveillance is a challenging task, especially when large amounts of data, such as 24/7 CCTV footage, must be processed. When implementing such systems in real-time environments, the high computational resource requirements for automated surveillance becomes a major bottleneck. Another challenge is to recognize anomalies accurately as achieving high accuracy while reducing computational cost is more challenging. To address these challenge, this research is based on the developing a system that is both efficient and cost effective. Although 3D convolutional neural networks have proven to be accurate, they are prohibitively expensive for practical use, particularly in real-time surveillance. In this article, we present two contributions: a resource-efficient framework for anomaly recognition problems and two-class and multi-class anomaly recognition on spatially augmented surveillance videos. This research aims to address the problem of computation overhead while maintaining recognition accuracy. The proposed Temporal based Anomaly Recognizer (TAR) framework combines a partial shift strategy with a 2D convolutional architecture-based model, namely MobileNetV2. Extensive experiments were carried out to evaluate the model's performance on the UCF Crime dataset, with MobileNetV2 as the baseline architecture; it achieved an accuracy of 88% which is 2.47% increased performance than available state-of-the-art. The proposed framework achieves 52.7% accuracy for multiclass anomaly recognition on the UCF Crime2Local dataset. The proposed model has been tested in real-time camera stream settings and can handle six streams simultaneously without the need for additional resources. © Copyright 2022 Saleem et al.
- Authors: Saleem, Gulshan , Bajwa, Usama , Raza, Rana , Alqahtani, Fayez , Tolba, Amr , Xia, Feng
- Date: 2022
- Type: Text , Journal article
- Relation: PeerJ Computer Science Vol. 8, no. (2022), p.
- Full Text:
- Reviewed:
- Description: Smart surveillance is a difficult task that is gaining popularity due to its direct link to human safety. Today, many indoor and outdoor surveillance systems are in use at public places and smart cities. Because these systems are expensive to deploy, these are out of reach for the vast majority of the public and private sectors. Due to the lack of a precise definition of an anomaly, automated surveillance is a challenging task, especially when large amounts of data, such as 24/7 CCTV footage, must be processed. When implementing such systems in real-time environments, the high computational resource requirements for automated surveillance becomes a major bottleneck. Another challenge is to recognize anomalies accurately as achieving high accuracy while reducing computational cost is more challenging. To address these challenge, this research is based on the developing a system that is both efficient and cost effective. Although 3D convolutional neural networks have proven to be accurate, they are prohibitively expensive for practical use, particularly in real-time surveillance. In this article, we present two contributions: a resource-efficient framework for anomaly recognition problems and two-class and multi-class anomaly recognition on spatially augmented surveillance videos. This research aims to address the problem of computation overhead while maintaining recognition accuracy. The proposed Temporal based Anomaly Recognizer (TAR) framework combines a partial shift strategy with a 2D convolutional architecture-based model, namely MobileNetV2. Extensive experiments were carried out to evaluate the model's performance on the UCF Crime dataset, with MobileNetV2 as the baseline architecture; it achieved an accuracy of 88% which is 2.47% increased performance than available state-of-the-art. The proposed framework achieves 52.7% accuracy for multiclass anomaly recognition on the UCF Crime2Local dataset. The proposed model has been tested in real-time camera stream settings and can handle six streams simultaneously without the need for additional resources. © Copyright 2022 Saleem et al.
Emergency warning messages dissemination in vehicular social networks: A trust based scheme
- Ullah, Noor, Kong, Xiangjie, Ning, Zhaolong, Tolba, Amr, Alrashoud, Mubarak, Xia, Feng
- Authors: Ullah, Noor , Kong, Xiangjie , Ning, Zhaolong , Tolba, Amr , Alrashoud, Mubarak , Xia, Feng
- Date: 2020
- Type: Text , Journal article
- Relation: Vehicular Communications Vol. 22 (2020)
- Full Text:
- Reviewed:
- Description: To ensure users' safety on the road, a plethora of dissemination schemes for Emergency Warning Messages (EWMs) have been proposed in vehicular networks. However, the issue of false alarms triggered by malicious users still poses serious challenges, such as disruption of vehicular traffic especially on highways leading to precarious effects. This paper proposes a novel Trust based Dissemination Scheme (TDS) for EWMs in Vehicular Social Networks (VSNs) to solve the aforementioned issue. To ensure the authenticity of EWMs, we exploit the user-post credibility network for identifying true and false alarms. Moreover, we develop a reputation mechanism by calculating a trust-score for each node based on its social-utility, behavior, and contribution in the network. We utilize the hybrid architecture of VSNs by employing social-groups based dissemination in Vehicle-to-Infrastructure (V2I) mode, whereas nodes' friendship-network in Vehicle-to-Vehicle (V2V) mode. We analyze the proposed scheme for accuracy by extensive simulations under varying malicious nodes ratio in the network. Furthermore, we compare the efficiency of TDS with state-of-the-art dissemination schemes in VSNs for delivery ratio, transmission delay, number of transmissions, and hop-count. The experimental results validate the significant efficacy of TDS in accuracy and aforementioned network parameters. © 2019 Elsevier Inc.
- Authors: Ullah, Noor , Kong, Xiangjie , Ning, Zhaolong , Tolba, Amr , Alrashoud, Mubarak , Xia, Feng
- Date: 2020
- Type: Text , Journal article
- Relation: Vehicular Communications Vol. 22 (2020)
- Full Text:
- Reviewed:
- Description: To ensure users' safety on the road, a plethora of dissemination schemes for Emergency Warning Messages (EWMs) have been proposed in vehicular networks. However, the issue of false alarms triggered by malicious users still poses serious challenges, such as disruption of vehicular traffic especially on highways leading to precarious effects. This paper proposes a novel Trust based Dissemination Scheme (TDS) for EWMs in Vehicular Social Networks (VSNs) to solve the aforementioned issue. To ensure the authenticity of EWMs, we exploit the user-post credibility network for identifying true and false alarms. Moreover, we develop a reputation mechanism by calculating a trust-score for each node based on its social-utility, behavior, and contribution in the network. We utilize the hybrid architecture of VSNs by employing social-groups based dissemination in Vehicle-to-Infrastructure (V2I) mode, whereas nodes' friendship-network in Vehicle-to-Vehicle (V2V) mode. We analyze the proposed scheme for accuracy by extensive simulations under varying malicious nodes ratio in the network. Furthermore, we compare the efficiency of TDS with state-of-the-art dissemination schemes in VSNs for delivery ratio, transmission delay, number of transmissions, and hop-count. The experimental results validate the significant efficacy of TDS in accuracy and aforementioned network parameters. © 2019 Elsevier Inc.
Exploring human mobility for multi-pattern passenger prediction : a graph learning framework
- Kong, Xiangjiea, Wang, Kailai, Hou, Mingliang, Xia, Feng, Karmakar, Gour, Li, Jianxin
- Authors: Kong, Xiangjiea , Wang, Kailai , Hou, Mingliang , Xia, Feng , Karmakar, Gour , Li, Jianxin
- Date: 2022
- Type: Text , Journal article
- Relation: IEEE Transactions on Intelligent Transportation Systems Vol. 23, no. 9 (2022), p. 16148-16160
- Full Text:
- Reviewed:
- Description: Traffic flow prediction is an integral part of an intelligent transportation system and thus fundamental for various traffic-related applications. Buses are an indispensable way of moving for urban residents with fixed routes and schedules, which leads to latent travel regularity. However, human mobility patterns, specifically the complex relationships between bus passengers, are deeply hidden in this fixed mobility mode. Although many models exist to predict traffic flow, human mobility patterns have not been well explored in this regard. To address this research gap and learn human mobility knowledge from this fixed travel behaviors, we propose a multi-pattern passenger flow prediction framework, MPGCN, based on Graph Convolutional Network (GCN). Firstly, we construct a novel sharing-stop network to model relationships between passengers based on bus record data. Then, we employ GCN to extract features from the graph by learning useful topology information and introduce a deep clustering method to recognize mobility patterns hidden in bus passengers. Furthermore, to fully utilize spatio-temporal information, we propose GCN2Flow to predict passenger flow based on various mobility patterns. To the best of our knowledge, this paper is the first work to adopt a multi-pattern approach to predict the bus passenger flow by taking advantage of graph learning. We design a case study for optimizing routes. Extensive experiments upon a real-world bus dataset demonstrate that MPGCN has potential efficacy in passenger flow prediction and route optimization. © 2000-2011 IEEE.
- Authors: Kong, Xiangjiea , Wang, Kailai , Hou, Mingliang , Xia, Feng , Karmakar, Gour , Li, Jianxin
- Date: 2022
- Type: Text , Journal article
- Relation: IEEE Transactions on Intelligent Transportation Systems Vol. 23, no. 9 (2022), p. 16148-16160
- Full Text:
- Reviewed:
- Description: Traffic flow prediction is an integral part of an intelligent transportation system and thus fundamental for various traffic-related applications. Buses are an indispensable way of moving for urban residents with fixed routes and schedules, which leads to latent travel regularity. However, human mobility patterns, specifically the complex relationships between bus passengers, are deeply hidden in this fixed mobility mode. Although many models exist to predict traffic flow, human mobility patterns have not been well explored in this regard. To address this research gap and learn human mobility knowledge from this fixed travel behaviors, we propose a multi-pattern passenger flow prediction framework, MPGCN, based on Graph Convolutional Network (GCN). Firstly, we construct a novel sharing-stop network to model relationships between passengers based on bus record data. Then, we employ GCN to extract features from the graph by learning useful topology information and introduce a deep clustering method to recognize mobility patterns hidden in bus passengers. Furthermore, to fully utilize spatio-temporal information, we propose GCN2Flow to predict passenger flow based on various mobility patterns. To the best of our knowledge, this paper is the first work to adopt a multi-pattern approach to predict the bus passenger flow by taking advantage of graph learning. We design a case study for optimizing routes. Extensive experiments upon a real-world bus dataset demonstrate that MPGCN has potential efficacy in passenger flow prediction and route optimization. © 2000-2011 IEEE.
Expressing metaphorically, writing creatively: Metaphor identification for creativity assessment
- Zhang, Dongyu, Zhang, Minghao, Peng, Ciyuan, Xia, Feng
- Authors: Zhang, Dongyu , Zhang, Minghao , Peng, Ciyuan , Xia, Feng
- Date: 2022
- Type: Text , Conference proceedings
- Relation: WWW '22: Companion Proceedings of the Web Conference , Virtual event , April 2022 p. 1198-
- Full Text:
- Reviewed:
- Description: Metaphor, which can implicitly express profound meanings and emotions, is a unique writing technique frequently used in human language. In writing, meaningful metaphorical expressions can enhance the literariness and creativity of texts. Therefore, the usage of metaphor is a significant impact factor when assessing the creativity and literariness of writing. However, little to no automatic writing assessment system considers metaphorical expressions when giving the score of creativity. For improving the accuracy of automatic writing assessment, this paper proposes a novel creativity assessment model that imports a token-level metaphor identification method to extract metaphors as the indicators for creativity scoring. The experimental results show that our model can accurately assess the creativity of different texts with precise metaphor identification. To the best of our knowledge, we are the first to apply automatic metaphor identification to assess writing creativity. Moreover, identifying features (e.g., metaphors) that influence writing creativity using computational approaches can offer fair and reliable assessment methods for educational settings.
- Authors: Zhang, Dongyu , Zhang, Minghao , Peng, Ciyuan , Xia, Feng
- Date: 2022
- Type: Text , Conference proceedings
- Relation: WWW '22: Companion Proceedings of the Web Conference , Virtual event , April 2022 p. 1198-
- Full Text:
- Reviewed:
- Description: Metaphor, which can implicitly express profound meanings and emotions, is a unique writing technique frequently used in human language. In writing, meaningful metaphorical expressions can enhance the literariness and creativity of texts. Therefore, the usage of metaphor is a significant impact factor when assessing the creativity and literariness of writing. However, little to no automatic writing assessment system considers metaphorical expressions when giving the score of creativity. For improving the accuracy of automatic writing assessment, this paper proposes a novel creativity assessment model that imports a token-level metaphor identification method to extract metaphors as the indicators for creativity scoring. The experimental results show that our model can accurately assess the creativity of different texts with precise metaphor identification. To the best of our knowledge, we are the first to apply automatic metaphor identification to assess writing creativity. Moreover, identifying features (e.g., metaphors) that influence writing creativity using computational approaches can offer fair and reliable assessment methods for educational settings.
Familiarity-based collaborative team recognition in academic social networks
- Yu, Shuo, Xia, Feng, Zhang, Chen, Wei, Haoran, Keogh, Kathleen, Chen, Honglong
- Authors: Yu, Shuo , Xia, Feng , Zhang, Chen , Wei, Haoran , Keogh, Kathleen , Chen, Honglong
- Date: 2022
- Type: Text , Journal article
- Relation: IEEE Transactions on Computational Social Systems Vol. 9, no. 5 (2022), p. 1432-1445
- Full Text:
- Reviewed:
- Description: Collaborative teamwork is key to major scientific discoveries. However, the prevalence of collaboration among researchers makes team recognition increasingly challenging. Previous studies have demonstrated that people are more likely to collaborate with individuals they are familiar with. In this work, we employ the definition of familiarity and then propose faMiliarity-based cOllaborative Team recOgnition (MOTO) algorithm to recognize collaborative teams. MOTO calculates the shortest distance matrix within the global collaboration network and the local density of each node. Central team members are initially recognized based on local density. Then, MOTO recognizes the remaining team members by using the familiarity metric and shortest distance matrix. Extensive experiments have been conducted upon a large-scale dataset. The experimental results show that compared with baseline methods, MOTO can recognize the largest number of teams. The teams recognized by the MOTO possess more cohesive team structures and lower team communication costs compared with other methods. MOTO utilizes familiarity in team recognition to identify cohesive academic teams. The recognized teams are in line with real-world collaborative teamwork patterns. Based on team recognition using MOTO, the research team structure and performance are further analyzed for given time periods. The number of teams that consist of members from different institutions increases gradually. Such teams are found to perform better in comparison with those whose members are from the same institution. © 2014 IEEE.
- Authors: Yu, Shuo , Xia, Feng , Zhang, Chen , Wei, Haoran , Keogh, Kathleen , Chen, Honglong
- Date: 2022
- Type: Text , Journal article
- Relation: IEEE Transactions on Computational Social Systems Vol. 9, no. 5 (2022), p. 1432-1445
- Full Text:
- Reviewed:
- Description: Collaborative teamwork is key to major scientific discoveries. However, the prevalence of collaboration among researchers makes team recognition increasingly challenging. Previous studies have demonstrated that people are more likely to collaborate with individuals they are familiar with. In this work, we employ the definition of familiarity and then propose faMiliarity-based cOllaborative Team recOgnition (MOTO) algorithm to recognize collaborative teams. MOTO calculates the shortest distance matrix within the global collaboration network and the local density of each node. Central team members are initially recognized based on local density. Then, MOTO recognizes the remaining team members by using the familiarity metric and shortest distance matrix. Extensive experiments have been conducted upon a large-scale dataset. The experimental results show that compared with baseline methods, MOTO can recognize the largest number of teams. The teams recognized by the MOTO possess more cohesive team structures and lower team communication costs compared with other methods. MOTO utilizes familiarity in team recognition to identify cohesive academic teams. The recognized teams are in line with real-world collaborative teamwork patterns. Based on team recognition using MOTO, the research team structure and performance are further analyzed for given time periods. The number of teams that consist of members from different institutions increases gradually. Such teams are found to perform better in comparison with those whose members are from the same institution. © 2014 IEEE.
Graduate employment prediction with bias
- Guo, Teng, Xia, Feng, Zhen, Shihao, Bai, Xiaomei, Zhang, Dongyu
- Authors: Guo, Teng , Xia, Feng , Zhen, Shihao , Bai, Xiaomei , Zhang, Dongyu
- Date: 2020
- Type: Text , Conference paper
- Relation: AAAI 2020 - 34th AAAI Conference on Artificial Intelligence p. 670-677
- Full Text:
- Reviewed:
- Description: The failure of landing a job for college students could cause serious social consequences such as drunkenness and suicide. In addition to academic performance, unconscious biases can become one key obstacle for hunting jobs for graduating students. Thus, it is necessary to understand these unconscious biases so that we can help these students at an early stage with more personalized intervention. In this paper, we develop a framework, i.e., MAYA (Multi-mAjor emploYment stAtus) to predict students’ employment status while considering biases. The framework consists of four major components. Firstly, we solve the heterogeneity of student courses by embedding academic performance into a unified space. Then, we apply a generative adversarial network (GAN) to overcome the class imbalance problem. Thirdly, we adopt Long Short-Term Memory (LSTM) with a novel dropout mechanism to comprehensively capture sequential information among semesters. Finally, we design a bias-based regularization to capture the job market biases. We conduct extensive experiments on a large-scale educational dataset and the results demonstrate the effectiveness of our prediction framework. Copyright © 2020, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. **Please note that there are multiple authors for this article therefore only the name of the first 5 including Federation University Australia affiliate “Feng Xia” is provided in this record**
- Authors: Guo, Teng , Xia, Feng , Zhen, Shihao , Bai, Xiaomei , Zhang, Dongyu
- Date: 2020
- Type: Text , Conference paper
- Relation: AAAI 2020 - 34th AAAI Conference on Artificial Intelligence p. 670-677
- Full Text:
- Reviewed:
- Description: The failure of landing a job for college students could cause serious social consequences such as drunkenness and suicide. In addition to academic performance, unconscious biases can become one key obstacle for hunting jobs for graduating students. Thus, it is necessary to understand these unconscious biases so that we can help these students at an early stage with more personalized intervention. In this paper, we develop a framework, i.e., MAYA (Multi-mAjor emploYment stAtus) to predict students’ employment status while considering biases. The framework consists of four major components. Firstly, we solve the heterogeneity of student courses by embedding academic performance into a unified space. Then, we apply a generative adversarial network (GAN) to overcome the class imbalance problem. Thirdly, we adopt Long Short-Term Memory (LSTM) with a novel dropout mechanism to comprehensively capture sequential information among semesters. Finally, we design a bias-based regularization to capture the job market biases. We conduct extensive experiments on a large-scale educational dataset and the results demonstrate the effectiveness of our prediction framework. Copyright © 2020, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. **Please note that there are multiple authors for this article therefore only the name of the first 5 including Federation University Australia affiliate “Feng Xia” is provided in this record**
Graph augmentation learning
- Yu, Shuo, Huang, Huafei, Dao, Minh, Xia, Feng
- Authors: Yu, Shuo , Huang, Huafei , Dao, Minh , Xia, Feng
- Date: 2022
- Type: Text , Conference paper
- Relation: 31st ACM Web Conference, WWW 2022, Virtual, online, 25 April 2022, WWW 2022 - Companion Proceedings of the Web Conference 2022 p. 1063-1072
- Full Text:
- Reviewed:
- Description: Graph Augmentation Learning (GAL) provides outstanding solutions for graph learning in handling incomplete data, noise data, etc. Numerous GAL methods have been proposed for graph-based applications such as social network analysis and traffic flow forecasting. However, the underlying reasons for the effectiveness of these GAL methods are still unclear. As a consequence, how to choose optimal graph augmentation strategy for a certain application scenario is still in black box. There is a lack of systematic, comprehensive, and experimentally validated guideline of GAL for scholars. Therefore, in this survey, we in-depth review GAL techniques from macro (graph), meso (subgraph), and micro (node/edge) levels. We further detailedly illustrate how GAL enhance the data quality and the model performance. The aggregation mechanism of augmentation strategies and graph learning models are also discussed by different application scenarios, i.e., data-specific, model-specific, and hybrid scenarios. To better show the outperformance of GAL, we experimentally validate the effectiveness and adaptability of different GAL strategies in different downstream tasks. Finally, we share our insights on several open issues of GAL, including heterogeneity, spatio-temporal dynamics, scalability, and generalization. © 2022 ACM.
- Authors: Yu, Shuo , Huang, Huafei , Dao, Minh , Xia, Feng
- Date: 2022
- Type: Text , Conference paper
- Relation: 31st ACM Web Conference, WWW 2022, Virtual, online, 25 April 2022, WWW 2022 - Companion Proceedings of the Web Conference 2022 p. 1063-1072
- Full Text:
- Reviewed:
- Description: Graph Augmentation Learning (GAL) provides outstanding solutions for graph learning in handling incomplete data, noise data, etc. Numerous GAL methods have been proposed for graph-based applications such as social network analysis and traffic flow forecasting. However, the underlying reasons for the effectiveness of these GAL methods are still unclear. As a consequence, how to choose optimal graph augmentation strategy for a certain application scenario is still in black box. There is a lack of systematic, comprehensive, and experimentally validated guideline of GAL for scholars. Therefore, in this survey, we in-depth review GAL techniques from macro (graph), meso (subgraph), and micro (node/edge) levels. We further detailedly illustrate how GAL enhance the data quality and the model performance. The aggregation mechanism of augmentation strategies and graph learning models are also discussed by different application scenarios, i.e., data-specific, model-specific, and hybrid scenarios. To better show the outperformance of GAL, we experimentally validate the effectiveness and adaptability of different GAL strategies in different downstream tasks. Finally, we share our insights on several open issues of GAL, including heterogeneity, spatio-temporal dynamics, scalability, and generalization. © 2022 ACM.
Graph Force Learning
- Sun, Ke, Liu, Jiaying, Yu, Shuo, Xu, Bo, Xia, Feng
- Authors: Sun, Ke , Liu, Jiaying , Yu, Shuo , Xu, Bo , Xia, Feng
- Date: 2020
- Type: Text , Conference paper
- Relation: 8th IEEE International Conference on Big Data, Big Data 2020 p. 2987-2994
- Full Text:
- Reviewed:
- Description: Features representation leverages the great power in network analysis tasks. However, most features are discrete which poses tremendous challenges to effective use. Recently, increasing attention has been paid on network feature learning, which could map discrete features to continued space. Unfortunately, current studies fail to fully preserve the structural information in the feature space due to random negative sampling strategy during training. To tackle this problem, we study the problem of feature learning and novelty propose a force-based graph learning model named GForce inspired by the spring-electrical model. GForce assumes that nodes are in attractive forces and repulsive forces, thus leading to the same representation with the original structural information in feature learning. Comprehensive experiments on three benchmark datasets demonstrate the effectiveness of the proposed framework. Furthermore, GForce opens up opportunities to use physics models to model node interaction for graph learning. © 2020 IEEE.
- Authors: Sun, Ke , Liu, Jiaying , Yu, Shuo , Xu, Bo , Xia, Feng
- Date: 2020
- Type: Text , Conference paper
- Relation: 8th IEEE International Conference on Big Data, Big Data 2020 p. 2987-2994
- Full Text:
- Reviewed:
- Description: Features representation leverages the great power in network analysis tasks. However, most features are discrete which poses tremendous challenges to effective use. Recently, increasing attention has been paid on network feature learning, which could map discrete features to continued space. Unfortunately, current studies fail to fully preserve the structural information in the feature space due to random negative sampling strategy during training. To tackle this problem, we study the problem of feature learning and novelty propose a force-based graph learning model named GForce inspired by the spring-electrical model. GForce assumes that nodes are in attractive forces and repulsive forces, thus leading to the same representation with the original structural information in feature learning. Comprehensive experiments on three benchmark datasets demonstrate the effectiveness of the proposed framework. Furthermore, GForce opens up opportunities to use physics models to model node interaction for graph learning. © 2020 IEEE.
Graph learning : a survey
- Xia, Feng, Sun, Ke, Yu, Shuo, Aziz, Abdul, Wan, Liangtian, Pan, Shirui, Liu, Huan
- Authors: Xia, Feng , Sun, Ke , Yu, Shuo , Aziz, Abdul , Wan, Liangtian , Pan, Shirui , Liu, Huan
- Date: 2021
- Type: Text , Journal article , Review
- Relation: IEEE Transactions on Artificial Intelligence Vol. 2, no. 2 (2021), p. 109-127
- Full Text:
- Reviewed:
- Description: Graphs are widely used as a popular representation of the network structure of connected data. Graph data can be found in a broad spectrum of application domains such as social systems, ecosystems, biological networks, knowledge graphs, and information systems. With the continuous penetration of artificial intelligence technologies, graph learning (i.e., machine learning on graphs) is gaining attention from both researchers and practitioners. Graph learning proves effective for many tasks, such as classification, link prediction, and matching. Generally, graph learning methods extract relevant features of graphs by taking advantage of machine learning algorithms. In this survey, we present a comprehensive overview on the state-of-the-art of graph learning. Special attention is paid to four categories of existing graph learning methods, including graph signal processing, matrix factorization, random walk, and deep learning. Major models and algorithms under these categories are reviewed, respectively. We examine graph learning applications in areas such as text, images, science, knowledge graphs, and combinatorial optimization. In addition, we discuss several promising research directions in this field. Impact Statement—Real-world intelligent systems generally rely on machine learning algorithms handling data of various types. Despite their ubiquity, graph data have imposed unprecedented challenges to machine learning due to their inherent complexity. Unlike text, audio and images, graph data are embedded in an irregular domain, making some essential operations of existing machine learning algorithms inapplicable. Many graph learning models and algorithms have been developed to tackle these challenges. This article presents a systematic review of the state-of-the-art graph learning approaches as well as their potential applications. The article serves multiple purposes. First, it acts as a quick reference to graph learning for researchers and practitioners in different areas such as social computing, information retrieval, computer vision, bioinformatics, economics, and e-commence. Second, it presents insights into open areas of research in the field. Third, it aims to stimulate new research ideas and more interests in graph learning. © IEEE Transactions on Artificial Intelligence 2020.
- Authors: Xia, Feng , Sun, Ke , Yu, Shuo , Aziz, Abdul , Wan, Liangtian , Pan, Shirui , Liu, Huan
- Date: 2021
- Type: Text , Journal article , Review
- Relation: IEEE Transactions on Artificial Intelligence Vol. 2, no. 2 (2021), p. 109-127
- Full Text:
- Reviewed:
- Description: Graphs are widely used as a popular representation of the network structure of connected data. Graph data can be found in a broad spectrum of application domains such as social systems, ecosystems, biological networks, knowledge graphs, and information systems. With the continuous penetration of artificial intelligence technologies, graph learning (i.e., machine learning on graphs) is gaining attention from both researchers and practitioners. Graph learning proves effective for many tasks, such as classification, link prediction, and matching. Generally, graph learning methods extract relevant features of graphs by taking advantage of machine learning algorithms. In this survey, we present a comprehensive overview on the state-of-the-art of graph learning. Special attention is paid to four categories of existing graph learning methods, including graph signal processing, matrix factorization, random walk, and deep learning. Major models and algorithms under these categories are reviewed, respectively. We examine graph learning applications in areas such as text, images, science, knowledge graphs, and combinatorial optimization. In addition, we discuss several promising research directions in this field. Impact Statement—Real-world intelligent systems generally rely on machine learning algorithms handling data of various types. Despite their ubiquity, graph data have imposed unprecedented challenges to machine learning due to their inherent complexity. Unlike text, audio and images, graph data are embedded in an irregular domain, making some essential operations of existing machine learning algorithms inapplicable. Many graph learning models and algorithms have been developed to tackle these challenges. This article presents a systematic review of the state-of-the-art graph learning approaches as well as their potential applications. The article serves multiple purposes. First, it acts as a quick reference to graph learning for researchers and practitioners in different areas such as social computing, information retrieval, computer vision, bioinformatics, economics, and e-commence. Second, it presents insights into open areas of research in the field. Third, it aims to stimulate new research ideas and more interests in graph learning. © IEEE Transactions on Artificial Intelligence 2020.