Most Read articles

    Published in last 1 year |  In last 2 years |  In last 3 years |  All

    Published in last 1 year
    Please wait a minute...
    For Selected: Toggle Thumbnails
    Review of Deep Learning Applied to Time Series Prediction
    LIANG Hongtao, LIU Shuo, DU Junwei, HU Qiang, YU Xu
    Journal of Frontiers of Computer Science and Technology    2023, 17 (6): 1285-1300.   DOI: 10.3778/j.issn.1673-9418.2211108
    The time series is generally a set of random variables that are observed and collected at a certain frequency in the course of something??s development. The task of time series forecasting is to extract the core patterns from a large amount of data and to make accurate estimates of future data based on known factors. Due to the access of a large number of IoT data collection devices, the explosive growth of multidimensional data and the increasingly demanding requirements for prediction accuracy, it is difficult for classical parametric models and traditional machine learning algorithms to meet high efficiency and high accuracy requirements of prediction tasks. In recent years, deep learning algorithms represented by convolutional neural networks, recurrent neural networks and Trans-former models have achieved fruitful results in time series forecasting tasks. To further promote the development of time series prediction technology, common characteristics of time series data, evaluation indexes of datasets and models are reviewed, and the characteristics, advantages and limitations of each prediction algorithm are experimentally compared and analyzed with time and algorithm architecture as the main research line. Several time series prediction methods based on Transformer model are highlighted and compared. Finally, according to the problems and challenges of deep learning applied to time series prediction tasks, this paper provides an outlook on the future research trends in this direction.
    Reference | Related Articles | Metrics
    Abstract3524
    PDF3931
    Research on Question Answering System on Joint of Knowledge Graph and Large Language Models
    ZHANG Heyi, WANG Xin, HAN Lifan, LI Zhao, CHEN Zirui, CHEN Zhe
    Journal of Frontiers of Computer Science and Technology    2023, 17 (10): 2377-2388.   DOI: 10.3778/j.issn.1673-9418.2308070
    The large language model (LLM), including ChatGPT, has shown outstanding performance in understanding and responding to human instructions, and has a profound impact on natural language question answering (Q&A). However, due to the lack of training in the vertical field, the performance of LLM in the vertical field is not ideal. In addition, due to its high hardware requirements, training and deploying LLM remains difficult. In order to address these challenges, this paper takes the application of traditional Chinese medicine formulas as an example, collects the domain related data and preprocesses the data. Based on LLM and knowledge graph, a vertical domain Q&A system is designed. The system has the following capabilities: (1) Information filtering. Filter out vertical domain related questions and input them into LLM to answer. (2) Professional Q&A. Generate answers with more professional knowledge based on LLM and self-built knowledge base. Compared with the fine-tuning method of introducing professional data, using this technology can deploy large vertical domain models without the need for retraining. (3) Extract conversion. By strengthening the information extraction ability of LLM and utilizing its generated natural language responses, structured knowledge is extracted and matched with a professional knowledge graph for professional verification. At the same time, structured knowledge can be transformed into readable natural language, achieving a deep integration of large models and knowledge graphs. Finally, the effect of the system is demonstrated and the performance of the system is verified from both subjective and objective perspectives through two experiments of subjective evaluation of experts and objective evaluation of multiple choice questions.
    Reference | Related Articles | Metrics
    Abstract2677
    PDF2583
    Review of Medical Image Segmentation Based on UNet
    XU Guangxian, FENG Chun, MA Fei
    Journal of Frontiers of Computer Science and Technology    2023, 17 (8): 1776-1792.   DOI: 10.3778/j.issn.1673-9418.2301044
    As one of the most important semantic segmentation frameworks in convolutional neural networks (CNN), UNet is widely used in image processing tasks such as classification, segmentation, and target detection of medical images. In this paper, the structural principles of UNet are described, and a comprehensive review of UNet-based networks and variant models is presented. The model algorithms are fully investigated from several perspectives, and an attempt is made to establish an evolutionary pattern among the models. Firstly, the UNet variant models are categorized according to the seven medical imaging systems they are applied to, and the algorithms with similar core composition are compared and described. Secondly, the principles, strengths and weaknesses, and applicable scenarios of each model are analyzed. Thirdly, the main UNet variant networks are summarized in terms of structural principles, core composition, datasets, and evaluation metrics. Finally, the inherent shortcomings and solutions of the UNet network structure are objectively described in light of the latest advances in deep learning, providing directions for continued improvement in the future. At the same time, other technological evolutions and application scenarios that can be combined with UNet are detailed, and the future development trend of UNet-based variant networks is further envisaged.
    Reference | Related Articles | Metrics
    Abstract1403
    PDF1514
    Research Review of Image Semantic Segmentation Method in High-Resolution Remote Sensing Image Interpretation
    MA Yan, Gulimila·Kezierbieke
    Journal of Frontiers of Computer Science and Technology    2023, 17 (7): 1526-1548.   DOI: 10.3778/j.issn.1673-9418.2211015
    Rapid acquisition of remote sensing information has important research significance for the development of image semantic segmentation methods in remote sensing image interpretation applications. With more and more types of data recorded by satellite remote sensing images and more and more complex feature information, accurate and effective extraction of information in remote sensing images has become the key to interpret remote sensing images by image semantic segmentation methods. In order to explore the image semantic segmentation method for fast and efficient interpretation of remote sensing images, a large number of image semantic segmentation methods for remote sensing images are summarized. Firstly, the traditional image semantic segmentation methods are reviewed and divided into edge detection-based segmentation methods, region-based segmentation methods, threshold-based segmentation methods and segmentation methods combined with specific theories. At the same time, the limitations of traditional image semantic segmentation methods are analyzed. Secondly, the semantic segmentation methods based on deep learning are elaborated in detail, and the basic ideas and technical characteristics of each method are used as the classification criteria. They are divided into four categories: FCN-based methods, codec-based methods, dilated convolution-based methods and attention-based methods. The sub-methods contained in each type of method are summarized, and the advantages and disadvantages of these methods are compared and analyzed. Then, the common datasets and performance evaluation indexes of remote sensing image semantic segmentation are briefly introduced. Experimental results of classical network models on different datasets are given, and the performance of different models is evaluated. Finally, the challenges of image semantic segmentation methods in high-resolution remote sensing image interpretation are analyzed, and the future development trend is prospected.
    Reference | Related Articles | Metrics
    Abstract1093
    PDF838
    Survey of Causal Inference for Knowledge Graphs and Large Language Models
    LI Yuan, MA Xinyu, YANG Guoli, ZHAO Huiqun, SONG Wei
    Journal of Frontiers of Computer Science and Technology    2023, 17 (10): 2358-2376.   DOI: 10.3778/j.issn.1673-9418.2307065
    In recent decades, causal inference has been a significant research topic in various fields, including statistics, computer science, education, public policy, and economics. Most causal inference methods focus on the analysis of sample observational data and text corpora. However, with the emergence of various knowledge graphs and large language models, causal inference tailored to knowledge graphs and large models has gradually become a research hotspot. In this paper, different causal inference methods are classified based on their orientation towards sample observational data, text data, knowledge graphs, and large language models. Within each classification, this paper provides a detailed analysis of classical research works, including their problem definitions, solution methods, contributions, and limitations. Additionally, this paper places particular emphasis on discussing recent advancements in the integration of causal inference methods with knowledge graphs and large language models. Various causal inference methods are analyzed and compared from the perspectives of efficiency and cost, and specific applications of knowledge graphs and large language models in causal inference tasks are summarized. Finally, future development directions of causal inference in combination with knowledge graphs and large models are prospected.
    Reference | Related Articles | Metrics
    Abstract1051
    PDF1214
    Review on Multi-lable Classification
    LI Dongmei, YANG Yu, MENG Xianghao, ZHANG Xiaoping, SONG Chao, ZHAO Yufeng
    Journal of Frontiers of Computer Science and Technology    2023, 17 (11): 2529-2542.   DOI: 10.3778/j.issn.1673-9418.2303082
    Multi-label classification refers to the classification problem where multiple labels may coexist in a single sample. It has been widely applied in fields such as text classification, image classification, music and video classification. Unlike traditional single-label classification problems, multi-label classification problems become more complex due to the possible correlation or dependence among labels. In recent years, with the rapid development of deep learning technology, many multi-label classification methods combined with deep learning have gradually become a research hotspot. Therefore, this paper summarizes the multi-label classification methods from the traditional and deep learning-based perspectives, and analyzes the key ideas, representative models, and advantages and disadvantages of each method. In traditional multi-label classification methods, problem transformation methods and algorithm adaptation methods are introduced. In deep learning-based multi-label classification methods, the latest multi-label classification methods based on Transformer are reviewed particularly, which have become one of the mainstream methods to solve multi-label classification problems. Additionally, various multi-label classification datasets from different domains are introduced, and 15 evaluation metrics for multi-label classification are briefly analyzed. Finally, future work is discussed from the perspectives of multi-modal data multi-label classification, prompt learning-based multi-label classification, and imbalanced data multi-label classification, in order to further promote the development and application of multi-label classification.
    Reference | Related Articles | Metrics
    Abstract900
    PDF824
    Survey of Fake News Detection with Multi-model Learning
    LIU Hualing, CHEN Shanghui, CAO Shijie, ZHU Jianliang, REN Qingqing
    Journal of Frontiers of Computer Science and Technology    2023, 17 (9): 2015-2029.   DOI: 10.3778/j.issn.1673-9418.2301064
    While social media brings convenience to people, it has also become a channel for the arbitrary spread of fake news. If not detected and stopped in time, it is easy to cause public panic and social unrest. Therefore, exploring accurate and efficient fake news detection technology has high theoretical value and practical significance. This paper provides a comprehensive overview of the related fake news detection techniques. Firstly, the relevant concepts of multi-modal fake news are sorted and summarized, and the trend of changes in single-modal and multi-modal news datasets is analyzed. Secondly, this paper introduces single-modal fake news detection techniques based on machine learning and deep learning, which have been widely used in the field of fake news detection. However, traditional single-modal techniques cannot fully explore the deep logic of fake news because fake news usually contains multiple data presentations. Thus, they are unable to effectively deal with the challenges brought by multi-modal fake news data. To address this issue, this paper summarizes and discusses advanced multi-modal fake news detection techniques from the perspectives of multi-stream and graph architectures, and explores their concepts and potential drawbacks. Finally, this paper analyzes the difficulties and bottlenecks in the current research on fake news detection and provides future research directions based on these analyses.
    Reference | Related Articles | Metrics
    Abstract725
    PDF706
    YOLOv8-VSC: Lightweight Algorithm for Strip Surface Defect Detection
    WANG Chunmei, LIU Huan
    Journal of Frontiers of Computer Science and Technology    2024, 18 (1): 151-160.   DOI: 10.3778/j.issn.1673-9418.2308060
    Currently, in the field of strip steel surface defect detection, the generalized target detection algorithm is highly complex and computationally large, while terminal equipment responsible for the detection of some small and medium-sized enterprises usually does not have strong computational capabilities, and the computational resources are limited, which leads to difficulties in the deployment of detection algorithms. To solve this problem, this paper proposes a lightweight strip steel surface defect detection model YOLOv8-VSC based on the YOLOv8n target detec-tion framework, which uses a lightweight VanillaNet network as the backbone feature extraction network and reduces the complexity of the model by reducing the unnecessary branching structure. Meanwhile, the SPD module is introduced to speed up the inference of the model while reducing the number of network layers. To further improve the detection accuracy, a lightweight up-sampling operator, CARAFE, is used in the feature fusion network to improve the quality and richness of the features. Finally, extensive experiments on the NEU-DET dataset yield a model with parametric and computational quantities of 1.96×106 and 6.0 GFLOPs, which are only 65.1% and 74.1% of the baseline, and the mAP reaches 80.8%, which is an improvement of 1.8 percentage points from the baseline. In addition, experimental results on the aluminum surface defect dataset and the VOC2012 dataset show that the proposed algorithm has good robustness. Compared with advanced target detection algorithms, the proposed algorithm requires fewer computational resources while ensuring high detection accuracy.
    Reference | Related Articles | Metrics
    Abstract667
    PDF549
    Advances in Knowledge Graph Embedding Based on Graph Neural Networks
    YAN Zhaoyao, DING Cangfeng, MA Lerong, CAO Lu, YOU Hao
    Journal of Frontiers of Computer Science and Technology    2023, 17 (8): 1793-1813.   DOI: 10.3778/j.issn.1673-9418.2212063
    As graph neural networks continue to develop, knowledge graph embedding methods based on graph neural networks are receiving increasing attention from researchers. Compared with traditional methods, they can better handle the diversity and complexity of entities, and capture the multiple features and complex relationships of entities, thereby improving the representation ability and application value of knowledge graphs. This paper firstly outlines the development history of knowledge graphs and the basic concepts of knowledge graphs and graph neural networks. Secondly, it focuses on discussing the design ideas and algorithm frameworks of knowledge graph embedding based on graph convolution, graph neural networks, graph attention, and graph autoencoders. Then, it describes the performance of graph neural network knowledge graph embedding in tasks such as link prediction, entity alignment, knowledge graph reasoning, and knowledge graph completion, while supplementing some research on commonsense knowledge graphs with graph neural networks. Finally, this paper makes a comprehensive summary, and future research directions are outlined with respect to some challenges and issues in knowledge graph embedding.
    Reference | Related Articles | Metrics
    Abstract588
    PDF566
    Named Entity Recognition Method of Large Language Model for Medical Question Answering System
    YANG Bo, SUN Xiaohu, DANG Jiayi, ZHAO Haiyan, JIN Zhi
    Journal of Frontiers of Computer Science and Technology    2023, 17 (10): 2389-2402.   DOI: 10.3778/j.issn.1673-9418.2307061
    In medical question answering systems, entity recognition plays a major role. Entity recognition based on deep learning has received more and more attention. However, in the medical question answering system, due to the lack of annotated training data, deep learning methods cannot well identify discontinuous and nested entities in medical text. Therefore, a large language model-based entity recognition application method is proposed, and it is applied to the medical problem system. Firstly, the dataset related to medical question answering is processed into text that can be analyzed and processed by a large language model. Secondly, the output of the large language model is classified, and different classifications are processed accordingly. Then, the input text is used for intent recognition, and  finally the results of entity recognition and intent recognition are sent to the medical knowledge graph for query, and the answer to the medical question and answer is obtained. Experiments are performed on 3 typical datasets and compared with several typical correlation methods. The results show that the method proposed in this paper performs better.
    Reference | Related Articles | Metrics
    Abstract579
    PDF426
    Review of Research on Fatigue Driving Detection Based on Driver Facial Features
    YANG Yanyan, LI Leixiao, LIN Hao
    Journal of Frontiers of Computer Science and Technology    2023, 17 (6): 1249-1267.   DOI: 10.3778/j.issn.1673-9418.2208041
    Fatigue driving is one of the main factors that threaten the safety of drivers and traffic. Efficient and accurate fatigue driving detection method can effectively ensure the safety of drivers and their surrounding traffic, maintain traffic order, and reduce property losses and casualties. The fatigue driving detection method based on the driver’s physiological characteristics and vehicle driving information has many limitations, such as being unfriendly to the driver and having numerous influencing factors. Therefore, the fatigue driving detection method based on the driver’s facial features has become a research hotspot. Firstly, the facial feature performance of fatigue driving is described, and advantages, disadvantages and application scenarios of common public datasets in the field of fatigue driving are summarized. Secondly, the advantages and disadvantages of common face detection algorithms in the field of fatigue driving detection are analyzed and studied by using open datasets and comparative experiments. Then, this paper generalizes the process of detection methods based on driver facial features, whose methods and technologies used in the key steps are reviewed. Furthermore, fatigue discriminant parameters and methods of fatigue driving results prediction are summarized. Finally, this paper ends with a discussion of current challenges of fatigue driving detection methods based on driver facial features and looks forward to the future research.
    Reference | Related Articles | Metrics
    Abstract526
    PDF617
    Word Embedding Methods in Natural Language Processing: a Review
    ZENG Jun, WANG Ziwei, YU Yang, WEN Junhao, GAO Min
    Journal of Frontiers of Computer Science and Technology    2024, 18 (1): 24-43.   DOI: 10.3778/j.issn.1673-9418.2303056
    Word embedding, as the first step in natural language processing (NLP) tasks, aims to transform input natural language text into numerical vectors, known as word vectors or distributed representations, which artificial intelligence models can process. Word vectors, the foundation of NLP tasks, are a prerequisite for accomplishing various NLP downstream tasks. However, most existing review literature on word embedding methods focuses on the technical routes of different word embedding methods, neglecting comprehensive analysis of the tokenization methods and the complete evolutionary trends of word embedding. This paper takes the introduction of the word2vec model and the Transformer model as pivotal points. From the perspective of whether generated word vectors can dynamically change their implicit semantic information to adapt to the overall semantics of input sentences, this paper categorizes word embedding methods into static and dynamic approaches and extensively discusses this classification. Simultaneously, it compares and analyzes tokenization methods in word embedding, including whole and sub-word segmentation. This paper also provides a detailed enumeration of the evolution of language models used to train word vectors, progressing from probability language models to neural probability language models and the current deep contextual language models. Additionally, this paper summarizes and explores the training strategies employed in pre-training language models. Finally, this paper concludes with a summary of methods for evaluating word vector quality, an analysis of the current state of word embedding methods, and a prospective outlook on their development.
    Reference | Related Articles | Metrics
    Abstract517
    PDF370
    Survey on Inductive Learning for Knowledge Graph Completion
    LIANG Xinyu, SI Guannan, LI Jianxin, TIAN Pengxin, AN Zhaoliang, ZHOU Fengyu
    Journal of Frontiers of Computer Science and Technology    2023, 17 (11): 2580-2604.   DOI: 10.3778/j.issn.1673-9418.2303063
    Knowledge graph completion can make knowledge graph more complete. However, traditional knowledge graph completion methods assume that all test entities and relations appear in the training process. Due to the evolving nature of real world KG, once unseen entities or relations appear, the knowledge graph needs to be retrained. Inductive learning for knowledge graph completion aims to complete triples containing unseen entities or unseen relations without training the knowledge graph from scratch, so it has received much attention in recent years. Firstly, starting from the basic concept of knowledge graph, this paper divides knowledge graph completion into two categories: transductive and inductive. Secondly, from the theoretical perspective of inductive knowledge graph completion, it is divided into two categories: semi-inductive and fully-inductive, and the models are summarized from this perspective. Then, from the technical perspective of inductive knowledge graph completion, it is divided into two categories: based on structural information and based on additional information. The methods based on structural information are subdivided into three categories: based on inductive embedding, based on logical rules and based on meta learning, and the methods based on additional information are subdivided into two categories: based on text information and other information. The current methods are further subdivided, analyzed and compared. Finally, it forecasts the main research directions in the future.
    Reference | Related Articles | Metrics
    Abstract485
    PDF571
    Survey on Blockchain-Based Cross-Domain Authentication for Internet of Things Terminals
    HUO Wei, ZHANG Qionglu, OU Wei, HAN Wenbao
    Journal of Frontiers of Computer Science and Technology    2023, 17 (9): 1995-2014.   DOI: 10.3778/j.issn.1673-9418.2211004
    Internet of things (IoT) devices are widely distributed, numerous and complex, which are involved in multiple management domains. They are often in uncontrollable environments and are more vulnerable to attacks than traditional Internet terminals, the security management and protection of IoT terminals face greater risks and challenges. As “the first line of defense” for the security protection of IoT devices, authentication plays an irreplaceable and important role in the development of IoT security. The blockchain technology has the characteristics and advantages of decentralization, distribution, immutability and traceability. And thus, it can effectively solve the single-point trust failure of trusted third parties and satisfy the principle of least authorization for multi-domain heterogeneity in cross-domain authentication for IoT terminals. Using the blockchain technology is an important trend in the future development of the IoT device cross-domain authentication. This paper categorizes and summarizes the main research achievements of IoT cross-domain authentication based on blockchain technology in recent years according to three categories: integrating traditional identity authentication mechanisms such as PKI and IBS/IBC, adopting inter-blockchain technology, and other cross-domain authentication technologies based on blockchain. Then this paper analyzes the technical characteristics, advantages and disadvantages of each different scheme. On this basis, the current problems and issues in the field of cross-domain authentication of IoT devices are summarized, and the future research directions and development suggestions for the cross-domain authentication of IoT terminals are given, so as to achieve a general and overall grasp of the research progress and development trend of IoT device cross-domain authentication schemes based on blockchain technology.
    Reference | Related Articles | Metrics
    Abstract480
    PDF528
    Review of Attention Mechanisms in Image Processing
    QI Xuanhao, ZHI Min
    Journal of Frontiers of Computer Science and Technology    2024, 18 (2): 345-362.   DOI: 10.3778/j.issn.1673-9418.2305057
    Attention mechanism in image processing has become one of the popular and important techniques in the field of deep learning, and is widely used in various deep learning models in image processing because of its excellent plug-and-play convenience. By weighting the input features, the attention mechanism focuses the model’s attention on the most important regions to improve the accuracy and performance of image processing tasks. Firstly, this paper divides the development process of attention mechanism into four stages, and on this basis, reviews and summarizes the research status and progress of four aspects: channel attention, spatial attention, channel and spatial mixed attention, and self-attention. Secondly, this paper provides a detailed discussion on the core idea, key structure and specific implementation of attention mechanism, and further summarizes the advantages and disadvantages of used models. Finally, by comparing the current mainstream attention mechanisms and analyzing the results, this paper discusses the problems of attention mechanisms in the image processing field at this stage, and provides an outlook on the future development of attention mechanisms in image processing, so as to provide references for further research.
    Reference | Related Articles | Metrics
    Abstract478
    PDF361
    Review of Privacy-Preserving Research in Recommendation Systems
    FENG Han, YI Huawei, LI Xiaohui, LI Rui
    Journal of Frontiers of Computer Science and Technology    2023, 17 (8): 1814-1832.   DOI: 10.3778/j.issn.1673-9418.2211069
    The recommendation system needs to extract the historical data information of the relevant users on a large scale as the training set of the prediction model. The larger and more specific the amount of data provided by users is, the easier it is for the personal information to be inferred, which is easy to lead to the leakage of personal privacy, so that the user’s trust in the service provider is reduced and the relevant data are no longer provided for the system, resulting in the reduction of the system recommendation accuracy or even more difficult to complete the recommendation. Therefore, how to obtain user information for effective recommendation with high accuracy under the premise of protecting user privacy has become a research hotspot. This paper firstly summarizes the privacy-preserving technology, including differential privacy technology, homomorphic encryption technology, federated learning and secure multi-party computing technology, and compares these commonly used privacy-preserving tech-nologies. Then, from the perspective of balancing the relationship between privacy injection and recommendation accuracy, the privacy-preserving technologies adopted by the user side, the server side and the user-server side are introduced, and the research results of privacy-preserving in recommendation systems at home and abroad are systematically elaborated. Based on this, a summary, comparison, and analysis are conducted. Next, the experi-mental results of the recommendation algorithm based on differential privacy technology are compared, and the shortcomings of the corresponding technology are analyzed. Finally, the future development directions of the recommendation system based on privacy-preserving are prospected.
    Reference | Related Articles | Metrics
    Abstract476
    PDF638
    Review of Research on Rolling Bearing Health Intelligent Monitoring and Fault Diagnosis Mechanism
    WANG Jing, XU Zhiwei, LIU Wenjing, WANG Yongsheng, LIU Limin
    Journal of Frontiers of Computer Science and Technology    2024, 18 (4): 878-898.   DOI: 10.3778/j.issn.1673-9418.2307005
    As one of the most critical and failure-prone parts of the mechanical systems of industrial equipment, bearings are subjected to high loads for long periods of time. When they fail or wear irreversibly, they may cause accidents or even huge economic losses. Therefore, effective health monitoring and fault diagnosis are of great significance to ensure safe and stable operation of industrial equipment. In order to further promote the development of bearing health monitoring and fault diagnosis technology, the current existing models and methods are analyzed and summarized, and the existing technologies are divided and compared. Starting from the distribution of vibration signal data used, firstly, the relevant methods under uniform data distribution are sorted out, the classification, analysis and summary of the current research status are carried out mainly according to signal-based analysis and data-driven-based, and the shortcomings and defects of the fault detection methods in this case are outlined. Secondly, considering the problem of uneven data acquisition under actual working conditions, the detection methods for dealing with such cases are summarized, and different processing techniques for this problem in existing research are classified into data processing methods, feature extraction methods, and model improvement methods according to their different focuses, and the existing problems are analyzed and summarized. Finally, the challenges and future development directions of bearing fault detection in existing industrial equipment are summarized and prospected.
    Reference | Related Articles | Metrics
    Abstract468
    PDF232
    Review of Visual Question Answering Technology
    WANG Yu, SUN Haichun
    Journal of Frontiers of Computer Science and Technology    2023, 17 (7): 1487-1505.   DOI: 10.3778/j.issn.1673-9418.2303025
    Visual question answering (VQA) is a popular cross-modal task that combines natural language pro-cessing and computer vision techniques. The main objective of this task is to enable computers to intelligently recognize and retrieve visual content and provide accurate answers. VQA involves the integration of multiple technologies such as object recognition and detection, intelligent question answering, image attribute classification, and scene analysis. It can support a wide range of cutting-edge interactive AI tasks such as visual dialogue and visual navigation, and has broad application prospects and great value. Over the past few years, the development of computer vision, natural language processing, and cross-modal AI models has provided many new technologies and methods for achieving the task of visual question answering. This paper mainly summarizes the mainstream models and specialized datasets in the field of visual question answering between 2019 and 2022. Firstly, this paper provides a review and discussion of the mainstream technical methods used in the key steps of implementing the visual question answering task, based on the module framework. Next, it subdivides various types of models in this field according to the technical methods adopted by mainstream models, and briefly introduces their improvement focus and limitations. Then, it summarizes the commonly used datasets and evaluation metrics for visual question answering, and compares and discusses the performance of several typical models. Finally, this paper focuses on the key issues that need to be addressed in the current visual question answering field, and predicts and prospects the future application and technological development in this field.
    Reference | Related Articles | Metrics
    Abstract467
    PDF534
    Survey on Few-Shot Knowledge Graph Completion Technology
    PENG Yanfei, ZHANG Ruisi, WANG Ruihua, GUO Jialong
    Journal of Frontiers of Computer Science and Technology    2023, 17 (6): 1268-1284.   DOI: 10.3778/j.issn.1673-9418.2209069
    Few-shot knowledge graph completion (FKGC) is a new research hotspot in the field of knowledge graph completion, which aims to complete knowledge graph with a few samples of data. This task is of great importance in practical application and the fields of knowledge graph. In order to further promote the development of the field of FKGC, this paper summarizes and analyzes the current methods. Firstly, this paper describes the concept of FKGC and related content. Secondly, three types of FKGC methods are summarized based on technical methods, including scale learning-based methods, meta learning-based methods, and other model-based methods. In addition, this paper analyzes and summarizes each method from the perspectives of model core, model ideas, advantages and disadvantages, etc. Then, the datasets and evaluation indexes of FKGC method are summarized, and the FKGC method is analyzed from two aspects of model characteristics and experimental results. Finally, starting from the practical problems, this paper summarizes the difficult problems of the current FKGC task, analyses the difficulties behind the problems, gives the corresponding solutions, and prospects several development directions that deserve attention in this field in the future.
    Reference | Related Articles | Metrics
    Abstract453
    PDF438
    Deep Learning-Based Infrared and Visible Image Fusion: A Survey
    WANG Enlong, LI Jiawei, LEI Jia, ZHOU Shihua
    Journal of Frontiers of Computer Science and Technology    2024, 18 (4): 899-915.   DOI: 10.3778/j.issn.1673-9418.2306061
    How to preserve the complementary information in multiple images to represent the scene in one image is a challenging topic. Based on this topic, various image fusion methods have been proposed. As an important branch of image fusion, infrared and visible image fusion (IVIF) has a wide range of applications in segmentation, target detection and military reconnaissance fields. In recent years, deep learning has led the development direction of image fusion. Researchers have explored the field of IVIF using deep learning. Relevant experimental work has proven that applying deep learning to achieving IVIF has significant advantages compared with traditional methods. This paper provides a detailed analysis on the advanced algorithms for IVIF based on deep learning. Firstly, this paper reports on the current research status from the aspects of network architecture, method innovation, and limitations. Secondly, this paper introduces the commonly used datasets in IVIF methods and provides the definition of commonly used evaluation metrics in quantitative experiments. Qualitative and quantitative evaluation experiments of fusion and segmentation and fusion efficiency analysis experiments are conducted on some representative methods mentioned in the paper to comprehensively evaluate the performance of the methods. Finally, this paper provides conclusions and prospects for possible future research directions in the field.
    Reference | Related Articles | Metrics
    Abstract453
    PDF436
    Review of Research on 3D Reconstruction of Dynamic Scenes
    SUN Shuifa, TANG Yongheng, WANG Ben, DONG Fangmin, LI Xiaolong, CAI Jiacheng, WU Yirong
    Journal of Frontiers of Computer Science and Technology    2024, 18 (4): 831-860.   DOI: 10.3778/j.issn.1673-9418.2305016
    As static scene 3D reconstruction algorithms become more mature, dynamic scene 3D reconstruction has become a hot and challenging research topic in recent years. Existing static scene 3D reconstruction algorithms have good reconstruction results for stationary objects. However, when objects in the scene undergo deformation or relative motion, their reconstruction results are not ideal. Therefore, developing research on 3D reconstruction of dynamic scenes is essential. This paper first introduces the related concepts and basic knowledge of 3D reconstruction, as well as the research classification and current status of static and dynamic scene 3D reconstruction. Then, the latest research progress on dynamic scene 3D reconstruction is comprehensively summarized, and the reconstruction algorithms are classified into dynamic 3D reconstruction based on RGB data sources and dynamic 3D reconstruction based on RGB-D data sources. RGB data sources can be further divided into template based dynamic 3D reconstruction, non rigid motion recovery structure based dynamic 3D reconstruction, and learning based dynamic 3D reconstruction under RGB data sources. The RGB-D data source mainly summarizes dynamic 3D reconstruction based on learning, with typical examples introduced and compared. The applications of dynamic scene 3D reconstruction in medical, intelligent manufacturing, virtual reality and augmented reality, and transportation fields are also discussed. Finally, future research directions for dynamic scene 3D reconstruction are proposed, and an outlook on the research progress in this rapidly developing field is presented.
    Reference | Related Articles | Metrics
    Abstract403
    PDF394
    Survey of Deformable Convolutional Networks
    LIU Weiguang, LIU Dong, WANG Lu
    Journal of Frontiers of Computer Science and Technology    2023, 17 (7): 1549-1564.   DOI: 10.3778/j.issn.1673-9418.2209076
    In recent years, with the rapid development of deep learning, deformable convolutional networks have received extensive attention because of their powerful feature extraction capabilities, overcoming some problems that are difficult to solve in convolutional neural networks, and have played an important role in computer vision, natural language processing and other related fields. Since there is a little research on systematic summary of the deformable convolutional network, in order to provide a detailed reference for subsequent research, this paper summarizes the related work since the introduction of the deformable convolutional network. Firstly, this paper reviews the high-quality literature in recent years, and introduces the core technologies such as deformable convolution and deformable region of interest pooling in deformable convolutional networks from the perspective of invariant features. Secondly, the collected relevant literature is classified according to different research fields, and the appli-cation of deformable convolutional networks in image recognition and classification, target detection, image seg-mentation, target tracking and other research fields is comprehensively summarized. At the same time, the perfor-mance, advantages and disadvantages of important network models are listed. Thirdly, by combing the literature, the advantages and disadvantages of the deformable convolutional network are analyzed, and the possible future research trends of the deformable convolutional network are discussed according to some problems existing at the present stage. Finally, the deformable convolutional networks are summarized and prospected based on invariant feature extraction.
    Reference | Related Articles | Metrics
    Abstract400
    PDF355
    Traffic Prediction Method Integrating Graph Wavelet and Attention Mechanism
    XUE Yanming, LI Guanghui, QI Tao
    Journal of Frontiers of Computer Science and Technology    2023, 17 (6): 1405-1416.   DOI: 10.3778/j.issn.1673-9418.2111012
    Traffic predicting is a critical component of modern intelligent transportation systems for traffic management and control. However, the traffic flow is complex. On one hand, the urban road structure is highly correlative, and there often exists a nonlinear structural dependence between different roads. On the other hand, traffic flow data often change dynamically over time. In recent years, many studies have tried to use deep learning methods to extract complex structural features in traffic flow. However, the process of local feature extraction still lacks flexibility, and ignores the dynamic variability as well as the correlation of spatio-temporal features. To this end, this paper proposes a new traffic prediction method integrating graph wavelet and attention mechanism. This method uses wavelet transform and an adaptive matrix to extract local and global spatial features of traffic flow respectively, and combines the improved recurrent neural network to extract local temporal characteristic information. Meanwhile, the attention mechanism is adopted in this method to capture the temporal and spatial dynamic variability. Then this method applies a spatio-temporal feature fusion mechanism to fusing local and global temporal and spatial features. Experimental results show that this method can extract spatial and temporal features of real traffic datasets well, and it outperforms the existing methods.
    Reference | Related Articles | Metrics
    Abstract398
    PDF162
    Research Progress of Graph Neural Network in Knowledge Graph Construction and Application
    XU Xinran, WANG Tengyu, LU Cai
    Journal of Frontiers of Computer Science and Technology    2023, 17 (10): 2278-2299.   DOI: 10.3778/j.issn.1673-9418.2302059
    As an effective representation of knowledge, knowledge graph network can be used to represent rich factual information between different categories and become an effective knowledge management tool. It has achieved great results in the application and research of knowledge engineering and artificial intelligence. Know-ledge graph is usually expressed as a complex network structure. Its unstructured characteristics make the applica-tion of graph neural network to the analysis and research of knowledge graph become a research hotspot in academia. The purpose of this paper is to provide extensive research on knowledge graph construction technology based on graph neural network to solve two types of knowledge graph construction tasks, including knowledge extraction (entity, relationship and attribute extraction) and knowledge merging and processing (link prediction, entity alignment and knowledge reasoning, etc.). Through these tasks, the structure of knowledge graph can be further improved and new knowledge and reasoning relationships can be discovered. This paper also studies the advanced graph neural network method for knowledge graph related applications, such as recommendation system, question answering system and computer vision. Finally, the future research directions of knowledge graph application based on graph neural network are proposed.
    Reference | Related Articles | Metrics
    Abstract393
    PDF501
    Review on Research of Knowledge Tracking
    WU Shuixiu, LUO Xianzeng, XIONG Jian, ZHONG Maosheng, WANG Mingwen
    Journal of Frontiers of Computer Science and Technology    2023, 17 (7): 1506-1525.   DOI: 10.3778/j.issn.1673-9418.2210056
    Knowledge tracking, which aims to model students’ changing knowledge states over learning time based on their historical answer records and then predict students’ answer performance, is a core module supporting smart education systems and has received increasing attention from researchers. This paper comprehensively compares the research progress in this field, analyzes the basic theoretical research related to knowledge tracking, and analyzes the knowledge tracking models from probabilistic models, logical models, and deep learning-based models according to different research methods. Probabilistic models assume that learning follows Markov processes, logical models are a class of logic function-based models, and deep learning-based knowledge tracking models relying on the powerful feature extraction ability of deep learning have become a hot research topic in recent years. The improvement methods proposed for the problems faced by deep learning-based knowledge tracking models such as interpretability and lack of learning features are presented. The public datasets currently available to researchers are given as well as a comparison of the performance of different models. Finally, this paper concludes with a summary of this rapidly growing field on knowledge tracking, suggesting some possible future research directions for the problems of research in this area.
    Reference | Related Articles | Metrics
    Abstract384
    PDF322
    Survey on Visual Transformer for Image Classification
    PENG Bin, BAI Jing, LI Wenjing, ZHENG Hu, MA Xiangyu
    Journal of Frontiers of Computer Science and Technology    2024, 18 (2): 320-344.   DOI: 10.3778/j.issn.1673-9418.2310092
    Transformer is a deep learning model based on the self-attention mechanism, showing tremendous potential in computer vision. In image classification tasks, the key challenge lies in efficiently and accurately capturing both local and global features of input images. Traditional approaches rely on convolutional neural networks to extract local features at the lower layers, expanding the receptive field through stacked convolutional layers to obtain global features. However, this strategy aggregates information over relatively short distances, making it difficult to model long-term dependencies. In contrast, the self-attention mechanism of Transformer directly compares features across all spatial positions, capturing long-range dependencies at both local and global levels and exhibiting stronger global modeling capabilities. Therefore, a thorough exploration of the challenges faced by Transformer in image classification tasks is crucial. Taking Vision Transformer as an example, this paper provides a detailed overview of the core principles and architecture of Transformer. It then focuses on image classification tasks, summarizing key issues and recent advancements in visual Transformer research related to performance enhancement, computational costs, and training optimization. Furthermore, applications of Transformer in specific domains such as medical imagery, remote sensing, and agricultural images are summarized, highlighting its versatility and generality. Finally, a comprehensive analysis of the research progress in visual Transformer for image classification is presented, offering insights into future directions for the development of visual Transformer.
    Reference | Related Articles | Metrics
    Abstract377
    PDF298
    Low-Light Enhancement Method for Light Field Images by Fusing Multi-scale Features
    LI Mingyue, YAN Tao, JING Huahua, LIU Yuan
    Journal of Frontiers of Computer Science and Technology    2023, 17 (8): 1904-1916.   DOI: 10.3778/j.issn.1673-9418.2202064
    Light field images (LFI) record rich 3D structural and textural details of target scenes, which have great advantages in a wide range of computer vision tasks. However, LFI captured under low-light conditions always suffer from low brightness and strong noise, which may seriously degrade the quality of LFI. In this paper, a low-light LFI enhancement method by fusing multi-scale light field (LF) structural features is proposed , which adopts digital single-lens reflex camera (DSLR) images to supervise the generated enlightened LFI. To explore and exploit light field structural features, angular and spatial Transformers are introduced to extract LF structural features from LFI at different scales, i.e., the complementary information between sub-views and local and long-range dependencies within each sub-view. A recurrent fusion module is proposed to preserve the long-term memory of features at diffe-rent scales by using a long-short-term memory network, while adaptively aggregating LF structural features in the entire feature space through local and global fusion layers. A 4D residual reconstruction module is designed to reconstruct target LFI sub-views from the aggregated features. In additional, a dataset of low-light LFI and normal-light DSLR image pairs is constructed to train the proposed network. Extensive experiment demonstrates that the proposed network can effectively improve the quality of low-light LFI, and it obviously outperforms other state-of-the-art methods.
    Reference | Related Articles | Metrics
    Abstract362
    PDF425
    Design and Implementation of Efficient Multi-branch Predictor
    YANG Ling, ZHOU Jinwen, WANG Jing, LAN Mengqiao, DING Zijian, YANG Shi, WANG Yongwen, HUANG Libo
    Journal of Frontiers of Computer Science and Technology    2023, 17 (8): 1842-1851.   DOI: 10.3778/j.issn.1673-9418.2207069
    Branch prediction is a momentous technology guarantee for processor performance, especially for the widely used superscalar processor. The properties of the branch predictor significantly affect the overall performance, power consumption, and area of the processor. To obtain a more cost-effective branch predictor in the superscalar processor, an attempt is made to use a single TAGE (tagged geometric history length) predictor to predict the branches within the fetch width. The championship branch prediction platform is used to evaluate the performance of the predictor, and its prediction ability is sufficient to meet the prediction conditions. However, in practice, conflicts in both the predictor and branch target buffer affect its performance. To solve the above problem, this paper adds additional prediction paths based on a single TAGE branch predictor and independently saves and predicts additional branch instruction information. This predictor is implemented in the processor using hardware description language and compared with a single TAGE branch predictor to perform standard benchmark programs for embedded processors, dhrystone, coremark and embench. Experimental results show that the performance of the optimized branch predictor is improved by 14.1 percentage points, while the storage overhead is only increased by 9.06%. Finally, through the analysis of the experimental data, it is found that this scheme is not only conducive to the prediction of additional branch instructions, but also can achieve more accurate prediction of single branch instruction through more accurate acquisition of branch history information.
    Reference | Related Articles | Metrics
    Abstract360
    PDF374
    Influence Evaluation of Telecom Fraud Case Types Based on ChatGPT
    PEI Bingsen, LI Xin, WU Yue
    Journal of Frontiers of Computer Science and Technology    2023, 17 (10): 2413-2425.   DOI: 10.3778/j.issn.1673-9418.2306044
    At present, telecommunications fraud crimes are on the rise, posing a serious threat to the safety of people??s property. In order to optimize anti-fraud strategies, objectively and accurately analyze the trends and characteristics of different types of telecommunications fraud cases, and determine the most influential criminal methods, a ChatGPT based telecommunications fraud case type impact assessment method is proposed. By utilizing a knowledge graph, the content of the case text is structured, and the methods of telecommunications fraud are quantified by taking the time of the incident, the amount involved, and the number of individuals involved as factors to evaluate the impact of the case. Firstly, ChatGPT is used to preprocess and extract knowledge from the text corpus of telecommu-nications fraud cases through multiple rounds of Q&A, in order to quickly and timely construct a case knowledge graph in the field of telecommunications fraud with low resources. Based on the knowledge graph, various factors such as incident time, amount involved, and the number of involved parties are statistically analyzed, and the impact of different types of cases is abstracted into influencing factors. The influencing factors are used to depict the trend and characteristics of incidents, to conduct comprehensive analysis and judgment. This paper analyzes existing case data and calculates the impact factors of case types, obtaining the changes in impact factors of different case types, verifying the scientific and effective calculation methods of impact factors, and providing a new method for the evaluation of telecommunications fraud types. Combining the advantages of ChatGPT and knowledge graph helps to timely grasp the trend of case development and changes, provides strong support and guidance to combat teleco-mmunications fraud, and is of great significance for protecting public property safety and social stability.
    Reference | Related Articles | Metrics
    Abstract358
    PDF273
    Review of Application of Generative Adversarial Networks in Image Restoration
    GONG Ying, XU Wentao, ZHAO Ce, WANG Binjun
    Journal of Frontiers of Computer Science and Technology    2024, 18 (3): 553-573.   DOI: 10.3778/j.issn.1673-9418.2307073
    With the rapid development of generative adversarial networks, many image restoration problems that are difficult to solve based on traditional methods have gained new research approaches. With its powerful generation ability, generative adversarial networks can restore intact images from damaged images, so they are widely used in image restoration. In order to summarize the relevant theories and research on the problem of using generative adversarial networks to repair damaged images in recent years, based on the categories of damaged images and their adapted repair methods, the applications of image restoration are divided into three main aspects: image inpainting, image deblurring, and image denoising. For each aspect, the applications are further subdivided through technical principles, application objects and other dimensions. For the field of image inpainting, different image completion methods based on generative adversarial networks are discussed from the perspectives of using conditional guidance and latent coding. For the field of image deblurring, the essential differences between motion blurred images and static blurred images and their repair methods are explained. For the field of image denoising, personalized denoising methods for different categories of images are summarized. For each type of applications, the characteristics of the specific GAN models employed are summarized. Finally, the advantages and disadvantages of GAN applied to image restoration are summarized, and the future application scenarios are prospected.
    Reference | Related Articles | Metrics
    Abstract355
    PDF383
    Research on Blockchain Interoperability and Cross-Chain Technology
    WANG Qun, LI Fujuan, NI Xueli, XIA Lingling, LIANG Guangjun, MA Zhuo
    Journal of Frontiers of Computer Science and Technology    2023, 17 (8): 1749-1775.   DOI: 10.3778/j.issn.1673-9418.2212081
    Blockchain is a distributed ledger technology with multi-party consensus, traceability and tamper-proof, which provides a broad application prospect for constructing efficient, trusted and secure data sharing mechanism and optimizing business processes. However, when blockchain is in a stage of rapid development, how to realize cross-chain interaction of information and cross-chain transfer of value has become an urgent problem to be solved in the process of blockchain extension to depth. Firstly, based on the review of the existing research results, the concept of blockchain interoperability is proposed, and it is divided into five aspects: inter-chain interoperability, inter-layer interoperability, inter-fork interoperability, inter-slice interoperability and interoperability between on-chain and off-chain. Secondly, by sorting out the evolution and implementation of cross-chain operation of blockchain, and referring to the TCP/IP architecture, a cross-chain operation model is designed, and the main implementation steps are functionally described. Thirdly, in view of the current research status of cross-chain operation, four key technologies of cross-chain operation, including notary mechanism, side chain/relay, hash lock and distributed private key control, are selected for analysis. Then, combined with the technical characteristics and application scenarios, the application demonstration is highlighted, some typical cross-chain application projects are introduced, and the security of blockchain cross-chain operation is analyzed. Finally, the future development trend of blockchain interoperability and cross-chain technology is summarized and explored.
    Reference | Related Articles | Metrics
    Abstract346
    PDF387
    Improved YOLOv4-Tiny Lightweight Target Detection Algorithm
    HE Xiangjie, SONG Xiaoning
    Journal of Frontiers of Computer Science and Technology    2024, 18 (1): 138-150.   DOI: 10.3778/j.issn.1673-9418.2301034
    Object detection is an important branch of deep learning. A large number of edge devices need lightweight object detection algorithms, but the existing lightweight universal object detection algorithms have problems of low detection accuracy and slow detection speed. To solve this problem, an improved YOLOv4-Tiny algorithm based on attention mechanism is proposed. The structure of the original backbone network of YOLOv4-Tiny algorithm is adjusted, the ECA (efficient channel attention) attention mechanism is introduced, the traditional spatial pyramid pooling (SPP) structure is improved to DC-SPP structure by using void convolution, and the CSATT (channel spatial attention) attention mechanism is proposed. The neck network of CSATT-PAN (channel spatial attention path aggregation network) is formed with the feature fusion network PAN, which improves the feature fusion capability of the network. Compared with the original YOLOv4-Tiny algorithm, the proposed YOLOv4-CSATT algorithm is significantly more sensitive to information and accurate in classification when the detection speed is basically the same. The accuracy is increased by 12.3 percentage points on VOC dataset and 6.4 percentage points is increased on COCO dataset. Moreover, the accuracy is 3.3,5.5,6.3,17.4,10.3,0.9 and 0.6 percentage points higher than the Faster R-CNN, SSD, Efficientdet-d1, YOLOv3-Tiny, YOLOv4-MobileNetv1, YOLOv4-MobileNetv2 and PP-YOLO algorithms respectively on VOC dataset, and 2.8, 7.1, 4.2, 18.0, 12.2, 2.1 and 4.0 percentage points higher in recall rate, respectively, with an FPS of 94. In this paper, the CSATT attention mechanism is proposed to improve the model’s ability to capture spatial channel information, and the ECA attention mechanism is combined with the feature fusion pyramid algorithm to improve the model’s feature fusion ability and target detection accuracy.
    Reference | Related Articles | Metrics
    Abstract333
    PDF215
    Research Progress in Application of Deep Learning in Animal Behavior Analysis
    SHEN Tong, WANG Shuo, LI Meng, QIN Lunming
    Journal of Frontiers of Computer Science and Technology    2024, 18 (3): 612-626.   DOI: 10.3778/j.issn.1673-9418.2306033
    In recent years, animal behavior analysis has become one of the most important methods in the fields of neuroscience and artificial intelligence. Taking advantage of the powerful deep-learning-based image analysis technology, researchers have developed state-of-the-art automatic animal behavior analysis methods with complex functions. Compared with traditional methods of animal behavior analysis, special labeling is not required in these methods, animal pose can be efficiently estimated and tracked. These methods like in a natural environment, which hold the potential for complex animal behavior experiments. Therefore, the application of deep learning in animal behavior analysis is reviewed. Firstly, this paper analyzes the tasks and current status of animal behavior analysis. Then, it highlights and compares existing deep learning-based animal behavior analysis tools. According to the dimension of experimental analysis, the deep learning-based animal behavior analysis tools are divided into two-dimensional animal behavior analysis tools and three-dimensional animal behavior analysis tools, and the functions, performance and scope of application of tools are discussed. Furthermore, the existing animal datasets and evaluation metrics are introduced, and the algorithm mechanism used in the existing animal behavior analysis tool is summarized from the advantages, limitations and applicable scenarios. Finally, the deep learning-based animal behavior analysis tools are prospected from the aspects of dataset, experimental paradigm and low latency.
    Reference | Related Articles | Metrics
    Abstract333
    PDF401
    Survey of Deep Feature Instance Level Image Retrieval Algorithms
    JI Changqing, WANG Bingbing, QIN Jing, WANG Zumin
    Journal of Frontiers of Computer Science and Technology    2023, 17 (7): 1565-1575.   DOI: 10.3778/j.issn.1673-9418.2210125
    Content-based image retrieval algorithm (CBIR) aims to find semantically matching or similar images with query images. It analyzes visual content in a large number of image databases. It is important to obtain discriminant image representation by feature extraction. With the continuous development of deep learning, the image feature representation method used in image retrieval has gradually changed. The original extraction method is  based on manual features. Now it is based on deep features. From the perspective of different feature extraction methods, the recent image retrieval algorithms based on depth feature are reviewed and traced. The image retrieval algorithms based on depth feature are divided into two aspects: depth global feature and depth local feature. The deep convolution feature aggregation technique is emphasized in the deep local feature algorithm. The widely used image retrieval methods of deep global and local feature fusion are summarized. This paper discusses the practical application of deep feature image retrieval technology in remote sensing image retrieval, e-commerce product retrieval and medical image retrieval. And it compares the performance of these feature extraction algorithms in image retrieval accuracy. Finally, the future research trend of depth feature extraction in case image retrieval is forecasted.
    Reference | Related Articles | Metrics
    Abstract330
    PDF375
    Overview of Cross-Chain Identity Authentication Based on DID
    BAI Yirui, TIAN Ning, LEI Hong, LIU Xuefeng, LU Xiang, ZHOU Yong
    Journal of Frontiers of Computer Science and Technology    2024, 18 (3): 597-611.   DOI: 10.3778/j.issn.1673-9418.2304003
    With the emergence of concepts such as metaverse and Web3.0, blockchain plays a very important role in many fields. Cross-chain technology is an important technical means to achieve inter-chain interconnection and value transfer. At this stage, traditional cross-chain technologies such as notary and sidechain have trust issues. At the same time, in the field of cross-chain identity authentication, there are problems that the identities of each chain are not unified and users do not have control over their own identities. Firstly, it systematically summarizes the development process and technical solutions of digital identity and cross-chain technology, and analyzes and compares four digital identity models and nine mainstream cross-chain projects. Secondly, by analyzing the main research results of cross-chain identity authentication in recent years, a general model of cross-chain identity authentication is designed, and the shortcomings of existing solutions are summarized. Then, it focuses on the cross-chain identity authentication implementation scheme based on DID, and analyzes the technical characteristics, advantages and disadvantages of different solutions. On this basis, three DID-based cross-chain identity authentication models are summarized, the main implementation steps are functionally described, and their advantages, limitations and efficiency are analyzed. Finally, in view of the shortcomings of the current DID-based cross-chain identity authentication model, its development difficulties are discussed and five possible future research directions are given.
    Reference | Related Articles | Metrics
    Abstract329
    PDF369
    Using HLS to Develop FPGA Heterogeneous Acceleration System: Problems, Optimization Methods and Opportunities
    XU Cheng, GUO Jinyang, LI Chao, WANG Jing, WANG Taolei, ZHAO Jieru
    Journal of Frontiers of Computer Science and Technology    2023, 17 (8): 1729-1748.   DOI: 10.3778/j.issn.1673-9418.2210102
    Currently, field programmable gate arrays (FPGAs) are favored by both academia and industry due to their programmability and excellent energy efficiency ratio. However, traditional FPGA development based on hardware description languages faces programming challenges. Hardware description languages, which are different from commonly used high-level languages, hinder software developers from utilizing FPGAs. High-level synthesis (HLS) enables developers to directly develop FPGA hardware from high-level languages such as C/C++, and is widely regarded as the preferred solution to this problem. In recent years, there have been many works in academia on HLS, dedicated to solving various problems in the HLS application process and improving the performance of systems developed through HLS. This paper lists feasible optimization directions from the perspective of heterogeneous system developers around the issue of developing FPGA heterogeneous systems using HLS. At the compilation optimization level, HLS tools can automatically generate high-performance RTL designs by inserting compilation guidance and designing efficient spatial exploration algorithms. At the memory access optimization level, HLS tools can set up buffers, split and replicate data to improve the overall system bandwidth. At the parallel optimization level, HLS tools can implement statement-level, task-level and board-level parallelism. Meanwhile, some technologies such as DSL, although they cannot directly improve the performance of heterogeneous acceleration systems, can further enhance the usability of HLS tools. Finally, this paper summarizes some challenges currently faced by HLS and prospects the future research on HLS.
    Reference | Related Articles | Metrics
    Abstract327
    PDF398
    Review of Computing Offloading Schemes for Multi-access Edge Computing
    ZHANG Bingjie, YANG Yanhong, CAO Shaozhong
    Journal of Frontiers of Computer Science and Technology    2023, 17 (9): 2030-2046.   DOI: 10.3778/j.issn.1673-9418.2301068
    Under the background of the Internet of things, the development of massive and large-scale machine communication has brought about the explosive growth of data traffic. The traditional cloud computing model can no longer meet the needs of low delay and low energy consumption of terminal data processing. Multi-access edge computing (MEC) with distributed multi-nodes near the terminal side is becoming the best choice to solve this problem. Computational offloading is the key technology of MEC, the offloading performance is affected by many factors, and there is a large space for optimization. How to design a high-performance computational offloading scheme has become the main research goal of scholars at home and abroad. This paper reviews the research of computing offloading scheme for MEC, introduces the concept of MEC, sorts out the development and application of MEC, and the execution process of computing offloading, analyzes and compares the recent research methods of computing offloading. According to different improvements, a computing offloading scheme is summarized which takes the offloading system environment, offloading delay, energy consumption of mobile devices and multiple evaluation indexes as the optimization direction. The problems of resource allocation, universality and security in MEC-oriented computing offloading are presented, and the future research directions are forecasted based on these problems.
    Reference | Related Articles | Metrics
    Abstract326
    PDF424
    Differentiable Rule Extraction with Large Language Model for Knowledge Graph Reasoning
    PAN Yudai, ZHANG Lingling, CAI Zhongmin, ZHAO Tianzhe, WEI Bifan, LIU Jun
    Journal of Frontiers of Computer Science and Technology    2023, 17 (10): 2403-2412.   DOI: 10.3778/j.issn.1673-9418.2306049
    Knowledge graph (KG) reasoning is to predict missing entities or relationships in incomplete triples, complete structured knowledge, and apply to different downstream tasks. Different from black-box methods which are widely studied, such as methods based on representation learning, the method based on rule extraction achieves an interpretable reasoning paradigm by generalizing first-order logic rules from the KG. To address the gap between discrete symbolic space and continuous embedding space, a differentiable rule extracting method based on the large pre-trained language model (DRaM) is proposed, which integrates discrete first-order logical rules with continuous vector space. In view of the influence of atom sequences in first-order logic rules for the reasoning process, a large pre-trained language model is introduced to encode the reasoning process. The differentiable method DRaM, which integrates first-order logical rules, achieves good results in link prediction tasks on three knowledge graph datasets, Family, Kinship and UMLS, especially for the indicator Hits@10. Comprehensive experimental results show that DRaM can effectively solve the problems of differentiable reasoning on the KGs, and can extract first-order logic rules with confidences from the reasoning process. DRaM not only enhances the reasoning performance with the help of first-order logic rules, but also enhances the interpretability of the method.
    Reference | Related Articles | Metrics
    Abstract321
    PDF287
    MFFNet: Image Semantic Segmentation Network of Multi-level Feature Fusion
    WANG Yan, NAN Peiqi
    Journal of Frontiers of Computer Science and Technology    2024, 18 (3): 707-717.   DOI: 10.3778/j.issn.1673-9418.2209110
    In the task of image semantic segmentation, most methods do not make full use of features of different scales and levels, but directly upsampling, which will cause some effective information to be dismissed as redundant information, thus reducing the accuracy and sensitivity of segmentation of some small categories and similar categories. Therefore, a multi-level feature fusion network (MFFNet) is proposed. MFFNet uses encoder-decoder structure, during the encoding stage, the context information and spatial detail information are obtained through the context information extraction path and spatial information extraction path respectively to enhance the inter-pixel correlation and boundary accuracy. During the decoding stage, a multi-level feature fusion path is designed, and the context information is fused by the mixed bilateral fusion module. Deep information and spatial information are fused by high-low feature fusion module. The global channel-attention fusion module is used to obtain the connections between different channels and realize global fusion of different scale information. The MIoU (mean intersection over union) of MFFNet network on the PASCAL VOC 2012 and Cityscapes validation sets is 80.70% and 76.33%, respectively, achieving better segmentation results.
    Reference | Related Articles | Metrics
    Abstract321
    PDF210
    Deep Learning Compiler Load Balancing Optimization Method for Model Training
    WANG Li, GAO Kai, ZHAO Yaqian, LI Rengang, CAO Fang, GUO Zhenhua
    Journal of Frontiers of Computer Science and Technology    2024, 18 (1): 111-126.   DOI: 10.3778/j.issn.1673-9418.2209026
    For computing-intensive artificial intelligence (AI) training tasks, the computational graph is more complex, and data loading, task division of the computational graph, and load balancing of task scheduling have become the key factors affecting the computing performance. This paper proposes three optimization methods to make the task scheduling of model training in deep learning compilers reach the load balance state. Firstly, the load balance between CPU and back-end computing devices is realized by automatically establishing an efficient pipeline for data loading and model training, which improves the overall energy efficiency of the system. Secondly, the layered optimization technology of computational graph is used to realize the load balance of computational graph when the back-end devices are scheduling. Finally, this paper improves the resource utilization of back-end devices by automatically establishing efficient pipeline between layers. Experimental results show that the proposed optimization method achieves the system load balancing in the process of automatically mapping the training tasks to underlying hardware devices. Compared with traditional deep learning frameworks and compilers such as TensorFlow, nGraph, etc., this paper achieves 2%~10% performance improvement in the training of different AI models, and the overall power consumption of the training system can be reduced by more than 10%.
    Reference | Related Articles | Metrics
    Abstract315
    PDF485