Skip to main content
Top

Intelligent Information Processing XII

13th IFIP TC 12 International Conference, IIP 2024, Shenzhen, China, May 3–6, 2024, Proceedings, Part I

  • 2024
  • Book

About this book

The two-volume set IFIP AICT 703 and 704 constitutes the refereed conference proceedings of the 13th IFIP TC 12 International Conference on Intelligent Information Processing XII, IIP 2024, held in Shenzhen, China, during May 3–6, 2024.

The 49 full papers and 5 short papers presented in these proceedings were carefully reviewed and selected from 58 submissions.

The papers are organized in the following topical sections:

Volume I: Machine Learning; Natural Language Processing; Neural and Evolutionary Computing; Recommendation and Social Computing; Business Intelligence and Risk Control; and Pattern Recognition.

Volume II: Image Understanding.

Table of Contents

Next
  • current Page 1
  • 2
  • 3
  1. Frontmatter

  2. Machine Learning

    1. Frontmatter

    2. Dual Contrastive Learning for Anomaly Detection in Attributed Networks

      Shijie Xue, He Kong, Qi Wang
      The chapter focuses on the critical issue of anomaly detection in attributed networks, which model complex real-world scenarios by including both node interactions and rich attributes. Traditional methods often fall short in effectively identifying anomalies at different levels. The proposed Cobra method addresses this by employing a dual contrastive learning framework that considers both contextual and behavioral anomalies. By sampling subgraphs and utilizing self-supervised learning strategies, Cobra captures the intricate relationships and behaviors within the network, providing a more accurate and comprehensive evaluation of node abnormality. Extensive experiments on various datasets demonstrate the superior performance of Cobra compared to state-of-the-art methods, highlighting its potential in enhancing anomaly detection across diverse applications.
    3. Online Learning in Varying Feature Spaces with Informative Variation

      Peijia Qin, Liyan Song
      The chapter addresses the limitations of classical online learning, which assumes a constant feature space. It introduces the concept of Varying Feature Space (VFS), where features can appear and disappear over time. The research focuses on informative variations that can indicate class labels, enhancing predictive performance. The proposed approach, OVFIV, combines a sparse learner for the variation space with an ensemble method to integrate predictions from both the feature and variation streams. Experimental results demonstrate the effectiveness of this method in various datasets, highlighting its potential to significantly improve predictive models in dynamic feature spaces.
    4. Towards a Flexible Accuracy-Oriented Deep Learning Module Inference Latency Prediction Framework for Adaptive Optimization Algorithms

      Jingran Shen, Nikos Tziritas, Georgios Theodoropoulos
      The chapter discusses the challenges of deploying large deep neural networks in resource-constrained edge environments and the need for adaptive optimization algorithms. It introduces a new framework that allows flexible configurations of input parameters and automatic selection of regression models for predicting module inference latency. The proposed Multi-task Encoder-Decoder Network (MEDN) is highlighted as a more accurate and efficient alternative to existing regression models. The framework's ability to measure device dynamics and handle various input parameters, including Inferable Parameters, is emphasized. Experimental results demonstrate MEDN's superior performance and the effectiveness of the Time/Space-efficient Auto-selection algorithm. Future research directions are also outlined, focusing on further enhancing the framework's capabilities.
    5. Table Orientation Classification Model Based on BERT and TCSMN

      Dawei Jin, Rongxin Mi, Tianhang Song
      The chapter focuses on the classification of table orientations using a deep learning model that combines BERT for contextual understanding and TCSMN for capturing sequential features. The model is designed to handle the diverse structures of tables found in scientific literature, offering a more accurate and efficient method for table analysis. The authors introduce row and column-based attention mechanisms to enhance the extraction of structural semantic features, contributing to the model's high performance. Experimental results demonstrate that the proposed TableTC model outperforms traditional and deep learning baselines, showcasing its effectiveness in table classification tasks. The chapter also discusses related work and future research directions, making it a valuable resource for professionals and researchers in the field of natural language processing and data analysis.
    6. Divide-and-Conquer Strategy for Large-Scale Dynamic Bayesian Network Structure Learning

      Hui Ouyang, Cheng Chen, Ke Tang
      This chapter introduces a divide-and-conquer strategy for large-scale dynamic Bayesian network structure learning, focusing on 2 Time-sliced Bayesian Networks (2-TBNs). The method, adapted from the Partition-Estimation-Fusion (PEF) strategy used in static Bayesian networks, demonstrates significant improvements in accuracy and efficiency. By leveraging prior knowledge of 2-TBNs, the approach enhances the learning process, particularly for the transition model. Experimental validation using large-scale datasets shows that the proposed strategy outperforms baseline methods in terms of edge classification accuracy and runtime. The chapter also highlights the potential for further improvements in the partition and fusion phases, and discusses future research directions.
    7. Entropy-Based Logic Explanations of Differentiable Decision Tree

      Yuanyuan Liu, Jiajia Zhang, Yifan Li
      This chapter delves into the challenge of interpreting complex decision-making processes in deep reinforcement learning. By leveraging entropy-based logic explanations, the authors introduce a method to actively intervene in the training of differentiable decision trees, reducing parameter explosion and enhancing interpretability. Experimental results demonstrate that this approach not only maintains high performance but also achieves superior interpretability compared to baseline methods. The novelty lies in the use of entropy penalty terms and state preprocessing techniques, which steer the training process towards more explainable models. The chapter concludes with compelling experimental evidence, showcasing the effectiveness of the proposed method in multiple reinforcement learning environments.
    8. Deep Friendly Embedding Space for Clustering

      Haiwei Hou, Shifei Ding, Xiao Xu, Lili Guo
      The chapter 'Deep Friendly Embedding Space for Clustering' delves into the advancements of deep clustering methods, highlighting the limitations of traditional clustering algorithms in handling large-scale, high-dimensional datasets. It introduces a unified deep clustering algorithm that leverages autoencoders for feature extraction and dimensionality reduction, incorporating deep metric learning to enhance feature discrimination. The proposed algorithm preserves the data manifold structure, leading to improved clustering results. Experimental validation on benchmark datasets and a case study on rolling bearing fault diagnosis showcase the algorithm's superior performance and potential for industrial applications. The chapter concludes by discussing future research directions, emphasizing the importance of semi-supervised learning and pseudo-label technology in guiding neural networks to learn more suitable representations for clustering.
    9. Bayesian Personalized Sorting Based on Time Factors and Hot Recommendations

      Wenhua Zeng, Junjie Liu, Bo Zhang
      The chapter introduces a Bayesian Personalized Ranking model, BPR-TH, designed to address information overload in digital libraries. By incorporating time factors and hot recommendations, BPR-TH effectively handles massive distributed data and cold start problems, outperforming traditional BPR and File-path algorithms. The model is realized through user behavior feature extraction, model construction, and optimization, resulting in improved personalized recommendations both online and offline. Experimental results demonstrate BPR-TH's superior performance in accuracy, coverage, and recall, making it a promising solution for personalized digital library recommendations.
    10. Design and Implementation of Risk Control Model Based on Deep Ensemble Learning Algorithm

      Maoguang Wang, Ying Cui
      The chapter delves into the critical issue of credit risk in internet credit loans and proposes a groundbreaking credit risk control model based on deep ensemble learning. By building a two-layer ensemble learner, the model effectively identifies potential defaulting users, achieving an impressive F1-Score of 0.98 on the Lending Club credit dataset. This innovative approach outperforms conventional methods like logistic regression and decision trees, showcasing excellent generalization capabilities. The model's architecture, including the selection of base learners and ensemble methods, is meticulously designed to capture both general and nuanced patterns in the data. The chapter also provides a comprehensive analysis of related work, experimental results, and future research directions, making it a valuable resource for professionals seeking to enhance credit risk management strategies.
    11. More Teachers Make Greater Students: Compression of CycleGAN

      Xiaoxi Liu, Lin Lv, Ju Liu, Yanyang Han, Mengnan Liang, Xiao Jiang
      The chapter introduces the MGFD framework, designed to compress CycleGAN models effectively. By integrating an Inception-enhanced network and a multi-granularity distillation scheme, MGFD simplifies the compression process and reduces computational costs. The framework eliminates the need for a discriminator, optimizing the student generator directly through knowledge distillation. Experimental results demonstrate that MGFD outperforms existing methods in terms of computational efficiency and image quality, making it a promising solution for practical applications on mobile devices and IoT systems.
    12. Hybrid Integrated Dimensionality Reduction Method Based on Conformal Homeomorphism Mapping

      Bianping Su, Chaoyin Liang, Chunkai Wang, Yufan Guo, Shicong Wu, Yan Chen, Longqing Zhang, Jiao Peng
      This chapter introduces a hybrid integrated dimensionality reduction method that leverages conformal homeomorphism mapping to address the challenges of maintaining geometric structure and reversibility in data reduction. The method combines linear, nonlinear, and hybrid dimensionality reduction techniques to ensure that the intrinsic rigidity and geometric topology structure of the data are preserved. The chapter discusses various dimensionality reduction methods, including Principal Component Analysis (PCA), Locally Linear Embedding (LLE), and Laplacian Eigenmap (LE), highlighting their advantages and limitations. The proposed method integrates these techniques to create a robust dimensionality reduction framework that is both efficient and interpretable. The chapter also includes a detailed explanation of the conformal homeomorphism mapping process and its application to text data, demonstrating the method's effectiveness through experimental results. By reading this chapter, professionals in the field of data science and machine learning will gain valuable insights into advanced dimensionality reduction techniques and their practical applications.
  3. Natural Language Processing

    1. Frontmatter

    2. Are Mixture-of-Modality-Experts Transformers Robust to Missing Modality During Training and Inferring?

      Yan Gao, Tong Xu, Enhong Chen
      The chapter delves into the robustness of Mixture-of-Modality-Experts Transformers (MoME) when faced with missing modality during training and inference. It begins by discussing the limitations of current multi-modal Transformers and the need for models that can handle incomplete data. The authors propose a series of sub-questions to guide their research and conduct experiments to compare the robustness of MoME Transformers with vanilla Transformers. They also explore the use of multi-task learning and data imputation techniques, such as Mixup, to improve the model's performance. The chapter concludes with a novel method based on MoME Transformers and multi-task learning, which demonstrates high robustness to missing modalities with no extra computational requirements. The method is validated through experiments on three popular multi-modal datasets, showing significant improvements over existing approaches.
    3. Question Answering Systems Based on Pre-trained Language Models: Recent Progress

      Xudong Luo, Ying Luo, Binxia Yang
      The chapter delves into the pivotal role of Pre-trained Language Models (PLMs) in enhancing Question Answering Systems (QASs), emphasizing their superiority in understanding complex language patterns and providing accurate, relevant answers. It explores various PLM-based methods for information retrieval, QA performance improvement, and addressing other QA challenges. Additionally, the chapter highlights the applications of PLM-based QASs in domains such as legal, medical, and multimodal QA, showcasing their versatility and potential. The discussion includes performance evaluations and future research directions, making it a valuable resource for professionals seeking to understand the state-of-the-art in QASs.
    4. A BERT-Based Model for Legal Document Proofreading

      Jinlong Liu, Xudong Luo
      The chapter introduces a BERT-based model for legal document proofreading, addressing the critical need for accuracy and precision in legal texts. The model integrates two structurally identical MLMs with varying training methods to enhance performance. It includes a grammar check module and a spelling check module, complementing each other to correct errors effectively. The spelling check module employs a Limiter algorithm to balance precision and recall by considering pinyin similarity. The model is trained on a million artificially generated legal document sentences, demonstrating superior performance compared to baseline models and large language models. The chapter also discusses the model's architecture, training process, and experimental evaluation, highlighting its potential to revolutionize legal document proofreading.
    5. Entity Relation Joint Extraction with Data Augmentation Based on Large Language Model

      Manman Zhang, Shuocan Zhu, Jingmin Zhang, Yu Han, Xiaoxuan Zhu, Leilei Zhang
      The chapter delves into the application of large language models for entity relation extraction, focusing on data augmentation techniques to address data scarcity. It introduces two methods—Entity pairs-Dominant Data Generation (EDDG) and Relation-Dominant Data Generation (RDDG)—using ChatGPT for generating annotated data. The study also emphasizes the importance of prompt engineering strategies, such as expression diversity, length diversity, and domain diversity, to enhance model performance. Experimental results on the DuIE dataset showcase the effectiveness of these strategies, highlighting a notable increase in F1 scores across various models. The chapter concludes by validating the proposed methods and strategies, demonstrating their potential for improving entity relation extraction tasks.
Next
  • current Page 1
  • 2
  • 3
Title
Intelligent Information Processing XII
Editors
Zhongzhi Shi
Jim Torresen
Shengxiang Yang
Copyright Year
2024
Electronic ISBN
978-3-031-57808-3
Print ISBN
978-3-031-57807-6
DOI
https://doi.org/10.1007/978-3-031-57808-3

Accessibility information for this book is coming soon. We're working to make it available as quickly as possible. Thank you for your patience.

Premium Partner

    Image Credits
    Neuer Inhalt/© ITandMEDIA, Nagarro GmbH/© Nagarro GmbH, AvePoint Deutschland GmbH/© AvePoint Deutschland GmbH, AFB Gemeinnützige GmbH/© AFB Gemeinnützige GmbH, USU GmbH/© USU GmbH, Ferrari electronic AG/© Ferrari electronic AG