Traceable Group-Wise Self-Optimizing Feature Transformation Learning: A Dual Optimization Perspective

Author:

Xiao Meng1,Wang Dongjie2,Wu Min3,Liu Kunpeng4,Xiong Hui5,Zhou Yuanchun6,Fu Yanjie7

Affiliation:

1. 1.Computer Network Information Center, Chinese Academy of Sciences, Beijing; 2.University of Chinese Academy of Sciences, Beijing, China

2. Department of Computer Science, University of Central Florida, Orlando, USA

3. Institute for Infocomm Research, Agency for Science, Singapore

4. Department of Computer Science, Portland State University, USA

5. 1.The Hong Kong University of Science and Technology (Guangzhou); 2. Guangzhou HKUST Fok Ying Tung Research Institute, China

6. Computer Network Information Center, Chinese Academy of Sciences; 2.University of Chinese Academy of Sciences, Beijing, China

7. Arizona State University, School of Computing and AI, Tempe, USA

Abstract

Feature transformation aims to reconstruct an effective representation space by mathematically refining the existing features. It serves as a pivotal approach to combat the curse of dimensionality, enhance model generalization, mitigate data sparsity, and extend the applicability of classical models. Existing research predominantly focuses on domain knowledge-based feature engineering or learning latent representations. However, these methods, while insightful, lack full automation and fail to yield a traceable and optimal representation space. An indispensable question arises: Can we concurrently address these limitations when reconstructing a feature space for a machine learning task? Our initial work took a pioneering step towards this challenge by introducing a novel self-optimizing framework. This framework leverages the power of three cascading reinforced agents to automatically select candidate features and operations for generating improved feature transformation combinations. Despite the impressive strides made, there was room for enhancing its effectiveness and generalization capability. In this extended journal version, we advance our initial work from two distinct yet interconnected perspectives: 1) We propose a refinement of the original framework, which integrates a graph-based state representation method to capture the feature interactions more effectively and develop different Q-learning strategies to alleviate Q-value overestimation further. 2) We utilize a new optimization technique (actor-critic) to train the entire self-optimizing framework in order to accelerate the model convergence and improve the feature transformation performance. Finally, to validate the improved effectiveness and generalization capability of our framework, we perform extensive experiments and conduct comprehensive analyses. These provide empirical evidence of the strides made in this journal version over the initial work, solidifying our framework’s standing as a substantial contribution to the field of automated feature transformation. To improve the reproducibility, we have released the associated code and data by the Github link https://github.com/coco11563/TKDD2023_code.

Publisher

Association for Computing Machinery (ACM)

Subject

General Computer Science

Reference38 articles.

1. Yoshua Bengio , Aaron Courville , and Pascal Vincent . 2013. Representation learning: A review and new perspectives . IEEE transactions on pattern analysis and machine intelligence 35, 8( 2013 ), 1798–1828. Yoshua Bengio, Aaron Courville, and Pascal Vincent. 2013. Representation learning: A review and new perspectives. IEEE transactions on pattern analysis and machine intelligence 35, 8(2013), 1798–1828.

2. David  M Blei , Andrew  Y Ng , and Michael  I Jordan . 2003. Latent dirichlet allocation. the Journal of machine Learning research 3 ( 2003 ), 993–1022. David M Blei, Andrew Y Ng, and Michael I Jordan. 2003. Latent dirichlet allocation. the Journal of machine Learning research 3 (2003), 993–1022.

3. Emmanuel  J Candès , Xiaodong Li , Yi Ma , and John Wright . 2011. Robust principal component analysis?Journal of the ACM (JACM) 58, 3 ( 2011 ), 1–37. Emmanuel J Candès, Xiaodong Li, Yi Ma, and John Wright. 2011. Robust principal component analysis?Journal of the ACM (JACM) 58, 3 (2011), 1–37.

4. Neural Feature Search: A Neural Architecture for Automated Feature Engineering

5. Techniques for Automated Machine Learning

Cited by 1 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. Resolving the Imbalance Issue in Hierarchical Disciplinary Topic Inference via LLM-based Data Augmentation;2023 IEEE International Conference on Data Mining Workshops (ICDMW);2023-12-04

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3