Vol. 3 No. 2 (2023): Journal of AI-Assisted Scientific Discovery
Articles

The Dialectics of Unsupervised Learning: A Synthesis of Anomaly Detection Methodologies

Dr. Yu Han
Associate Professor of Computer Science, Shanghai Jiao Tong University, China
Cover

Published 30-12-2023

How to Cite

[1]
Dr. Yu Han, “The Dialectics of Unsupervised Learning: A Synthesis of Anomaly Detection Methodologies”, Journal of AI-Assisted Scientific Discovery, vol. 3, no. 2, pp. 261–283, Dec. 2023, Accessed: Nov. 21, 2024. [Online]. Available: https://scienceacadpress.com/index.php/jaasd/article/view/105

Abstract

The most frequently used category of learning is by far supervised learning. Supervised learning has become a field of applied research everywhere where computation facilities are available. So why do we care about unsupervised learning? Can unsupervised learning do what supervised learning can do (classification) and does every learned classifier have a supervised learning's unsupervised counterpart? The answer to this last question concerns the family of the product of base classifiers, namely the ensemble classifier. Unlike classifiers learned by other supervised learning methodologies, ensemble classifiers do not always have a ready-made unsupervised counterpart. This is so because they differ from models learned by other supervised methodologies in that their learning objective is indirect training error minimization. Unsupervised learning, while often overshadowed by the prominence of supervised learning, offers unique advantages and capabilities that cannot be ignored. It enables us to venture into uncharted territories and uncover hidden patterns and structures within data without the need for explicit labels or guidance. By autonomously exploring and analyzing the intrinsic properties of the data, unsupervised learning opens up a whole new dimension of possibilities. In the realm of classification tasks, supervised learning has long held the crown as the go-to approach. However, this does not negate the potential of unsupervised learning to tackle classification challenges. While unsupervised learning may not directly mirror the classification capabilities of supervised learning, it possesses the ability to discover underlying patterns and groupings that can indirectly contribute to classification tasks. Think of unsupervised learning as a powerful detective, tirelessly working to uncover the subtle nuances and relationships within the data, ultimately aiding in the classification process. Now, let's address the query regarding the existence of unsupervised counterparts for every learned classifier in supervised learning. While it would be ideal for every supervised learning classifier to have a neatly packaged unsupervised counterpart, this is not always the case. Enter the ensemble classifier, a family of base classifiers that combine their collective wisdom to make highly accurate predictions. Unlike classifiers learned through traditional supervised methodologies, ensemble classifiers have a distinct learning objective: the minimization of training error through an indirect approach. This distinction in learning objectives is what sets ensemble classifiers apart and makes the creation of a ready-made unsupervised counterpart more challenging. The intricate nature of ensemble learning necessitates a specialized unsupervised counterpart that can effectively capture the collaborative wisdom of the base classifiers. Despite this hurdle, researchers continue to push the boundaries of unsupervised methods in the quest for adaptable and robust ensemble classifiers that can empower diverse applications. In conclusion, while supervised learning reigns supreme in popularity, the importance of unsupervised learning should not be underestimated. From unlocking hidden insights to indirectly aiding in classification tasks, unsupervised learning offers a rich landscape of possibilities. Although every supervised learning classifier may not have a readily available unsupervised counterpart, the ensemble classifier stands as a testament to the evolving nature of learning methodologies. As the field progresses, the synergy between supervised and unsupervised learning continues to pave the way for exciting advancements in the realm of machine learning.

Downloads

Download data is not yet available.

References

  1. C. C. Aggarwal, "Outlier Analysis," Springer, 2017.
  2. M. Goldstein and S. Uchida, "A Comparative Evaluation of Unsupervised Anomaly Detection Algorithms for Multivariate Data," PLoS ONE, vol. 11, no. 4, pp. 1-31, Apr. 2016.
  3. Tatineni, Sumanth. "Embedding AI Logic and Cyber Security into Field and Cloud Edge Gateways." International Journal of Science and Research (IJSR) 12.10 (2023): 1221-1227.
  4. Vemori, Vamsi. "Towards a Driverless Future: A Multi-Pronged Approach to Enabling Widespread Adoption of Autonomous Vehicles-Infrastructure Development, Regulatory Frameworks, and Public Acceptance Strategies." Blockchain Technology and Distributed Systems 2.2 (2022): 35-59.
  5. Tatineni, Sumanth. "Addressing Privacy and Security Concerns Associated with the Increased Use of IoT Technologies in the US Healthcare Industry." Technix International Journal for Engineering Research (TIJER) 10.10 (2023): 523-534.
  6. Z. Zong, H. Chen, M. Shi, J. Wu, Y. Wang, and W. Zheng, "Anomaly Detection in Time Series of Images Using Deep Learning," IEEE Access, vol. 7, pp. 67756-67768, 2019.
  7. K. Chalapathy and S. Chawla, "Deep Learning for Anomaly Detection: A Survey," arXiv preprint arXiv:1901.03407, 2019.
  8. X. Xu, H. Wang, B. Li, and X. Zhang, "Anomaly Detection in Multimedia Data Through Multimodal Deep Learning," IEEE Transactions on Neural Networks and Learning Systems, vol. 32, no. 1, pp. 191-204, Jan. 2021.
  9. Y. Zhao and M. Comiter, "A Survey on Anomaly Detection Using Data Mining Techniques," ACM SIGKDD Explorations Newsletter, vol. 15, no. 1, pp. 1-14, Jun. 2014.
  10. M. Kloft and P. Laskov, "Online Anomaly Detection Under Adversarial Impact," in Proc. 13th International Conference on Artificial Intelligence and Statistics, Chia Laguna Resort, Sardinia, Italy, 2010, pp. 405-412.
  11. R. Chalapathy, A. K. Menon, and S. Chawla, "Anomaly Detection Using One-Class Neural Networks," arXiv preprint arXiv:1802.06360, 2018.
  12. S. Liu, T. Cheng, and Y. Wang, "Anomaly Detection in Dynamic Graphs Using Attention-based Temporal Graph Neural Networks," in Proc. 2020 IEEE International Conference on Data Mining (ICDM), Sorrento, Italy, 2020, pp. 987-996.
  13. A. Ahmed and A. Pothen, "Anomaly Detection in Streaming Sensor Data," IEEE Transactions on Knowledge and Data Engineering, vol. 30, no. 3, pp. 545-558, Mar. 2018.
  14. S. Eskin, "Anomaly Detection Over Noisy Data Using Ensemble Methods," in Proc. 17th International Conference on Machine Learning, Stanford, CA, 2000, pp. 255-262.
  15. J. Zhuang, J. Tang, and H. Wang, "Local Outlier Detection with Interpretable Rules," in Proc. 2019 SIAM International Conference on Data Mining, Calgary, AB, Canada, 2019, pp. 237-245.
  16. L. Ruff, R. A. Vandermeulen, N. Görnitz, L. Deecke, S. A. Siddiqui, A. Binder, E. Müller, and M. Kloft, "Deep One-Class Classification," in Proc. 35th International Conference on Machine Learning (ICML), Stockholm, Sweden, 2018, pp. 4390-4399.
  17. E. Eskin, A. Arnold, M. Prerau, L. Portnoy, and S. Stolfo, "A Geometric Framework for Unsupervised Anomaly Detection: Detecting Intrusions in Unlabeled Data," in Proc. 2002 Data Mining for Security Applications Workshop, Arlington, VA, 2002, pp. 77-101.
  18. R. Chalapathy and S. Chawla, "Anomalous: Deep Anomaly Detection in Time Series," arXiv preprint arXiv:1901.07329, 2019.
  19. D. G. Akila and B. Tamilarasi, "Anomaly Detection in Network Traffic Using K-Means Clustering," in Proc. 2018 International Conference on Computing, Power and Communication Technologies (GUCON), Greater Noida, India, 2018, pp. 169-173.
  20. X. Li, X. Song, M. Wu, X. Li, C. Wu, and X. Hu, "Anomaly Detection in Multivariate Data With Graph Learning-Based Approaches," IEEE Transactions on Knowledge and Data Engineering, vol. 33, no. 6, pp. 2454-2465, Jun. 2021.
  21. M. Pang, Y. Zhou, J. Gao, and B. Wu, "Deep Autoencoder-Based Anomaly Detection in High-Dimensional Data," IEEE Access, vol. 7, pp. 28218-28229, 2019.