Improving Automated Data Annotation with Self-Supervised Learning: A Pathway to Robust AI Models

Improving Automated Data Annotation with Self-Supervised Learning: A Pathway to Robust AI Models

Authors

  • Arunkumar Thirunagalingam

Abstract

The need for large, high-quality annotated datasets has become critical in the rapidly developing field of artificial intelligence (AI). Manual labeling of data is a major component of traditional supervised learning methods, which are labor-intensive and prone to human error. Automated data annotation attempts to overcome these issues, but current methods frequently fall short in terms of accuracy and consistency. This paper investigates the incorporation of self-supervised learning (SSL) into automated data annotation processes to improve the robustness and reliability of AI models. Without the need for human intervention, SSL generates pseudo-labels by utilizing the inherent structure of data. Our proposed methodology displays considerable increases in model performance and generalization when applied to varied datasets. Experimental results reveal that SSL-based annotation not only decreases labeling costs but also boosts the robustness of AI models against noisy and missing input. This research has broad implications for various AI applications, such as natural language processing and computer vision, among others.

References

. A. Dosovitskiy, P. Fischer, J. Springenberg, M. Riedmiller, and T. Brox, “Discriminative unsupervised feature learning with exemplar convolutional neural networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 38, no. 9, pp. 1734-1747, Sep. 2016.

. T. Chen, S. Kornblith, M. Norouzi, and G. Hinton, “A simple framework for contrastive learning of visual representations,” in Proc. of the 37th International Conference on Machine Learning, 2020, pp. 1597-1607.

. A. Radford, K. Narasimhan, T. Salimans, and I. Sutskever, “Improving language understanding by generative pre-training,” arXiv preprint arXiv:1801.06146, 2018.

. Y. Bengio, A. Courville, and P. Vincent, “Representation learning: A review and new perspectives,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 35, no. 8, pp. 1798-1828, Aug. 2013.

. Z. Zhuang, Q. Li, A. Luo, Z. Tang, and J. Liu, “Self-supervised learning for large-scale medical image analysis: A survey,” arXiv preprint arXiv:2109.08682, 2021.

. L. van der Maaten and G. Hinton, “Visualizing data using t-SNE,” Journal of Machine Learning Research, vol. 9, no. 11, pp. 2579-2605, Nov. 2008.

. O. J. Hénaff, A. Razavi, C. Doersch, S. M. A. Eslami, and A. van den Oord, “Data-efficient image recognition with contrastive predictive coding,” in Proc. of the 37th International Conference on Machine Learning, 2020, pp. 4182-4192.

. X. Chen, H. Fan, R. Girshick, and K. He, “Improved baselines with momentum contrastive learning,” in Proc. of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 9729-9738.

. J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of deep bidirectional transformers for language understanding,” in Proc. of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, vol. 1, 2019, pp. 4171-4186.

. T. Mikolov, K. Chen, G. Corrado, and J. Dean, “Efficient estimation of word representations in vector space,” arXiv preprint arXiv:1301.3781, 2013.

. S. Xie, Z. Liu, J. E. Mao, and A. Yuille, “Rethinking image classification with a self-supervised balanced sampling approach,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 43, no. 6, pp. 1919-1933, Jun. 2021.

. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proc. of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 770-778.

. A. Krizhevsky, I. Sutskever, and G. Hinton, “ImageNet classification with deep convolutional neural networks,” Communications of the ACM, vol. 60, no. 6, pp. 84-90, Jun. 2017.

. D. P. Kingma and M. Welling, “Auto-encoding variational Bayes,” arXiv preprint arXiv:1312.6114, 2013.

. P. Vincent, H. Larochelle, I. Lajoie, Y. Bengio, and P.-A. Manzagol, “Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion,” Journal of Machine Learning Research, vol. 11, pp. 3371-3408, Dec. 2010.

. Z. Liu, X. Li, P. Luo, C. C. Loy, and X. Tang, “Deep learning face attributes in the wild,” in Proc. of the IEEE/CVF International Conference on Computer Vision (ICCV), 2015, pp. 3730-3738.

. G. Hinton, L. Deng, D. Yu, et al., “Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups,” IEEE Signal Processing Magazine, vol. 29, no. 6, pp. 82-97, Nov. 2012.

. I. Goodfellow, J. Pouget-Abadie, M. Mirza, et al., “Generative adversarial nets,” in Proc. of the 27th International Conference on Neural Information Processing Systems (NeurIPS), 2014, pp. 2672-2680.

. Y. Lecun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521, no. 7553, pp. 436-444, May 2015.

. A. Voulodimos, N. Doulamis, A. Doulamis, and E. Protopapadakis, “Deep learning for computer vision: A brief review,” Computational Intelligence and Neuroscience, vol. 2018, Article ID 7068349, 2018.

Nikhil Yogesh Joshi. (2022). Implementing Automated Testing Frameworks in CI/CD Pipelines: Improving Code Quality and Reducing Time to Market. International Journal on Recent and Innovation Trends in Computing and Communication, 10(6), 106–113. Retrieved from https://www.ijritcc.org/index.php/ijritcc/article/view/11166

Nikhil Yogesh Joshi. (2021). Enhancing Deployment Efficiency: A Case Study On Cloud Migration And Devops Integration For Legacy Systems. (2021). Journal Of Basic Science And Engineering, 18(1).

Downloads

Published

2023-06-30

How to Cite

Thirunagalingam , A. (2023). Improving Automated Data Annotation with Self-Supervised Learning: A Pathway to Robust AI Models. International Transactions in Artificial Intelligence, 7(7), 1–22. Retrieved from https://isjr.co.in/index.php/ITAI/article/view/266

Issue

Section

Articles
Loading...