PRIVACY-PRESERVING SPLIT AND COLLABORATIVE TRAINING FOR LARGE AND SMALL NEURAL ARCHITECTURES: THREATS, PROTOCOLS, AND A UNIFIED FRAMEWORK

Authors

  • Dr. A. R. Menon Institute for Secure Intelligent Systems, New Delhi University, India

Keywords:

Split learning, privacy-preserving training, multiparty computation

Abstract

This article presents a unified, publication-ready exposition of privacy-preserving collaborative learning architectures with emphasis on split-learning for one-dimensional convolutional neural networks (1D-CNNs) and split/fine-tuning paradigms applied to large language models (LLMs). Building on empirical and theoretical work in split learning (Abuadbba et al., 2020), secret-sharing primitives and randomization techniques from multiparty computation (Beaver, 1992; Chase et al., 2020), and recent findings of vulnerabilities in split-based LLM fine-tuning (Chen et al., 2024), we synthesize secure protocol building blocks and evaluate their applicability across multiple deployment scenarios including mobile LLM-enhanced applications (Chandra, 2025) and decentralized health monitoring (Sharma et al., 2024). The article details threat models, attacker capabilities, and formal desiderata for privacy-preserving collaborative training; describes how classic cryptographic primitives can be combined with split learning to mitigate identified leakage; and proposes a practical hybrid protocol that trades off computation, communication, and privacy. We also present a careful, text-based methodological narrative on measuring privacy leakage without resorting to equations, illustrate expected empirical behaviours through descriptive analysis, and discuss operational limitations and future research directions. The article aims to serve researchers and practitioners seeking to design, evaluate, and deploy collaborative learning systems that are resilient to the kinds of bidirectional attacks recently demonstrated against split fine-tuning.

References

Sharif Abuadbba, K., Kim, K., Kim, M., Thapa, C., Çamtepe, S. A., Gao, Y., Kim, H., & Nepal, S. (2020). Can we use split learning on 1D CNN models for privacy preserving training? In H.-M. Sun, S.-P. Shieh, G. Gu, & G. Ateniese (Eds.), ASIACCS 2020: 15th ACM Symposium on Information, Computer and Communications Security (pp. 305–318). ACM Press.

Almeida, T. A., Hidalgo, J. M. G., & Yamakami, A. (2011). Contributions to the study of SMS spam filtering: new collection and results. In Proceedings of the 11th ACM Symposium on Document Engineering (pp. 259–262).

Beaver, D. (1992). Efficient multiparty protocols using circuit randomization. In J. Feigenbaum (Ed.), Advances in Cryptology – CRYPTO’91 (Lecture Notes in Computer Science, vol. 576) (pp. 420–432). Springer.

Black, S., et al. (2021). GPT-Neo: Large Scale Autoregressive Language Modeling with Mesh-Tensorflow. March 2021.

Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., ... & Amodei, D. (2020). Language models are few-shot learners. Advances in Neural Information Processing Systems, 33, 1877–1901.

Chandra, R. (2025). Security and privacy testing automation for LLM-enhanced applications in mobile devices. International Journal of Networks and Security, 5(2), 30–41.

Chase, M., Ghosh, E., & Poburinnaya, O. (2020). Secret-shared shuffle. In S. Moriai & H. Wang (Eds.), Advances in Cryptology – ASIACRYPT 2020, Part III (Lecture Notes in Computer Science, vol. 12493) (pp. 342–372). Springer.

Chen, G., et al. (2024). Unveiling the vulnerability of private fine-tuning in split-based frameworks for large language models: A bidirectionally enhanced attack. Proceedings of the 2024 ACM SIGSAC Conference on Computer and Communications Security, 2904–2918. ACM.

Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. In NAACL (pp. 4171–4186).

Gangadhar, C., Chanthirasekaran, K., Chandra, K. R., Sharma, A., Thangamani, M., & Kumar, P. S. (2022). An energy efficient NOMA-based spectrum sharing techniques for cell-free massive MIMO. International Journal of Engineering Systems Modelling and Simulation, 13(4), 284–288.

Geeitha, S., & Thangamani, M. (2018). Incorporating EBO-HSIC with SVM for gene selection associated with cervical cancer classification. Journal of Medical Systems, 42(11), 225.

JYOTHI, D., Vijay, P. J., Kumar, M. K., Lakshmi, R. V., Popelo, O., Marhasova, V., ... & Kumar, D. V. (2025). Design of an improved method for intrusion detection using CNN, LSTM, and blockchain. Journal of Theoretical and Applied Information Technology, 102(1).

Kalaiselvi, B., & Thangamani, M. (2020). An efficient Pearson correlation based improved random forest classification for protein structure prediction techniques. Measurement, 162, 107885.

Kavin, P., Karki, S., Hemalatha, S., Singh, D., Vijayalakshmi, R., Thangamani, M., ... & Adigo, A. G. (2022). Machine learning‐based secure data acquisition for fake accounts detection in future mobile communication networks. Wireless Communications and Mobile Computing, 2022, Article ID 6356152.

Sharma, T., Reddy, D. N., Kaur, C., Godla, S. R., Salini, R., Gopi, A., & Baker El-Ebiary, Y. A. (2024). Federated Convolutional Neural Networks for Predictive Analysis of Traumatic Brain Injury: Advancements in Decentralized Health Monitoring. International Journal of Advanced Computer Science & Applications, 15(4).

Tom Brown et al. (2020). Language models are few-shot learners. Advances in Neural Information Processing Systems, 33, 1877–1901.

Downloads

Published

2025-10-31

How to Cite

Dr. A. R. Menon. (2025). PRIVACY-PRESERVING SPLIT AND COLLABORATIVE TRAINING FOR LARGE AND SMALL NEURAL ARCHITECTURES: THREATS, PROTOCOLS, AND A UNIFIED FRAMEWORK. Ethiopian International Journal of Multidisciplinary Research, 12(10), 860–866. Retrieved from https://eijmr.org/index.php/eijmr/article/view/3972