RETHINKING AI LIFECYCLE MANAGEMENT IN A SERVERLESS WORLD
Keywords:
Serverless Computing, DevOps Pipelines, Artificial Intelligence, Lifecycle ManagementAbstract
As Artificial Intelligence becomes more widely used across industries, it's becoming clear that traditional DevOps approaches don’t quite fit the way AI development works. AI projects are inherently experimental and require constant iteration, frequent model updates, and access to specialized infrastructure—demands that typical operational frameworks aren't built to handle smoothly. This mismatch often creates friction in development workflows. Serverless computing offers a promising solution. With its event-driven design, automatic scaling, and pay-per-use pricing, serverless aligns naturally with the unpredictable and bursty nature of AI workloads. It can scale up during model training or inference, then scale down when not in use—making it both flexible and cost-effective. This paper takes a close look at how serverless architectures can transform the way we manage the AI lifecycle. By abstracting away infrastructure management, serverless allows data scientists to focus on building better models instead of wrestling with operational challenges. It supports smoother transitions between development, testing, and production environments, and includes helpful features like versioning, A/B testing, and progressive model rollout. The event-driven model of serverless platforms also makes it easier to build continuous training pipelines, automated evaluation systems, and real-time inference services that can adapt based on live performance data. In short, rethinking AI workflows around serverless principles can help organizations move faster, use resources more efficiently, and keep the flexibility needed to experiment and innovate in AI.
References
E. Jonas, J. Schleier-Smith, V. Sreekanti, et al., "Cloud Programming Simplified: A Berkeley View on Serverless Computing," arXiv preprint arXiv:1902.03383, 2019.
Amazon Web Services, "Serverless Architectures for Machine Learning," AWS Whitepapers, 2024. [Online]. Available: https://aws.amazon.com/whitepapers
Google Cloud, "AI Platform Documentation," Google Cloud Docs, 2024. [Online]. Available: https://cloud.google.com/ai-platform
M. Zaharia et al., "Accelerating the Machine Learning Lifecycle with MLflow," in Proc. 4th SysML Conference, 2018.
Microsoft Azure, "Serverless Compute Guide," Microsoft Docs, 2023. [Online]. Available: https://docs.microsoft.com/azure/architecture/serverless.
S. M. S. Tuli, R. Mahmud, S. Tuli and R. Buyya, "Machine Learning Model Management for Edge-AI: Challenges, Solutions, and Future Directions," IEEE Internet of Things Journal, vol. 9, no. 1, pp. 495-516, Jan. 2022.
C. Spillner, C. Matthies and A. Schill, "FAASdom for Data Science: A Serverless Platform for Reproducible Data Analytics," in Proceedings of the 2020 IEEE International Conference on Big Data (Big Data), Atlanta, GA, USA, 2020, pp. 2562–2571.
D. S. Linthicum, "Serverless Computing: Economic and Architectural Impact," IEEE Cloud Computing, vol. 5, no. 5, pp. 6-10, Sept.-Oct. 2018.
A. Arjona, E. Cabrera and P. Cordero, "Serverless Machine Learning Inference on Cloud Functions: Performance, Cost and Developer Experience," in Proceedings of the 2021 IEEE International Conference on Cloud Engineering (IC2E), San Francisco, CA, USA, 2021, pp. 172-178.
M. Biffl, P. Grunbacher and A. Mäder, "Applying DevOps Principles to Improve AI Lifecycle Management," in Proceedings of the 2020 IEEE International Conference on Software Architecture Companion (ICSA-C), Salvador, Brazil, 2020, pp. 180-183.
Downloads
Published
Issue
Section
License
Copyright (c) 2025 Sowjanya Pandruju (Author)

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.