[edit]

Scaling Machine Learning as a Service

Li Erran Li, Eric Chen, Jeremy Hermann, Pusheng Zhang, Luming Wang

; Proceedings of The 3rd International Conference on Predictive Applications and APIs, PMLR 67:14-29, 2017.

Abstract

Machine learning as a service (MLaaS) is imperative to the success of many companies as they need to gain business intelligence from big data. Building a scalable MLaaS for mission-critical and real-time applications is a very challenging problem. In this paper, we present the scalable MLaaS we built for Uber that operates globally. We focus on several scalability challenges. First, how to scale feature computation for many machine learning use cases. Second, how to build accurate models using global data and account for individual city or region characteristics. Third, how to enable scalable model deployment and real-time serving for hundreds of thousands of models across multiple data centers. Our technical solutions are the design and implementation of a scalable feature computing engine and feature store, a framework to manage and train a hierarchy of models as a single logical entity, and an automated one-click deployment system and scalable real-time serving service.

Cite this Paper

BibTeX @InProceedings{pmlr-v67-li17a, title = {Scaling Machine Learning as a Service}, author = {Li Erran Li and Eric Chen and Jeremy Hermann and Pusheng Zhang and Luming Wang}, pages = {14--29}, year = {2017}, editor = {Claire Hardgrove and Louis Dorard and Keiran Thompson and Florian Douetteau}, volume = {67}, series = {Proceedings of Machine Learning Research}, address = {Microsoft NERD, Boston, USA}, month = {11--12 Oct}, publisher = {PMLR}, pdf = {http://proceedings.mlr.press/v67/li17a/li17a.pdf}, url = {http://proceedings.mlr.press/v67/li17a.html}, abstract = {Machine learning as a service (MLaaS) is imperative to the success of many companies as they need to gain business intelligence from big data. Building a scalable MLaaS for mission-critical and real-time applications is a very challenging problem. In this paper, we present the scalable MLaaS we built for Uber that operates globally. We focus on several scalability challenges. First, how to scale feature computation for many machine learning use cases. Second, how to build accurate models using global data and account for individual city or region characteristics. Third, how to enable scalable model deployment and real-time serving for hundreds of thousands of models across multiple data centers. Our technical solutions are the design and implementation of a scalable feature computing engine and feature store, a framework to manage and train a hierarchy of models as a single logical entity, and an automated one-click deployment system and scalable real-time serving service.} } Copy to Clipboard Download

Endnote %0 Conference Paper %T Scaling Machine Learning as a Service %A Li Erran Li %A Eric Chen %A Jeremy Hermann %A Pusheng Zhang %A Luming Wang %B Proceedings of The 3rd International Conference on Predictive Applications and APIs %C Proceedings of Machine Learning Research %D 2017 %E Claire Hardgrove %E Louis Dorard %E Keiran Thompson %E Florian Douetteau %F pmlr-v67-li17a %I PMLR %J Proceedings of Machine Learning Research %P 14--29 %U http://proceedings.mlr.press %V 67 %W PMLR %X Machine learning as a service (MLaaS) is imperative to the success of many companies as they need to gain business intelligence from big data. Building a scalable MLaaS for mission-critical and real-time applications is a very challenging problem. In this paper, we present the scalable MLaaS we built for Uber that operates globally. We focus on several scalability challenges. First, how to scale feature computation for many machine learning use cases. Second, how to build accurate models using global data and account for individual city or region characteristics. Third, how to enable scalable model deployment and real-time serving for hundreds of thousands of models across multiple data centers. Our technical solutions are the design and implementation of a scalable feature computing engine and feature store, a framework to manage and train a hierarchy of models as a single logical entity, and an automated one-click deployment system and scalable real-time serving service. Copy to Clipboard Download

APA Li, L.E., Chen, E., Hermann, J., Zhang, P. & Wang, L.. (2017). Scaling Machine Learning as a Service. Proceedings of The 3rd International Conference on Predictive Applications and APIs, in PMLR 67:14-29 Copy to Clipboard Download

Related Material