ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1812.01776
  4. Cited By
InferLine: ML Prediction Pipeline Provisioning and Management for Tight
  Latency Objectives
v1v2 (latest)

InferLine: ML Prediction Pipeline Provisioning and Management for Tight Latency Objectives

5 December 2018
D. Crankshaw
Gur-Eyal Sela
Corey Zumar
Xiangxi Mo
Joseph E. Gonzalez
Ion Stoica
Alexey Tumanov
ArXiv (abs)PDFHTML

Papers citing "InferLine: ML Prediction Pipeline Provisioning and Management for Tight Latency Objectives"

11 / 11 papers shown
Title
SuperServe: Fine-Grained Inference Serving for Unpredictable Workloads
SuperServe: Fine-Grained Inference Serving for Unpredictable Workloads
Alind Khare
Dhruv Garg
Sukrit Kalra
Snigdha Grandhi
Ion Stoica
Alexey Tumanov
96
11
0
27 Dec 2023
Cocktail: Leveraging Ensemble Learning for Optimized Model Serving in
  Public Cloud
Cocktail: Leveraging Ensemble Learning for Optimized Model Serving in Public Cloud
Jashwant Raj Gunasekaran
Cyan Subhra Mishra
P. Thinakaran
M. Kandemir
Chita R. Das
89
3
0
09 Jun 2021
AutonoML: Towards an Integrated Framework for Autonomous Machine
  Learning
AutonoML: Towards an Integrated Framework for Autonomous Machine Learning
D. Kedziora
Katarzyna Musial
Bogdan Gabrys
164
18
0
23 Dec 2020
Online Learning Demands in Max-min Fairness
Online Learning Demands in Max-min Fairness
Kirthevasan Kandasamy
Gur-Eyal Sela
Joseph E. Gonzalez
Sai Li
Ion Stoica
FaML
124
16
0
15 Dec 2020
MicroRec: Efficient Recommendation Inference by Hardware and Data
  Structure Solutions
MicroRec: Efficient Recommendation Inference by Hardware and Data Structure Solutions
Wenqi Jiang
Zhen He
Shuai Zhang
Thomas B. Preußer
Kai Zeng
...
Tongxuan Liu
Yong Li
Jingren Zhou
Ce Zhang
Gustavo Alonso
84
7
0
12 Oct 2020
A Tensor Compiler for Unified Machine Learning Prediction Serving
A Tensor Compiler for Unified Machine Learning Prediction ServingUSENIX Symposium on Operating Systems Design and Implementation (OSDI), 2020
Supun Nakandala Karla Saur
Karla Saur
Gyeong-In Yu
Konstantinos Karanasos
Carlo Curino
Markus Weimer
Matteo Interlandi
230
55
0
09 Oct 2020
MeDaS: An open-source platform as service to help break the walls
  between medicine and informatics
MeDaS: An open-source platform as service to help break the walls between medicine and informatics
Liang Zhang
Johann Li
Ping Li
Xiaoyuan Lu
Peiyi Shen
Guangming Zhu
Syed Afaq Ali Shah
Bennamoun
Kun Qian
Björn W. Schuller
MedIm
144
6
0
12 Jul 2020
Optimizing Prediction Serving on Low-Latency Serverless Dataflow
Optimizing Prediction Serving on Low-Latency Serverless Dataflow
Vikram Sreekanti
Harikaran Subbaraj
Chenggang Wu
Joseph E. Gonzalez
J. M. Hellerstein
79
25
0
11 Jul 2020
Orchestrating the Development Lifecycle of Machine Learning-Based IoT
  Applications: A Taxonomy and Survey
Orchestrating the Development Lifecycle of Machine Learning-Based IoT Applications: A Taxonomy and Survey
Bin Qian
Jie Su
Z. Wen
D. N. Jha
Yinhao Li
...
Albert Y. Zomaya
Omer F. Rana
Lizhe Wang
Maciej Koutny
R. Ranjan
165
4
0
11 Oct 2019
Parity Models: A General Framework for Coding-Based Resilience in ML
  Inference
Parity Models: A General Framework for Coding-Based Resilience in ML Inference
J. Kosaian
K. V. Rashmi
Shivaram Venkataraman
168
14
0
02 May 2019
The OoO VLIW JIT Compiler for GPU Inference
The OoO VLIW JIT Compiler for GPU Inference
Paras Jain
Xiangxi Mo
Ajay Jain
Alexey Tumanov
Joseph E. Gonzalez
Ion Stoica
147
18
0
28 Jan 2019
1