Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2106.01674
Cited By
JIZHI: A Fast and Cost-Effective Model-As-A-Service System for Web-Scale Online Inference at Baidu
3 June 2021
Hao Liu
Qian Gao
Jiang Li
X. Liao
Hao Xiong
Guangxing Chen
Wenlin Wang
Guobao Yang
Zhiwei Zha
Daxiang Dong
Dejing Dou
Haoyi Xiong
VLM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"JIZHI: A Fast and Cost-Effective Model-As-A-Service System for Web-Scale Online Inference at Baidu"
3 / 3 papers shown
Title
ElasticRec: A Microservice-based Model Serving Architecture Enabling Elastic Resource Scaling for Recommendation Models
Yujeong Choi
Jiin Kim
Minsoo Rhu
32
1
0
11 Jun 2024
Clover: Toward Sustainable AI with Carbon-Aware Machine Learning Inference Service
Baolin Li
S. Samsi
V. Gadepally
Devesh Tiwari
20
27
0
19 Apr 2023
Supporting Massive DLRM Inference Through Software Defined Memory
E. K. Ardestani
Changkyu Kim
Seung Jae Lee
Luoshang Pan
Valmiki Rampersad
...
Krishnakumar Nair
Maxim Naumov
Christopher Peterson
M. Smelyanskiy
Vijay Rao
BDL
31
20
0
21 Oct 2021
1