Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2405.17245
Cited By
Galaxy: A Resource-Efficient Collaborative Edge AI System for In-situ Transformer Inference
27 May 2024
Shengyuan Ye
Jiangsu Du
Liekang Zeng
Wenzhong Ou
Xiaowen Chu
Yutong Lu
Xu Chen
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Galaxy: A Resource-Efficient Collaborative Edge AI System for In-situ Transformer Inference"
11 / 11 papers shown
Title
Large Language Model Partitioning for Low-Latency Inference at the Edge
Dimitrios Kafetzis
Ramin Khalili
Iordanis Koutsopoulos
24
0
0
05 May 2025
Quantifying Privacy Leakage in Split Inference via Fisher-Approximated Shannon Information Analysis
Ruijun Deng
Zhihui Lu
Qiang Duan
FedML
36
0
0
14 Apr 2025
Jupiter: Fast and Resource-Efficient Collaborative Inference of Generative LLMs on Edge Devices
Shengyuan Ye
Bei Ouyang
Liekang Zeng
Tianyi Qian
Xiaowen Chu
Jian Tang
Xu Chen
29
0
0
11 Apr 2025
PRIMA.CPP: Speeding Up 70B-Scale LLM Inference on Low-Resource Everyday Home Clusters
Zonghang Li
Tao Li
Wenjiao Feng
Mohsen Guizani
Hongfang Yu
22
0
0
07 Apr 2025
Edge Graph Intelligence: Reciprocally Empowering Edge Networks with Graph Intelligence
Liekang Zeng
Shengyuan Ye
Xu Chen
Xiaoxi Zhang
Ju Ren
Jian Tang
Yang Yang
Xuemin
Shen
52
2
0
08 Jan 2025
TPI-LLM: Serving 70B-scale LLMs Efficiently on Low-resource Edge Devices
Zonghang Li
Wenjiao Feng
Mohsen Guizani
Hongfang Yu
36
2
0
01 Oct 2024
Pluto and Charon: A Time and Memory Efficient Collaborative Edge AI Framework for Personal LLMs Fine-Tuning
Bei Ouyang
Shengyuan Ye
Liekang Zeng
Tianyi Qian
Jingyi Li
Xu Chen
47
3
0
20 Aug 2024
Asteroid: Resource-Efficient Hybrid Pipeline Parallelism for Collaborative DNN Training on Heterogeneous Edge Devices
Shengyuan Ye
Liekang Zeng
Xiaowen Chu
Guoliang Xing
Xu Chen
33
11
0
15 Aug 2024
ZeRO++: Extremely Efficient Collective Communication for Giant Model Training
Guanhua Wang
Heyang Qin
S. A. Jacobs
Connor Holmes
Samyam Rajbhandari
Olatunji Ruwase
Feng Yan
Lei Yang
Yuxiong He
VLM
55
56
0
16 Jun 2023
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
M. Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
243
1,817
0
17 Sep 2019
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
294
6,943
0
20 Apr 2018
1