Movatterモバイル変換


[0]ホーム

URL:


Skip to main content
Cornell University
We gratefully acknowledge support from the Simons Foundation,member institutions, and all contributors.Donate
arxiv logo>cs> arXiv:2502.17494
arXiv logo
Cornell University Logo

Computer Science > Information Retrieval

arXiv:2502.17494 (cs)
[Submitted on 20 Feb 2025 (v1), last revised 3 Mar 2025 (this version, v4)]

Title:External Large Foundation Model: How to Efficiently Serve Trillions of Parameters for Online Ads Recommendation

Authors:Mingfu Liang,Xi Liu,Rong Jin,Boyang Liu,Qiuling Suo,Qinghai Zhou,Song Zhou,Laming Chen,Hua Zheng,Zhiyuan Li,Shali Jiang,Jiyan Yang,Xiaozhen Xia,Fan Yang,Yasmine Badr,Ellie Wen,Shuyu Xu,Hansey Chen,Zhengyu Zhang,Jade Nie,Chunzhi Yang,Zhichen Zeng,Weilin Zhang,Xingliang Huang,Qianru Li,Shiquan Wang,Evelyn Lyu,Wenjing Lu,Rui Zhang,Wenjun Wang,Jason Rudy,Mengyue Hang,Kai Wang,Yinbin Ma,Shuaiwen Wang,Sihan Zeng,Tongyi Tang,Xiaohan Wei,Longhao Jin,Jamey Zhang,Marcus Chen,Jiayi Zhang,Angie Huang,Chi Zhang,Zhengli Zhao,Jared Yang,Qiang Jin,Xian Chen,Amit Anand Amlesahwaram,Lexi Song,Liang Luo,Yuchen Hao,Nan Xiao,Yavuz Yetim,Luoshang Pan,Gaoxiang Liu,Yuxi Hu,Yuzhen Huang,Jackie Xu,Rich Zhu,Xin Zhang,Yiqun Liu,Hang Yin,Yuxin Chen,Buyun Zhang,Xiaoyi Liu,Xingyuan Wang,Wenguang Mao,Zhijing Li,Qin Huang,Chonglin Sun,Nancy Yu,Shuo Gu,Shupin Mao,Benjamin Au,Jingzheng Qin,Peggy Yao,Jae-Woo Choi,Bin Gao,Ernest Wang,Lei Zhang,Wen-Yen Chen,Ted Lee,Jay Zha,Yi Meng,Alex Gong,Edison Gao,Alireza Vahdatpour,Yiping Han,Yantao Yao,Toshinari Kureha,Shuo Chang,Musharaf Sultan,John Bocharov,Sagar Chordia,Xiaorui Gan,Peng Sun,Rocky Liu,Bo Long,Wenlin Chen et al. (2 additional authors not shown)
View PDFHTML (experimental)
Abstract:Ads recommendation is a prominent service of online advertising systems and has been actively studied. Recent studies indicate that scaling-up and advanced design of the recommendation model can bring significant performance improvement. However, with a larger model scale, such prior studies have a significantly increasing gap from industry as they often neglect two fundamental challenges in industrial-scale applications. First, training and inference budgets are restricted for the model to be served, exceeding which may incur latency and impair user experience. Second, large-volume data arrive in a streaming mode with data distributions dynamically shifting, as new users/ads join and existing users/ads leave the system. We propose the External Large Foundation Model (ExFM) framework to address the overlooked challenges. Specifically, we develop external distillation and a data augmentation system (DAS) to control the computational cost of training/inference while maintaining high performance. We design the teacher in a way like a foundation model (FM) that can serve multiple students as vertical models (VMs) to amortize its building cost. We propose Auxiliary Head and Student Adapter to mitigate the data distribution gap between FM and VMs caused by the streaming data issue. Comprehensive experiments on internal industrial-scale applications and public datasets demonstrate significant performance gain by ExFM.
Comments:Accepted by the ACM Web Conference (WWW) 2025 Industrial Track as Oral Presentation
Subjects:Information Retrieval (cs.IR); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
Cite as:arXiv:2502.17494 [cs.IR]
 (orarXiv:2502.17494v4 [cs.IR] for this version)
 https://doi.org/10.48550/arXiv.2502.17494
arXiv-issued DOI via DataCite

Submission history

From: Mingfu Liang [view email]
[v1] Thu, 20 Feb 2025 22:35:52 UTC (1,057 KB)
[v2] Wed, 26 Feb 2025 05:29:28 UTC (1,057 KB)
[v3] Thu, 27 Feb 2025 23:32:37 UTC (1,057 KB)
[v4] Mon, 3 Mar 2025 22:21:09 UTC (1,057 KB)
Full-text links:

Access Paper:

Current browse context:
cs.IR
Change to browse by:
export BibTeX citation

Bookmark

BibSonomy logoReddit logo

Bibliographic and Citation Tools

Bibliographic Explorer(What is the Explorer?)
Connected Papers(What is Connected Papers?)
scite Smart Citations(What are Smart Citations?)

Code, Data and Media Associated with this Article

CatalyzeX Code Finder for Papers(What is CatalyzeX?)
Hugging Face(What is Huggingface?)
Papers with Code(What is Papers with Code?)

Demos

Hugging Face Spaces(What is Spaces?)

Recommenders and Search Tools

Influence Flower(What are Influence Flowers?)
CORE Recommender(What is CORE?)

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community?Learn more about arXivLabs.

Which authors of this paper are endorsers? |Disable MathJax (What is MathJax?)

[8]ページ先頭

©2009-2025 Movatter.jp