ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2102.04887
  4. Cited By
NewsBERT: Distilling Pre-trained Language Model for Intelligent News
  Application
v1v2 (latest)

NewsBERT: Distilling Pre-trained Language Model for Intelligent News Application

Conference on Empirical Methods in Natural Language Processing (EMNLP), 2021
9 February 2021
Chuhan Wu
Fangzhao Wu
Yang Yu
Tao Qi
Yongfeng Huang
Qi Liu
    VLM
ArXiv (abs)PDFHTML

Papers citing "NewsBERT: Distilling Pre-trained Language Model for Intelligent News Application"

10 / 10 papers shown
Routing Distilled Knowledge via Mixture of LoRA Experts for Large Language Model based Bundle Generation
Routing Distilled Knowledge via Mixture of LoRA Experts for Large Language Model based Bundle Generation
Kaidong Feng
Zhu Sun
Hui Fang
Jie Yang
Wenyuan Liu
Yew-Soon Ong
MoE
96
0
0
24 Aug 2025
Revisiting Language Models in Neural News Recommender Systems
Revisiting Language Models in Neural News Recommender SystemsEuropean Conference on Information Retrieval (ECIR), 2025
Yuyue Zhao
Jin Huang
David Vos
Maarten de Rijke
KELM
796
4
0
20 Jan 2025
SCStory: Self-supervised and Continual Online Story Discovery
SCStory: Self-supervised and Continual Online Story DiscoveryThe Web Conference (WWW), 2023
Susik Yoon
Yu Meng
Dongha Lee
Jiawei Han
CLL
198
12
0
27 Nov 2023
Knowledge-aware Collaborative Filtering with Pre-trained Language Model
  for Personalized Review-based Rating Prediction
Knowledge-aware Collaborative Filtering with Pre-trained Language Model for Personalized Review-based Rating PredictionIEEE Transactions on Knowledge and Data Engineering (TKDE), 2023
Quanxiu Wang
Xinlei Cao
Jianyong Wang
Wei Zhang
VLM
85
10
0
02 Aug 2023
Gradient Knowledge Distillation for Pre-trained Language Models
Gradient Knowledge Distillation for Pre-trained Language Models
Lean Wang
Lei Li
Xu Sun
VLM
172
5
0
02 Nov 2022
User recommendation system based on MIND dataset
User recommendation system based on MIND dataset
Niran A. Abdulhussein
Ahmed J. Obaid
192
5
0
06 Sep 2022
Few-shot News Recommendation via Cross-lingual Transfer
Few-shot News Recommendation via Cross-lingual TransferThe Web Conference (WWW), 2022
Taicheng Guo
Lu Yu
B. Shihada
Xiangliang Zhang
254
10
0
28 Jul 2022
No One Left Behind: Inclusive Federated Learning over Heterogeneous
  Devices
No One Left Behind: Inclusive Federated Learning over Heterogeneous DevicesKnowledge Discovery and Data Mining (KDD), 2022
Ruixuan Liu
Fangzhao Wu
Chuhan Wu
Yanlin Wang
Lingjuan Lyu
Hong Chen
Xing Xie
FedML
217
96
0
16 Feb 2022
GateFormer: Speeding Up News Feed Recommendation with Input Gated
  Transformers
GateFormer: Speeding Up News Feed Recommendation with Input Gated Transformers
Peitian Zhang
Zheng liu
AI4TS
103
2
0
12 Jan 2022
One Teacher is Enough? Pre-trained Language Model Distillation from
  Multiple Teachers
One Teacher is Enough? Pre-trained Language Model Distillation from Multiple TeachersFindings (Findings), 2021
Chuhan Wu
Fangzhao Wu
Yongfeng Huang
182
71
0
02 Jun 2021
1
Page 1 of 1