Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

NASRec: Weight Sharing Neural Architecture Search for Recommender Systems (2207.07187v2)

Published 14 Jul 2022 in cs.IR and cs.LG

Abstract: The rise of deep neural networks offers new opportunities in optimizing recommender systems. However, optimizing recommender systems using deep neural networks requires delicate architecture fabrication. We propose NASRec, a paradigm that trains a single supernet and efficiently produces abundant models/sub-architectures by weight sharing. To overcome the data multi-modality and architecture heterogeneity challenges in the recommendation domain, NASRec establishes a large supernet (i.e., search space) to search the full architectures. The supernet incorporates versatile choice of operators and dense connectivity to minimize human efforts for finding priors. The scale and heterogeneity in NASRec impose several challenges, such as training inefficiency, operator-imbalance, and degraded rank correlation. We tackle these challenges by proposing single-operator any-connection sampling, operator-balancing interaction modules, and post-training fine-tuning. Our crafted models, NASRecNet, show promising results on three Click-Through Rates (CTR) prediction benchmarks, indicating that NASRec outperforms both manually designed models and existing NAS methods with state-of-the-art performance. Our work is publicly available at https://github.com/facebookresearch/NasRec.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Tunhou Zhang (16 papers)
  2. Dehua Cheng (10 papers)
  3. Yuchen He (53 papers)
  4. Zhengxing Chen (20 papers)
  5. Xiaoliang Dai (44 papers)
  6. Liang Xiong (16 papers)
  7. Feng Yan (67 papers)
  8. Hai Li (159 papers)
  9. Yiran Chen (176 papers)
  10. Wei Wen (49 papers)
Citations (13)

Summary

We haven't generated a summary for this paper yet.