Show simple item record

dc.contributor.authorMezentsev, Gleb
dc.contributor.authorGusak, Danil
dc.contributor.authorOseledets, Ivan
dc.contributor.authorFrolov, Evgeny
dc.date.accessioned2025-06-12T20:52:14Z
dc.date.available2025-06-12T20:52:14Z
dc.date.issued2024-10-08
dc.identifier.isbn979-8-4007-0505-2
dc.identifier.urihttps://hdl.handle.net/1721.1/159401
dc.descriptionRecSys ’24, October 14–18, 2024, Bari, Italyen_US
dc.description.abstractScalability issue plays a crucial role in productionizing modern recommender systems. Even lightweight architectures may suffer from high computational overload due to intermediate calculations, limiting their practicality in real-world applications. Specifically, applying full Cross-Entropy (CE) loss often yields state-of-the-art performance in terms of recommendations quality. Still, it suffers from excessive GPU memory utilization when dealing with large item catalogs. This paper introduces a novel Scalable Cross-Entropy (SCE) loss function in the sequential learning setup. It approximates the CE loss for datasets with large-size catalogs, enhancing both time efficiency and memory usage without compromising recommendations quality. Unlike traditional negative sampling methods, our approach utilizes a selective GPU-efficient computation strategy, focusing on the most informative elements of the catalog, particularly those most likely to be false positives. This is achieved by approximating the softmax distribution over a subset of the model outputs through the maximum inner product search. Experimental results on multiple datasets demonstrate the effectiveness of SCE in reducing peak memory usage by a factor of up to 100 compared to the alternatives, retaining or even exceeding their metrics values. The proposed approach also opens new perspectives for large-scale developments in different domains, such as large language models.en_US
dc.publisherACM|18th ACM Conference on Recommender Systemsen_US
dc.relation.isversionofhttps://doi.org/10.1145/3640457.3688140en_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceAssociation for Computing Machineryen_US
dc.titleScalable Cross-Entropy Loss for Sequential Recommendations with Large Item Catalogsen_US
dc.typeArticleen_US
dc.identifier.citationMezentsev, Gleb, Gusak, Danil, Oseledets, Ivan and Frolov, Evgeny. 2024. "Scalable Cross-Entropy Loss for Sequential Recommendations with Large Item Catalogs."
dc.identifier.mitlicensePUBLISHER_POLICY
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2025-06-01T07:47:49Z
dc.language.rfc3066en
dc.rights.holderThe author(s)
dspace.date.submission2025-06-01T07:47:49Z
mit.licensePUBLISHER_POLICY
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record