-
Notifications
You must be signed in to change notification settings - Fork 120
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[RMP] Provide to our customers best practices guidance for training Retrieval, Ranking and Multi-Stage RecSys models #553
Comments
@gabrielspmoreira @viswa-nvidia to update checkboxes and convert to issues across the releases. |
fyi, I have updated the check boxes and created the tasks a few days ago |
Created some slides with an overview of the proposal to discuss in the next grooming meeting |
I have written a new RMP #732 as a rewrite of this one in a more customer-centric and pragmatic approach: a quick start pipeline for customers. I keep this RMP unchanged as it was prioritized, but propose closing it and taking the #732 as the replacement. The new one also reduces the scope in terms of dataset experiments, as we are going to stick initially to a single dataset (TenRec) as an example on how to use the quick-start template pipeline. |
@viswa-nvidia @EvenOldridge I am closing this ticket, as I rewrote it in a customer-centric approach in #732 , which is now assigned to 23.01 |
Problem:
Merlin platform provides tools for building multi-stage recommender systems (which we are going to present at RecSys’22 conference on our tutorial and demo).
In particular, the retrieval and ranking stages are implemented in Merlin Models library and tied together during inference with Merlin Systems library.
Although we provide to our customers an implementation of retrieval and ranking models, and a few example notebooks with toy datasets to get them started, it might be hard for our customers to obtain a reasonable accuracy and performance without too much experimentation on their side with their datasets. In addition, it is expected that customers find some issues in their experiments related to our API flexibility and to models accuracy/performance with real datasets.
This scenario might reduce customers’ interest and engagement if they think Merlin has not been well experimented and refined with real datasets and is not mature enough for their purposes.
Goal:
The goal of this work is to improve customers experience when starting to experiment Merlin with their own datasets.
We want to leverage NVIDIA internal research and computational resources to perform comprehensive experimentation of our retrieval and ranking models with a diversity of public datasets, so that we can:
Constraints:
Starting Point:
Our research team has already done or is doing experimentation work for some RecSys use cases:
The effort in the retrieval research is tightly related to the scope of this RMP.
We have been refining/fixing and experimenting with a number of ranking models (like DCN-v2, DLRM, DeepFM, Wide&Deep) and MTL ranking models (e.g. MMoE, PLE) for AliCCP and Tenrec public dataset. That research work is directly related to this RMP.
Tasks
Ranking Models
PredictionBlock
instead ofPredictionTask
) #917AliCCP dataset experiments
Tenrec dataset experiments
Testing
Documentation
Retrieval Models
Finish the refactory of the YouTubeDNN retrieval model
- [x] NVIDIA-Merlin/models#540 - Provide flexibility to define the targets of YouTube DNN sequential input (e.g. next-item prediction x last item prediction)
- [x] NVIDIA-Merlin/models#622 - Make YouTubeDNN and retrieval model, that exports a Top-K recommender model for evaluation and inference
Datasets preprocessing
Hypertuning
Documentation
Testing
Two-stage recommendation
Investigate/research how to better train two-stage recsys pipelines
The text was updated successfully, but these errors were encountered: