SlideShare a Scribd company logo
1 of 27
Download to read offline
Local Collaborative Autoencoders
Minjin Choi1, Yoonki Jeong1, Joonseok Lee2, Jongwuk Lee1
Sungkyunkwan University (SKKU), South Korea1
Google Research, United States2
Motivation
2
Global Low-rank Assumption
➢Existing models are based on the global low-lank assumption.
• All users and items share the same latent features.
➢Limitation: some users/items may have different latent features.
3
1 1 ? ? 1
1 1 1 ? ?
1 1 ? 1 ?
? 1 1 1 ?
? 1 1 ? 1
0.5 ... 0.1
0.6 ... 0.2
0.7 ... 0.3
… ... ...
0.1 ... 0.7
0.2 0.4 1.2 … 2.3
… ... … ... ...
0.4 0.6 0.3 … 1.5
k features
k
features
Local Low-rank Assumption
➢A user-item matrix can be divided to several sub-matrices
with the local low-rank assumption.
• Each sub-matrix represents different communities.
• Local models represent various communities with different characteristics.
4
1 1 ? ? 1
1 1 1 ? ?
1 1 ? 1 ?
? 1 1 1 ?
? 1 1 ? 1
Limitation of Existing Local Models
➢If the local model is too large, it is close to the global model.
• Because LLORMA uses large local models, the local model may not
represent its unique characteristic.
• The performance gain may come from an ensemble effect.
5
1 1 ? ? 1
1 1 1 ? ?
1 1 ? 1 ?
? 1 1 1 ?
? 1 1 ? 1
1 1 ? ? 1
1 1 1 ? ?
1 1 ? 1 ?
? 1 1 1 ?
? 1 1 ? 1
Lee et al., "LLORMA: Local Low-Rank Matrix Approximation," JMLR 2016
Limitation of Existing Local Models
➢If the local model is too small, the accuracy is too low.
• Because sGLSVD uses small local models, some local model may have
insufficient training data.
6
1 1 ? ? 1
1 1 1 ? ?
1 1 ? 1 ?
? 1 1 1 ?
? 1 1 ? 1
Evangelia Christakopoulou and George Karypis, “Local Latent Space Models for Top-N Recommendation,” KDD 2018
1 1 ? ? 1
1 1 1 ? ?
1 1 ? 1 ?
? 1 1 1 ?
? 1 1 ? 1
Research Question
7
How to build coherent and accurate local models?
Our Key Contributions
➢When the local model keeps small and coherent,
we build the local model with a relatively large training data.
8
(Dataset: ML10M)
Training matrix
== Test matrix
training matrix
> test matrix
Our Key Contributions
➢Autoencoder-based models are used as the base model
to train the local model.
• They are useful for capturing non-linear and complicated patterns.
9
1 1 ? ? 1
1 1 1 ? ?
1 1 ? 1 ?
? 1 1 1 ?
? 1 1 ? 1
Sedhain et al., "AutoRec: Autoencoders Meet Collaborative Filtering," WWW 2015
2 1 3
…
2 1 3
…
…
𝒉(𝒓)
𝒓
ො
𝒓
𝑾, 𝒃
𝑾′
, 𝒃′
Autoencoder-based models take
each row as input.
Proposed Method
10
Local Collaborative Autoencoders (LOCA)
➢Overall architecture of LOCA
• Step 1: Discovering two local communities for an anchor user
• Step 2: Training a local model with an expanded community
• Step 3: Combining multiple local models
11
A local community
for an anchor user
An expanded local
community
Local model q
…
Final model
Local model 2
Local model 1
Step 1: Discovering Local Communities
➢For an anchor user, determine a local community
and expand the local community for training.
12
1.0
0.6
0.3
0.5
0.4
0.7
0.4
0.2
Calculate the similarities between the
anchor user and the other users.
…
…
1.0
0.7
0.6
1.0
0.7
0.6
0.5
0.5
0.4
0.4
Neighbors for the
local community
Expanded neighbors to
train the local community
Step 2: Training a Local Model
➢Train the local model with the expanded community.
• Use the autoencoder-based model for training the local model.
• Note: It is possible to utilize any base models, e.g., MF, AE and EASER.
• The similarities with the anchor are used for the user weights for training.
13
1.0
0.6
0.3
0.5
0.4
0.7
0.4
1 1 ? 1
1 1 1 1
1 1 ? ?
1 1 ? 1
1 ? ? ?
1 1 1 1
1 1 ? ?
…
…
…
…
…
…
…
…
Harald Steck., "Embarrassingly shallow autoencoders for sparse data,” WWW 2019.
Step 3: Combining Multiple Local Models
➢Aggregate multiple local models with the weight of
each local model.
14
Local model 1
Local model q
…
1 1 ? 1
1 1 1 1
1 1 ? ?
1 1 ? 1
1 ? ? ?
1 1 1 1
1 1 ? ?
? 1 1 ?
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
.9 .4 .5 .7
.8 .9 .6 .6
.7 .8 .5 .8
…
.7 .4 .5 .9
.7 .5 .5 .8
.8 .6 .7 .8
…
…
…
.9 .4 .5 .7
.8 .9 .6 .6
.7 .8 .5 .8
.8 .7 .5 1
.7 .4 .5 .9
.7 .5 .5 .8
.8 .6 .7 .8
.6 1 1 .5
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
Training Local Models in Detail
➢The loss function for the 𝑗-th local model
➢Aggregating all local models and a global model
15
argmin
𝜽(𝒋)
෍
𝒓𝒖∈𝑹
𝒕𝒖
(𝒋)
ℒ 𝒓𝒖, 𝑴𝒍𝒐𝒄𝒂𝒍
(𝒓𝒖; 𝜽(𝒋)
) + 𝝀𝛀(𝜽(𝒋)
)
෡
𝑹 = 𝜶𝑴𝒈𝒍𝒐𝒃𝒂𝒍 𝑹; 𝜽 𝒈 + 𝟏 − 𝜶 ෍
𝒋=𝟏
𝒒
𝒘(𝒋) ⊙ 𝑴𝒍𝒐𝒄𝒂𝒍(𝑹; 𝜽(𝒋)) ⊘ 𝒘
Parameter for the
𝒋-th local model
The 𝒋-th local model
User weight for training the 𝒋-th local model
User weight for inferring the 𝒋-th local model
The global model is used to handle the users
who are not covered by local models.
How to Choose Anchor Users
➢Random selection: Choose 𝑘 anchor users at random.
➢Need to maximize the coverage of users by local models.
• An optimal maximum coverage algorithm is the NP-hard problem.
➢Use the greedy method to maximize the coverage.
• Select the anchor user who has the most uncovered users iteratively.
16
Group1
Group 2
Group 3
Experiments
17
Experimental Setup: Dataset
➢We evaluate our model over five public datasets
with various characteristics (e.g., domain, sparsity).
18
Dataset # of users # of items # of ratings Sparsity
MovieLens 10M
(ML10M)
69,878 10,677 10,000,054 98.66%
MovieLens 20M
(ML20M)
138,493 26,744 20,000,263 99.46%
Amazon Music
(AMusic)
4,964 11,797 97,439 99.83%
Amazon Game
(AGame)
13,063 17,408 236,415 99.90%
Yelp 25,677 25,815 731,671 99.89%
Evaluation Protocol and Metrics
➢Evaluation protocol: leave-5-out
• Hold-out the last 5 interactions as the test data for each user.
➢Evaluation metrics
• Recall@100
• Measures the number of test items included in the top-N list.
• NDCG@100
• Measures the ranking of test items in the top-N list.
19
Time
Training data:
Test data:
User interaction:
Competitive Global/Local Models
➢Four autoencoder-based global models
• CDAE: a denoising autoencoder-based model with a latent user vector
• MultVAE: a VAE-based model
• RecVAE: a VAE-based model by improving MultVAE
• EASER: an item-to-item latent factor model
➢Two local models
• LLORMA: local model using MF as the base model
• sGLSVD: local model using SVD as the base model
20
Yao Wu et al., “Collaborative Denoising Auto-Encoders for Top-N Recommender Systems,” WSDM 2016
Dawen Liang et al., "Variational autoencoders for collaborative filtering,” WWW 2018.
Ilya Shenbin et al., "RecVAE: A new variational autoencoder for Top-N recommendations with implicit feedback,” WSDM 2020.
Harald Steck., "Embarrassingly shallow autoencoders for sparse data,” WWW 2019.
Lee et al., "LLORMA: Local Low-Rank Matrix Approximation," JMLR 2016
Evangelia Christakopoulou and George Karypis, “Local Latent Space Models for Top-N Recommendation,” KDD 2018
Accuracy: LOCA vs. Competing Models
21
Dataset Metric CDAE MultVAE EASER RecVAE LLORMA sGLSVD LOCAVAE LOCAEASE
ML10M
Recall@100 0.4685 0.4653 0.4648 0.4705 0.4692 0.4468 0.4865 0.4798
NDCG@100 0.1982 0.1945 0.2000 0.1996 0.2042 0.1953 0.2073 0.2049
ML20M
Recall@100 0.4324 0.4397 0.4468 0.4417 0.3355 0.4342 0.4419 0.4654
NDCG@100 0.1844 0.1860 0.1948 0.1857 0.1446 0.1919 0.1884 0.2024
Amusic
Recall@100 0.0588 0.0681 0.0717 0.0582 0.0517 0.0515 0.0748 0.0717
NDCG@100 0.712 0.0822 0.0821 0.0810 0.0638 0.0613 0.0893 0.0826
Agames
Recall@100 0.1825 0.2081 0.1913 0.1920 0.1223 0.1669 0.2147 0.1947
NDCG@100 0.0808 0.0920 0.0915 0.0849 0.0539 0.0777 0.0966 0.0922
Yelp
Recall@100 0.2094 0.2276 0.2187 0.2262 0.1013 0.1965 0.2354 0.2205
NDCG@100 0.0920 0.0982 0.0972 0.0975 0.0429 0.0857 0.1103 0.0981
Global Models Local Models Ours
➢LOCA consistently outperforms competitive
global/local models over five benchmark datasets.
Effect of the Number of Local Models
➢The accuracy of LOCA improved consistently with
an increase in the number of local models.
22
0.185
0.19
0.195
0.2
0.205
0.21
0 50 100 150 200 250 300
NDCG@100
The number of local models
ML10M
MultVAE Ensemble_VAE
LLORMA_VAE LOCA_VAE
0.078
0.08
0.082
0.084
0.086
0.088
0.09
0 50 100 150 200 250 300
NDCG@100 The number of local models
AMusic
MultVAE Ensemble_VAE
LLORMA_VAE LOCA_VAE
Effect of Anchor Selection Method
➢Our coverage-based anchor selection outperforms
the other methods in terms of accuracies and coverages.
23
0.196
0.198
0.2
0.202
0.204
0.206
0.208
50 100 150 200 250 300
NDCG@100
The number of local models
ML10M
Random K-means Ours
0.08
0.082
0.084
0.086
0.088
0.09
50 100 150 200 250 300
NDCG@100 The number of local models
AMusic
Random K-means Ours
Illustration of LOCA
➢When a user has multiple tastes, LOCA can capture
the user preference by combining different local patterns.
• For a user (66005 in ML10M) who likes Sci-Fi and Horror movies, LOCA
shows a better accuracy.
24
Recommendation Local 70 Local 179 Global Ground Truth
Top-1
Sci-Fi
Adventure
Horror
Action
Thriller
Action
Sci-Fi
Action
Top-2
Sci-Fi
Horror
Horror
Drama
Drama
Horror
Thriller
Top-3
Sci-Fi
Action
Horror
Drama
Drama
Mystery
Horror
Action
Conclusion
25
Conclusion
➢We propose a new local recommender framework,
namely local collaborative autoencoders (LOCA).
➢LOCA can handle a large number of local models effectively.
• Adopts a local model with different training/inference strategies.
• Utilizes autoencoder-based model as the base model.
• Makes use of a greedy maximum coverage method to build various
local models.
➢LOCA outperforms the state-of-the-art global and local models
over various benchmark datasets.
26
Q&A
27
Code: https://github.com/jin530/LOCA
Email: zxcvxd@skku.edu

More Related Content

What's hot

Session-aware Linear Item-Item Models for Session-based Recommendation (WWW 2...
Session-aware Linear Item-Item Models for Session-based Recommendation (WWW 2...Session-aware Linear Item-Item Models for Session-based Recommendation (WWW 2...
Session-aware Linear Item-Item Models for Session-based Recommendation (WWW 2...민진 최
 
Recommending for the World
Recommending for the WorldRecommending for the World
Recommending for the WorldYves Raimond
 
Conditional CycleGANによる食事画像変換
Conditional CycleGANによる食事画像変換Conditional CycleGANによる食事画像変換
Conditional CycleGANによる食事画像変換Ryosuke Tanno
 
Recurrent Neural Networks. Part 1: Theory
Recurrent Neural Networks. Part 1: TheoryRecurrent Neural Networks. Part 1: Theory
Recurrent Neural Networks. Part 1: TheoryAndrii Gakhov
 
Introduction to Tree-LSTMs
Introduction to Tree-LSTMsIntroduction to Tree-LSTMs
Introduction to Tree-LSTMsDaniel Perez
 
Contextualization at Netflix
Contextualization at NetflixContextualization at Netflix
Contextualization at NetflixLinas Baltrunas
 
Past, Present & Future of Recommender Systems: An Industry Perspective
Past, Present & Future of Recommender Systems: An Industry PerspectivePast, Present & Future of Recommender Systems: An Industry Perspective
Past, Present & Future of Recommender Systems: An Industry PerspectiveJustin Basilico
 
Time, Context and Causality in Recommender Systems
Time, Context and Causality in Recommender SystemsTime, Context and Causality in Recommender Systems
Time, Context and Causality in Recommender SystemsYves Raimond
 
Deep Learning for Recommender Systems
Deep Learning for Recommender SystemsDeep Learning for Recommender Systems
Deep Learning for Recommender SystemsYves Raimond
 
Past, present, and future of Recommender Systems: an industry perspective
Past, present, and future of Recommender Systems: an industry perspectivePast, present, and future of Recommender Systems: an industry perspective
Past, present, and future of Recommender Systems: an industry perspectiveXavier Amatriain
 
Rl chapter 1 introduction
Rl chapter 1 introductionRl chapter 1 introduction
Rl chapter 1 introductionConnorShorten2
 
Uber Data Analysis - SAS Project
Uber Data Analysis - SAS ProjectUber Data Analysis - SAS Project
Uber Data Analysis - SAS ProjectKushal417
 
Explicit Density Models
Explicit Density ModelsExplicit Density Models
Explicit Density ModelsSangwoo Mo
 
The Science and the Magic of User Feedback for Recommender Systems
The Science and the Magic of User Feedback for Recommender SystemsThe Science and the Magic of User Feedback for Recommender Systems
The Science and the Magic of User Feedback for Recommender SystemsXavier Amatriain
 
Movie Recommendation engine
Movie Recommendation engineMovie Recommendation engine
Movie Recommendation engineJayesh Lahori
 
Uncertainty Awareness in Integrating Machine Learning and Game Theory
Uncertainty Awareness in Integrating Machine Learning and Game TheoryUncertainty Awareness in Integrating Machine Learning and Game Theory
Uncertainty Awareness in Integrating Machine Learning and Game TheoryRikiya Takahashi
 

What's hot (20)

Session-aware Linear Item-Item Models for Session-based Recommendation (WWW 2...
Session-aware Linear Item-Item Models for Session-based Recommendation (WWW 2...Session-aware Linear Item-Item Models for Session-based Recommendation (WWW 2...
Session-aware Linear Item-Item Models for Session-based Recommendation (WWW 2...
 
Human Action Recognition
Human Action RecognitionHuman Action Recognition
Human Action Recognition
 
Recommending for the World
Recommending for the WorldRecommending for the World
Recommending for the World
 
Conditional CycleGANによる食事画像変換
Conditional CycleGANによる食事画像変換Conditional CycleGANによる食事画像変換
Conditional CycleGANによる食事画像変換
 
Recurrent Neural Networks. Part 1: Theory
Recurrent Neural Networks. Part 1: TheoryRecurrent Neural Networks. Part 1: Theory
Recurrent Neural Networks. Part 1: Theory
 
Introduction to Tree-LSTMs
Introduction to Tree-LSTMsIntroduction to Tree-LSTMs
Introduction to Tree-LSTMs
 
Contextualization at Netflix
Contextualization at NetflixContextualization at Netflix
Contextualization at Netflix
 
Past, Present & Future of Recommender Systems: An Industry Perspective
Past, Present & Future of Recommender Systems: An Industry PerspectivePast, Present & Future of Recommender Systems: An Industry Perspective
Past, Present & Future of Recommender Systems: An Industry Perspective
 
Tuning learning rate
Tuning learning rateTuning learning rate
Tuning learning rate
 
Time, Context and Causality in Recommender Systems
Time, Context and Causality in Recommender SystemsTime, Context and Causality in Recommender Systems
Time, Context and Causality in Recommender Systems
 
그림 그리는 AI
그림 그리는 AI그림 그리는 AI
그림 그리는 AI
 
Deep Learning for Recommender Systems
Deep Learning for Recommender SystemsDeep Learning for Recommender Systems
Deep Learning for Recommender Systems
 
Past, present, and future of Recommender Systems: an industry perspective
Past, present, and future of Recommender Systems: an industry perspectivePast, present, and future of Recommender Systems: an industry perspective
Past, present, and future of Recommender Systems: an industry perspective
 
Rl chapter 1 introduction
Rl chapter 1 introductionRl chapter 1 introduction
Rl chapter 1 introduction
 
XGBoost & LightGBM
XGBoost & LightGBMXGBoost & LightGBM
XGBoost & LightGBM
 
Uber Data Analysis - SAS Project
Uber Data Analysis - SAS ProjectUber Data Analysis - SAS Project
Uber Data Analysis - SAS Project
 
Explicit Density Models
Explicit Density ModelsExplicit Density Models
Explicit Density Models
 
The Science and the Magic of User Feedback for Recommender Systems
The Science and the Magic of User Feedback for Recommender SystemsThe Science and the Magic of User Feedback for Recommender Systems
The Science and the Magic of User Feedback for Recommender Systems
 
Movie Recommendation engine
Movie Recommendation engineMovie Recommendation engine
Movie Recommendation engine
 
Uncertainty Awareness in Integrating Machine Learning and Game Theory
Uncertainty Awareness in Integrating Machine Learning and Game TheoryUncertainty Awareness in Integrating Machine Learning and Game Theory
Uncertainty Awareness in Integrating Machine Learning and Game Theory
 

Similar to Local collaborative autoencoders (WSDM2021)

RS in the context of Big Data-v4
RS in the context of Big Data-v4RS in the context of Big Data-v4
RS in the context of Big Data-v4Khadija Atiya
 
Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...
Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...
Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...Harry McLaren
 
Lessons learnt at building recommendation services at industry scale
Lessons learnt at building recommendation services at industry scaleLessons learnt at building recommendation services at industry scale
Lessons learnt at building recommendation services at industry scaleDomonkos Tikk
 
@RISK Unchained Webinar
@RISK Unchained Webinar@RISK Unchained Webinar
@RISK Unchained WebinarAndrew Sich
 
Why biased matrix factorization works well?
Why biased matrix factorization works well?Why biased matrix factorization works well?
Why biased matrix factorization works well?Joonyoung Yi
 
Explain Yourself: Why You Get the Recommendations You Do
Explain Yourself: Why You Get the Recommendations You DoExplain Yourself: Why You Get the Recommendations You Do
Explain Yourself: Why You Get the Recommendations You DoDatabricks
 
Towards a Quality Assessment of Web Corpora for Language Technology Applications
Towards a Quality Assessment of Web Corpora for Language Technology ApplicationsTowards a Quality Assessment of Web Corpora for Language Technology Applications
Towards a Quality Assessment of Web Corpora for Language Technology ApplicationsMarina Santini
 
NLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobil
NLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobilNLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobil
NLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobilDatabricks
 
Deep Learning for Recommender Systems
Deep Learning for Recommender SystemsDeep Learning for Recommender Systems
Deep Learning for Recommender SystemsMarcel Kurovski
 
Deep Learning for Recommender Systems
Deep Learning for Recommender SystemsDeep Learning for Recommender Systems
Deep Learning for Recommender Systemsinovex GmbH
 
It Does What You Say, Not What You Mean: Lessons From A Decade of Program Repair
It Does What You Say, Not What You Mean: Lessons From A Decade of Program RepairIt Does What You Say, Not What You Mean: Lessons From A Decade of Program Repair
It Does What You Say, Not What You Mean: Lessons From A Decade of Program RepairClaire Le Goues
 
Using Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning ModelsUsing Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning ModelsScott Clark
 
Using Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning ModelsUsing Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning ModelsSigOpt
 
Machine Learning, Deep Learning and Data Analysis Introduction
Machine Learning, Deep Learning and Data Analysis IntroductionMachine Learning, Deep Learning and Data Analysis Introduction
Machine Learning, Deep Learning and Data Analysis IntroductionTe-Yen Liu
 
In the age of Big Data, what role for Software Engineers?
In the age of Big Data, what role for Software Engineers?In the age of Big Data, what role for Software Engineers?
In the age of Big Data, what role for Software Engineers?CS, NcState
 
Using SigOpt to Tune Deep Learning Models with Nervana Cloud
Using SigOpt to Tune Deep Learning Models with Nervana CloudUsing SigOpt to Tune Deep Learning Models with Nervana Cloud
Using SigOpt to Tune Deep Learning Models with Nervana CloudSigOpt
 
Moving Toward Deep Learning Algorithms on HPCC Systems
Moving Toward Deep Learning Algorithms on HPCC SystemsMoving Toward Deep Learning Algorithms on HPCC Systems
Moving Toward Deep Learning Algorithms on HPCC SystemsHPCC Systems
 
Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...
Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...
Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...Spark Summit
 
Performance evaluation of IR models
Performance evaluation of IR modelsPerformance evaluation of IR models
Performance evaluation of IR modelsNisha Arankandath
 

Similar to Local collaborative autoencoders (WSDM2021) (20)

RS in the context of Big Data-v4
RS in the context of Big Data-v4RS in the context of Big Data-v4
RS in the context of Big Data-v4
 
Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...
Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...
Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...
 
Lessons learnt at building recommendation services at industry scale
Lessons learnt at building recommendation services at industry scaleLessons learnt at building recommendation services at industry scale
Lessons learnt at building recommendation services at industry scale
 
Machine Learning Impact on IoT - Part 2
Machine Learning Impact on IoT - Part 2Machine Learning Impact on IoT - Part 2
Machine Learning Impact on IoT - Part 2
 
@RISK Unchained Webinar
@RISK Unchained Webinar@RISK Unchained Webinar
@RISK Unchained Webinar
 
Why biased matrix factorization works well?
Why biased matrix factorization works well?Why biased matrix factorization works well?
Why biased matrix factorization works well?
 
Explain Yourself: Why You Get the Recommendations You Do
Explain Yourself: Why You Get the Recommendations You DoExplain Yourself: Why You Get the Recommendations You Do
Explain Yourself: Why You Get the Recommendations You Do
 
Towards a Quality Assessment of Web Corpora for Language Technology Applications
Towards a Quality Assessment of Web Corpora for Language Technology ApplicationsTowards a Quality Assessment of Web Corpora for Language Technology Applications
Towards a Quality Assessment of Web Corpora for Language Technology Applications
 
NLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobil
NLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobilNLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobil
NLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobil
 
Deep Learning for Recommender Systems
Deep Learning for Recommender SystemsDeep Learning for Recommender Systems
Deep Learning for Recommender Systems
 
Deep Learning for Recommender Systems
Deep Learning for Recommender SystemsDeep Learning for Recommender Systems
Deep Learning for Recommender Systems
 
It Does What You Say, Not What You Mean: Lessons From A Decade of Program Repair
It Does What You Say, Not What You Mean: Lessons From A Decade of Program RepairIt Does What You Say, Not What You Mean: Lessons From A Decade of Program Repair
It Does What You Say, Not What You Mean: Lessons From A Decade of Program Repair
 
Using Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning ModelsUsing Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning Models
 
Using Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning ModelsUsing Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning Models
 
Machine Learning, Deep Learning and Data Analysis Introduction
Machine Learning, Deep Learning and Data Analysis IntroductionMachine Learning, Deep Learning and Data Analysis Introduction
Machine Learning, Deep Learning and Data Analysis Introduction
 
In the age of Big Data, what role for Software Engineers?
In the age of Big Data, what role for Software Engineers?In the age of Big Data, what role for Software Engineers?
In the age of Big Data, what role for Software Engineers?
 
Using SigOpt to Tune Deep Learning Models with Nervana Cloud
Using SigOpt to Tune Deep Learning Models with Nervana CloudUsing SigOpt to Tune Deep Learning Models with Nervana Cloud
Using SigOpt to Tune Deep Learning Models with Nervana Cloud
 
Moving Toward Deep Learning Algorithms on HPCC Systems
Moving Toward Deep Learning Algorithms on HPCC SystemsMoving Toward Deep Learning Algorithms on HPCC Systems
Moving Toward Deep Learning Algorithms on HPCC Systems
 
Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...
Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...
Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...
 
Performance evaluation of IR models
Performance evaluation of IR modelsPerformance evaluation of IR models
Performance evaluation of IR models
 

Recently uploaded

PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...Sérgio Sacani
 
Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)PraveenaKalaiselvan1
 
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.Nitya salvi
 
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...Lokesh Kothari
 
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...Monika Rani
 
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...ssuser79fe74
 
Biopesticide (2).pptx .This slides helps to know the different types of biop...
Biopesticide (2).pptx  .This slides helps to know the different types of biop...Biopesticide (2).pptx  .This slides helps to know the different types of biop...
Biopesticide (2).pptx .This slides helps to know the different types of biop...RohitNehra6
 
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptxSCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptxRizalinePalanog2
 
Pests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPirithiRaju
 
Creating and Analyzing Definitive Screening Designs
Creating and Analyzing Definitive Screening DesignsCreating and Analyzing Definitive Screening Designs
Creating and Analyzing Definitive Screening DesignsNurulAfiqah307317
 
Zoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdfZoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdfSumit Kumar yadav
 
Forensic Biology & Its biological significance.pdf
Forensic Biology & Its biological significance.pdfForensic Biology & Its biological significance.pdf
Forensic Biology & Its biological significance.pdfrohankumarsinghrore1
 
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPirithiRaju
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bSérgio Sacani
 
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...chandars293
 
Isotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoIsotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoSérgio Sacani
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​kaibalyasahoo82800
 
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls Agency
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls AgencyHire 💕 9907093804 Hooghly Call Girls Service Call Girls Agency
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls AgencySheetal Arora
 
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLKochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLkantirani197
 

Recently uploaded (20)

PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...
 
CELL -Structural and Functional unit of life.pdf
CELL -Structural and Functional unit of life.pdfCELL -Structural and Functional unit of life.pdf
CELL -Structural and Functional unit of life.pdf
 
Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)
 
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
 
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
 
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
 
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
 
Biopesticide (2).pptx .This slides helps to know the different types of biop...
Biopesticide (2).pptx  .This slides helps to know the different types of biop...Biopesticide (2).pptx  .This slides helps to know the different types of biop...
Biopesticide (2).pptx .This slides helps to know the different types of biop...
 
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptxSCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
 
Pests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdf
 
Creating and Analyzing Definitive Screening Designs
Creating and Analyzing Definitive Screening DesignsCreating and Analyzing Definitive Screening Designs
Creating and Analyzing Definitive Screening Designs
 
Zoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdfZoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdf
 
Forensic Biology & Its biological significance.pdf
Forensic Biology & Its biological significance.pdfForensic Biology & Its biological significance.pdf
Forensic Biology & Its biological significance.pdf
 
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
 
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
 
Isotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoIsotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on Io
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​
 
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls Agency
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls AgencyHire 💕 9907093804 Hooghly Call Girls Service Call Girls Agency
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls Agency
 
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLKochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
 

Local collaborative autoencoders (WSDM2021)

  • 1. Local Collaborative Autoencoders Minjin Choi1, Yoonki Jeong1, Joonseok Lee2, Jongwuk Lee1 Sungkyunkwan University (SKKU), South Korea1 Google Research, United States2
  • 3. Global Low-rank Assumption ➢Existing models are based on the global low-lank assumption. • All users and items share the same latent features. ➢Limitation: some users/items may have different latent features. 3 1 1 ? ? 1 1 1 1 ? ? 1 1 ? 1 ? ? 1 1 1 ? ? 1 1 ? 1 0.5 ... 0.1 0.6 ... 0.2 0.7 ... 0.3 … ... ... 0.1 ... 0.7 0.2 0.4 1.2 … 2.3 … ... … ... ... 0.4 0.6 0.3 … 1.5 k features k features
  • 4. Local Low-rank Assumption ➢A user-item matrix can be divided to several sub-matrices with the local low-rank assumption. • Each sub-matrix represents different communities. • Local models represent various communities with different characteristics. 4 1 1 ? ? 1 1 1 1 ? ? 1 1 ? 1 ? ? 1 1 1 ? ? 1 1 ? 1
  • 5. Limitation of Existing Local Models ➢If the local model is too large, it is close to the global model. • Because LLORMA uses large local models, the local model may not represent its unique characteristic. • The performance gain may come from an ensemble effect. 5 1 1 ? ? 1 1 1 1 ? ? 1 1 ? 1 ? ? 1 1 1 ? ? 1 1 ? 1 1 1 ? ? 1 1 1 1 ? ? 1 1 ? 1 ? ? 1 1 1 ? ? 1 1 ? 1 Lee et al., "LLORMA: Local Low-Rank Matrix Approximation," JMLR 2016
  • 6. Limitation of Existing Local Models ➢If the local model is too small, the accuracy is too low. • Because sGLSVD uses small local models, some local model may have insufficient training data. 6 1 1 ? ? 1 1 1 1 ? ? 1 1 ? 1 ? ? 1 1 1 ? ? 1 1 ? 1 Evangelia Christakopoulou and George Karypis, “Local Latent Space Models for Top-N Recommendation,” KDD 2018 1 1 ? ? 1 1 1 1 ? ? 1 1 ? 1 ? ? 1 1 1 ? ? 1 1 ? 1
  • 7. Research Question 7 How to build coherent and accurate local models?
  • 8. Our Key Contributions ➢When the local model keeps small and coherent, we build the local model with a relatively large training data. 8 (Dataset: ML10M) Training matrix == Test matrix training matrix > test matrix
  • 9. Our Key Contributions ➢Autoencoder-based models are used as the base model to train the local model. • They are useful for capturing non-linear and complicated patterns. 9 1 1 ? ? 1 1 1 1 ? ? 1 1 ? 1 ? ? 1 1 1 ? ? 1 1 ? 1 Sedhain et al., "AutoRec: Autoencoders Meet Collaborative Filtering," WWW 2015 2 1 3 … 2 1 3 … … 𝒉(𝒓) 𝒓 ො 𝒓 𝑾, 𝒃 𝑾′ , 𝒃′ Autoencoder-based models take each row as input.
  • 11. Local Collaborative Autoencoders (LOCA) ➢Overall architecture of LOCA • Step 1: Discovering two local communities for an anchor user • Step 2: Training a local model with an expanded community • Step 3: Combining multiple local models 11 A local community for an anchor user An expanded local community Local model q … Final model Local model 2 Local model 1
  • 12. Step 1: Discovering Local Communities ➢For an anchor user, determine a local community and expand the local community for training. 12 1.0 0.6 0.3 0.5 0.4 0.7 0.4 0.2 Calculate the similarities between the anchor user and the other users. … … 1.0 0.7 0.6 1.0 0.7 0.6 0.5 0.5 0.4 0.4 Neighbors for the local community Expanded neighbors to train the local community
  • 13. Step 2: Training a Local Model ➢Train the local model with the expanded community. • Use the autoencoder-based model for training the local model. • Note: It is possible to utilize any base models, e.g., MF, AE and EASER. • The similarities with the anchor are used for the user weights for training. 13 1.0 0.6 0.3 0.5 0.4 0.7 0.4 1 1 ? 1 1 1 1 1 1 1 ? ? 1 1 ? 1 1 ? ? ? 1 1 1 1 1 1 ? ? … … … … … … … … Harald Steck., "Embarrassingly shallow autoencoders for sparse data,” WWW 2019.
  • 14. Step 3: Combining Multiple Local Models ➢Aggregate multiple local models with the weight of each local model. 14 Local model 1 Local model q … 1 1 ? 1 1 1 1 1 1 1 ? ? 1 1 ? 1 1 ? ? ? 1 1 1 1 1 1 ? ? ? 1 1 ? … … … … … … … … … … … … … … … .9 .4 .5 .7 .8 .9 .6 .6 .7 .8 .5 .8 … .7 .4 .5 .9 .7 .5 .5 .8 .8 .6 .7 .8 … … … .9 .4 .5 .7 .8 .9 .6 .6 .7 .8 .5 .8 .8 .7 .5 1 .7 .4 .5 .9 .7 .5 .5 .8 .8 .6 .7 .8 .6 1 1 .5 … … … … … … … … … … … … … … … … … … … … …
  • 15. Training Local Models in Detail ➢The loss function for the 𝑗-th local model ➢Aggregating all local models and a global model 15 argmin 𝜽(𝒋) ෍ 𝒓𝒖∈𝑹 𝒕𝒖 (𝒋) ℒ 𝒓𝒖, 𝑴𝒍𝒐𝒄𝒂𝒍 (𝒓𝒖; 𝜽(𝒋) ) + 𝝀𝛀(𝜽(𝒋) ) ෡ 𝑹 = 𝜶𝑴𝒈𝒍𝒐𝒃𝒂𝒍 𝑹; 𝜽 𝒈 + 𝟏 − 𝜶 ෍ 𝒋=𝟏 𝒒 𝒘(𝒋) ⊙ 𝑴𝒍𝒐𝒄𝒂𝒍(𝑹; 𝜽(𝒋)) ⊘ 𝒘 Parameter for the 𝒋-th local model The 𝒋-th local model User weight for training the 𝒋-th local model User weight for inferring the 𝒋-th local model The global model is used to handle the users who are not covered by local models.
  • 16. How to Choose Anchor Users ➢Random selection: Choose 𝑘 anchor users at random. ➢Need to maximize the coverage of users by local models. • An optimal maximum coverage algorithm is the NP-hard problem. ➢Use the greedy method to maximize the coverage. • Select the anchor user who has the most uncovered users iteratively. 16 Group1 Group 2 Group 3
  • 18. Experimental Setup: Dataset ➢We evaluate our model over five public datasets with various characteristics (e.g., domain, sparsity). 18 Dataset # of users # of items # of ratings Sparsity MovieLens 10M (ML10M) 69,878 10,677 10,000,054 98.66% MovieLens 20M (ML20M) 138,493 26,744 20,000,263 99.46% Amazon Music (AMusic) 4,964 11,797 97,439 99.83% Amazon Game (AGame) 13,063 17,408 236,415 99.90% Yelp 25,677 25,815 731,671 99.89%
  • 19. Evaluation Protocol and Metrics ➢Evaluation protocol: leave-5-out • Hold-out the last 5 interactions as the test data for each user. ➢Evaluation metrics • Recall@100 • Measures the number of test items included in the top-N list. • NDCG@100 • Measures the ranking of test items in the top-N list. 19 Time Training data: Test data: User interaction:
  • 20. Competitive Global/Local Models ➢Four autoencoder-based global models • CDAE: a denoising autoencoder-based model with a latent user vector • MultVAE: a VAE-based model • RecVAE: a VAE-based model by improving MultVAE • EASER: an item-to-item latent factor model ➢Two local models • LLORMA: local model using MF as the base model • sGLSVD: local model using SVD as the base model 20 Yao Wu et al., “Collaborative Denoising Auto-Encoders for Top-N Recommender Systems,” WSDM 2016 Dawen Liang et al., "Variational autoencoders for collaborative filtering,” WWW 2018. Ilya Shenbin et al., "RecVAE: A new variational autoencoder for Top-N recommendations with implicit feedback,” WSDM 2020. Harald Steck., "Embarrassingly shallow autoencoders for sparse data,” WWW 2019. Lee et al., "LLORMA: Local Low-Rank Matrix Approximation," JMLR 2016 Evangelia Christakopoulou and George Karypis, “Local Latent Space Models for Top-N Recommendation,” KDD 2018
  • 21. Accuracy: LOCA vs. Competing Models 21 Dataset Metric CDAE MultVAE EASER RecVAE LLORMA sGLSVD LOCAVAE LOCAEASE ML10M Recall@100 0.4685 0.4653 0.4648 0.4705 0.4692 0.4468 0.4865 0.4798 NDCG@100 0.1982 0.1945 0.2000 0.1996 0.2042 0.1953 0.2073 0.2049 ML20M Recall@100 0.4324 0.4397 0.4468 0.4417 0.3355 0.4342 0.4419 0.4654 NDCG@100 0.1844 0.1860 0.1948 0.1857 0.1446 0.1919 0.1884 0.2024 Amusic Recall@100 0.0588 0.0681 0.0717 0.0582 0.0517 0.0515 0.0748 0.0717 NDCG@100 0.712 0.0822 0.0821 0.0810 0.0638 0.0613 0.0893 0.0826 Agames Recall@100 0.1825 0.2081 0.1913 0.1920 0.1223 0.1669 0.2147 0.1947 NDCG@100 0.0808 0.0920 0.0915 0.0849 0.0539 0.0777 0.0966 0.0922 Yelp Recall@100 0.2094 0.2276 0.2187 0.2262 0.1013 0.1965 0.2354 0.2205 NDCG@100 0.0920 0.0982 0.0972 0.0975 0.0429 0.0857 0.1103 0.0981 Global Models Local Models Ours ➢LOCA consistently outperforms competitive global/local models over five benchmark datasets.
  • 22. Effect of the Number of Local Models ➢The accuracy of LOCA improved consistently with an increase in the number of local models. 22 0.185 0.19 0.195 0.2 0.205 0.21 0 50 100 150 200 250 300 NDCG@100 The number of local models ML10M MultVAE Ensemble_VAE LLORMA_VAE LOCA_VAE 0.078 0.08 0.082 0.084 0.086 0.088 0.09 0 50 100 150 200 250 300 NDCG@100 The number of local models AMusic MultVAE Ensemble_VAE LLORMA_VAE LOCA_VAE
  • 23. Effect of Anchor Selection Method ➢Our coverage-based anchor selection outperforms the other methods in terms of accuracies and coverages. 23 0.196 0.198 0.2 0.202 0.204 0.206 0.208 50 100 150 200 250 300 NDCG@100 The number of local models ML10M Random K-means Ours 0.08 0.082 0.084 0.086 0.088 0.09 50 100 150 200 250 300 NDCG@100 The number of local models AMusic Random K-means Ours
  • 24. Illustration of LOCA ➢When a user has multiple tastes, LOCA can capture the user preference by combining different local patterns. • For a user (66005 in ML10M) who likes Sci-Fi and Horror movies, LOCA shows a better accuracy. 24 Recommendation Local 70 Local 179 Global Ground Truth Top-1 Sci-Fi Adventure Horror Action Thriller Action Sci-Fi Action Top-2 Sci-Fi Horror Horror Drama Drama Horror Thriller Top-3 Sci-Fi Action Horror Drama Drama Mystery Horror Action
  • 26. Conclusion ➢We propose a new local recommender framework, namely local collaborative autoencoders (LOCA). ➢LOCA can handle a large number of local models effectively. • Adopts a local model with different training/inference strategies. • Utilizes autoencoder-based model as the base model. • Makes use of a greedy maximum coverage method to build various local models. ➢LOCA outperforms the state-of-the-art global and local models over various benchmark datasets. 26