Recommender Systems

Poznámky na přípravu na zkoušku z RS 2023/2024

Collaborative Filtering (CF) (Lecture 1)


Pure CF Approaches


User-based Nearest-Neighbor Collaborative Filtering


Measuring User Similarity

  1. Pearson Correlation:

    sim(u,v)=iIuv(ruir¯u)(rvir¯v)iIuv(ruir¯u)2iIuv(rvir¯v)2
    • rui: Rating of user u for item i
    • r¯u: Average rating of user u
    • Iuv: Set of items rated by both u and v
  2. Jaccard Similarity:

    J(A,B)=|AB||AB|

Making Predictions


Memory-based vs. Model-based Approaches


Item-based Collaborative Filtering


Data Sparsity Problems


Matrix Factorization


Matrix Factorization Algorithm

On slides there are better pseudocode with more intuitive thinking


Hyperparameter Tuning


Advanced Collaborative Filtering Techniques


Collaborative Filtering Issues


Matrix Factorization Techniques


Optimizing Matrix Factorization


Ranking-oriented Optimization


Content Based Recommendation (Lecture 2)

Overview

Definitions


Item Representation and Content

Structured vs. Unstructured Data

Examples


Simple Content Representation

Keyword Overlap

Jaccard Similarity

Example


Term-Frequency - Inverse Document Frequency (TF-IDF)

TF-IDF Overview

Formula

Example


Improving the Vector Space Model

Techniques

Advanced Methods


Content-Based RS Beyond Text

Numeric and Nominal Attributes

Example Approaches


Content-Based RS: Multimedia and Attribute Data

Combined Approach

Practical Examples


Query-Based Retrieval: Rocchio's Method

Overview

Formula

Bayesian Relevance Feedback


Explicit Decision Models

Decision Trees

Rule Induction


Feature Selection

Strategies

Practical Tips


Limitations and Challenges

Common Issues

Solutions

Summary


Knowledge-Based Recommendation

Basic I/O Relationship

Types

Examples


Evaluation of Recommender Systems (Lecture 4)

Overview

Motivation

Scientific Requirements for Evaluation

Reproducibility

Evaluation Criteria

Components of Evaluation Research

Subjects

Settings

Measured Variables

Design Types

Evaluation Settings and Techniques

Online Experiments

Lab Studies

Offline Evaluation

Methodology

Bias and Variance

Metrics for Evaluating Recommender Systems

Relevance and Ranking Metrics

Rank-Based Metrics

Advanced Metrics

Practical Challenges and Solutions

Causality and Biases

Offline Evaluation Metrics

Summary and Best Practices

Key Takeaways

Future Directions


Recap

:
$$
\text{Diversity} = \frac{1}{N} \sum_{i=1}^{N} \sum_{j=i+1}^{N} \text{diversity}(i, j)
$$

Linear Monotone Preference Model (Lecture 5)

Overview

Conflicting Requirements


Ordering and Preference Scales

Human-Intuitive Ordering

User Preferences


Preferential Sets and Fuzzy Sets

Preferential Sets

Preference Scale


Decathlon Data Example

Multi-Criterial Scale Points

Dominance and Pareto Ordering

Data Cube and Preference Cube

Sum of Points


Linear Monotone Preference Model (LMPM)

Definition

Aggregation Function

Contour Lines


Building the LMPM Step-by-Step

Triangular Preference Functions

Trapezoidal Preference Functions


Calculating Preference Degrees

Mapping Data to Preferences

Aggregation and Ordering

Example Calculation


Dynamical Model and Time Simulation

Moving Ideal Points

Example Sessions


Aggregation and Visualization

Aggregation Function

Contour Line Visualization

Practical Applications


Extensions and Practical Considerations

Handling Complex Preferences

Market Segmentation

Real-World Applications

Simulation and Extensions


Querying Threshold Algorithm (Lecture 6)

Overview

User Preferences and Data Cube

Preference Model

Visualization

Threshold Algorithms (FLN TA)

Web Service Motivation

Data Model

Threshold Algorithm Steps

  1. Sorted Access:

    • Access each list Li sequentially.
    • Perform random access to other lists to find grades.
    • Compute:t(R)=t(x1R,,xmR)
    • Remember the top-k objects.
  2. Define Threshold:

    • For each list Li, let xi be the grade of the last object seen.
    • Define threshold t:t=t(x1,,xm)
  3. Output:

    • Set Y containing top-k objects with highest grades.

Algorithm Illustration

Example

Heuristics and Optimizations

Incremental Threshold Algorithm

FLN Data Model and Multiuser Extensions

Multiuser Model

Proof of Correctness

FLN Theorem 4.1

Proof Outline

No Random Access (NRA) Algorithm

NRA Algorithm Steps

  1. Access Each List:

    • Calculate x1d,,xmd.
    • For each item R, calculate Sd(R),Wd(R),Bd(R)
  2. Viability Check:

    • An object R is viable if:BSd(R)>Mkd
  3. Output Set:

    • The set Tkd contains the top-k objects.

Further Notes

Correctness of NRA

Conclusion


Hybrid Recommender Systems (Lecture 7)

Overview

Hybridization Designs

Hybridization Design Types

Monolithic Hybridization Design

Feature Augmentation

Visual Bayesian Personalized Ranking (BPR)

Parallelized Hybridization Design

Weighted Combination

Estimating Weights

Switching and Mixed Hybridization

Switching

Mixed

Pipelined Hybridization Design

Cascade

Meta-Level

Limitations and Considerations

Limitations of Hybrid Strategies

Example Formulas and Metrics

Weighted Combination Formula

Weighted Score=w1Score1+w2Score2

Dynamic Weight Adjustment

MAE=1ni=1n|yiy^i|

where yi is the actual rating and y^i is the predicted rating.

Reciprocal Recommender Systems (Lecture 8)

People Recommender Systems

Overview

Reciprocal Recommender Systems in Online Dating

Modeling Likes and Dislikes

Insights from Recent Studies

Practical Considerations

People Recommender Systems in Job Recruitment

Job Recommender Systems

Effectiveness of Job Title-Based Embeddings

Follow-Up Studies

Sentence-Pair Classification for Algorithmic Recruiting (2023)

Career Path Prediction

Task Definition

Approach

Practical Considerations and Challenges

Challenges in Reciprocal Recommender Systems

Improving Recommendations


Deep learning (Lecture 9)

Here are the detailed notes extracted from the provided slides on "Deep Learning for Recommender Systems" with headers indented one level deeper and all mathematical expressions formatted with double dollar signs for LaTeX rendering.


Deep Learning for Recommender Systems

Why Deep Learning?

Neural Networks and Learning

Neural Model

Feedforward Multilayered Network

Modern Deep Networks

Key Ingredients

Deep Learning for Recommender Systems

Main Topics as of 2017

Research Directions

Best Practices

Item Embeddings and 2vec Models

Embeddings

Matrix Factorization (MF)

Word2Vec

Continuous Bag of Words (CBOW)

Skip-gram

Paragraph2Vec (Doc2Vec)

Prod2Vec

Skip-gram on Products

Meta-Prod2Vec

Feature Extraction from Content

Images, Text, and Audio

Music Representations

Audio Features in Recommender Systems

Neural Collaborative Filtering (NCF)

NeuMF (He et al. 2017)

Session-Based Recommendations with RNNs

GRU4Rec (Hidasi et al. 2015)

Follow-Up Development

Utilizing More Information

Meta-Prod2Vec and Content2Vec

Context-Aware Recommender Systems

Fairness and Bias in Recommender Systems

Addressing Biases

Outlook

Future Directions