FYP 24108

Language-model based recommender algorithm

Combine the power of language model and the collaberative filtering data

Introduction

Background

•Cold-start challenge in recommender systems (new users/items with no interaction history).

•Potential of Language Models (LMs): Generate semantic representations from textual metadata.

•Goal: Hybrid LM + Collaborative Filtering (CF) for zero-shot recommendations.

Objective

•Design high-performance CF architecture.

•Develop LM-CF integration framework.

•Optimize LM fine-tuning strategies

Contribution

•Self-attention CF model (32% ROC-AUC and 87% PR-AUC improvement compared to cos-similarity model).

•Empirical validation of LM embeddings for cold-start scenarios.

•Decoupled finetuning framework for reduced computational costs.

Methodology

Our comprehensive suite of professional services caters to a diverse clientele, ranging from homeowners to commercial developers.

Self-Attention Model

Integrating with language model

  1. In dataset of this project, only item has text description
  2. Replace item embedder with LM embeddings
  3. User embedder remains unchanged

Evaluation

Metrics

ROC-AUC (Receiver Operating Characteristic AUC):

Measures classification discriminative power across all thresholds.

Baseline = 0.5 (random prediction performance).

PR-AUC (Precision-Recall AUC):

Focuses on precision-recall tradeoffs for positive class detection.

Baseline = 0.12 (equal to dataset sparsity with 12% interaction density).

Performance of Self-Attention Model for know items

Performance of LM integrated model for new items