This project focuses on implementing recommendation systems using large language models (LLMs). Specifically, we employ the P5 recommendation model for rating prediction and sequential recommendation tasks. Additionally, we explore the transformer-decoder architecture, including fine-tuning with GPT-2 and traditional RNN models.
The P5 model is utilized for two primary tasks:
- Rating Prediction: Predicting user ratings for items.
- Sequential Recommendation: Recommending the next item in a user's sequence of interactions.
We experiment with the transformer-decoder architecture, fine-tuning it using GPT-2 for enhanced sequential recommendation performance. This architecture helps in capturing long-term dependencies in user interaction sequences.
Traditional RNN models are also implemented to compare their performance with the transformer-based approaches. RNNs are useful for modeling sequential data, though they may struggle with long-term dependencies compared to transformer models.
The models are evaluated on various datasets, including:
- MovieLens
- Amazon Beauty
- Delicious
These datasets provide a diverse set of user-item interactions, enabling comprehensive evaluation of the models' performance.