Augmented recipe recommendation using flavor profile, ingredients, and cooking technique¶
As the amount of information on the internet has ballooned, recommendation systems have become increasingly crucial to help users find desired information without having to do extensive manual search. For a while I have been curious about how these systems work given the extensive way they touch my daily life. I decided to explore various approaches to recommendation algorithms in the domain of food and recipes, which is particularly relevant given the vast number of online recipes.
The goal of a recommender system is to predict the rating a user would give to a new item and to suggest to the user items for which the predicted rating is high. Several common approaches are:
Collaborative filtering (CF) is a method that uses the ratings of many users over many items to identify similar users and predict the rating a user would give to an item based on the ratings given by similar users. The only data necessary for this approach is ratings history on items [5].
Content-based (CB) is a method that uses information about items to calculate similarity between new items and items a user has historically rated to predict the rating a user would give to an item [1].
Hybrid approaches aim to combine the strengths of multiple previously mentioned approaches. This can be done in many ways, such as averaging results of CF and CB predictions, using the prediction from one approach as the input to another, and introducing more advanaced optimization techniques to apply elements from both CB and CF appraoches in one algorithm. [1, 7]
Project Goals¶
I aimed to explore and compare the performance of CB, CF, and hybrid recommender algorithms in the context of recipe recommendation, while leveraging recipe flavor profiles and recipe metadata (meal type, cooking technique, cooking time) in the recommendation.
Data Sources¶
The first data set is a set of recipes from Food.com, made available on as a Kaggle dataset (https://www.kaggle.com/datasets/shuyangli94/food-com-recipes-and-user-interactions?resource=download&select=RAW_recipes.csv). The recipes data ncludes the name of the recipe, a description of the recipe, recipe tags, the nutritional value, the steps in make the recipe, and the ingredients.\ The second data set is a set of recipe interactions also from Food.com and also made available through Kaggle (https://www.kaggle.com/datasets/shuyangli94/food-com-recipes-and-user-interactions?resource=download&select=RAW_interactions.csv). This data set includes the rating and review given to a recipe by various users (recipe id matches up to the recipe id in the recipes data set above).\ The third data set provides the flavor molecules and associated flavor profile for a given food (from https://cosylab.iiitd.edu.in/flavordb/). It is accessed via API calls for each ingredient in the recipes from the first datset.\ A final data set provides ingredient lemmatization, or in orther words associating each possible ingredient with a \"base\" version of that ingredient (e.g. all types of lettuce become \"lettuce\"). Also made available through Kaggle (https://www.kaggle.com/datasets/shuyangli94/food-com-recipes-and-user-interactions/discussion/118716?resource=download&select=ingr_map.pkl).
Algorithm Details¶
For this project I coded each of the following algorithms from scratch using numpy and Pytorch, among other packages.
Collaborative filtering (CF)¶
Nearest neighbors: calculate the nearest neighbors of a new user measured by cosine similarity: $$Sim(u_i, u_k) := \frac{r_i * r_k}{||r_i||*||r_k||} = \frac{\sum_{j=1}^mr_{ij}r_{kj}}{\sqrt{\sum_{j=1}^mr_{ij}^2\sum_{j=1}^mr_{kj}^2}}$$ where $r_i$ and $r_k$ are ratings vectors for users $u_i$ and $u_k$.
Predict a user's rating on a new recipe $r_{ij}$ by weighted average with bias avoided by by subtracting each user's average rating $\tilde{r_k}$ from their rating of the recipe and adding in the target user's average rating $\tilde{r_i}$: $$r_{ij} = \tilde{r_i}+\frac{\sum_kSim(u_i, u_k)(r_{kj}-\tilde{r}_k)}{\text{num ratings}}$$
Matrix Factorization: aims to decompose the user's preferences for into preferences for a set of latent factors. Matrix factorization can be performed using Singular Value Decomposition (SVD): $$M = U\Sigma V^T$$ By selecting the top $k$ singular values of matrix $\Sigma$, we can reconstruct matrix $M$ with less dimensions but still capturing much of the variability of the original matrix [9]. The concept here, when applied over recipe ratings, would be to find the dimensions of latent food preferences so as to avoid having to deal with the high dimensionality of individual recipe ratings.
However, when factoring a sparse matrix, it's more efficient to use Non-negative Matrix Factorization (NMF), which involves finding $P$ and $Q$ such that the reconstructed user-item rating $\hat{r}_{ui}= q_i^Tp_u$ is as close as possible to the true ${r}_{ui}$. In order to find $P$ and $Q$, the Mean Squared Error is minimized:
$$min_{q,p} \sum_{(u, i) \in TR} (r_{ui} - q_i^Tp_u)^2 + \lambda(||q_i||^2+||p_u||^2)$$where $p_u$ is the user vector, the $u$-th row of matrix $P$, and $q_i$ is the item vector, the $i$-th row of matrix $Q$, and $TR$ is the training set [9]. I implemented this optimization by hand by performing Gradient Decent according to the implementation in Luo et al. (2014). On each update of the Gradient Decent, the entries of the $P$ and $Q$ matrices are updated as below:
$$p_{u,k} \leftarrow p_{u,k}\frac{\sum_{i \in TR}q_{k,i} r_{u,i}}{|I_u|\lambda p_{u,k} + \sum_{i \in TR} \hat r_{u,i}}$$$$q_{k,i} \leftarrow q_{k,i}\frac{\sum_{i \in TR}p_{u,k} r_{u,i}}{|U_i|\lambda q_{k,i} + \sum_{i \in TR} \hat r_{u,i}}$$where $I_u$ is the number of ratings for that user in the item set, and $U_i$ is the number of rating for that item in the user set. A prediction for a new user-recipe pair is simply the $\hat r_{ui}$ entry in the reconstructed $\hat R = PQ^T$ matrix.\
Content-based (CB)¶
A rating for each user on each ingredient is calculated as the average of the ratings each user gave to all recipes including that ingredient: $$rat(u_i, ingr_j) = \frac{\sum_{l; ingr_j \in l}r_{il}}{l}$$ where $r_{il}$ is the rating user $i$ gave to recipe $l$. This formula is then applied over the flavor profile of each recipe, tags, and cooking techniques, to create a comprehensive recipe-based data source for each user. Predict a user's rating on a new recipe $r_{ij}$ by finding the average rating across all the ingredients, flavors, and cooking techniques in the new recipe:
$$r_{ij} = \frac{\sum_{l\in rec_j} rat(u_i, ingr_l)}{l}$$Hybrid¶
Content-augmented CF using cosine similarity: attempts to generate as many ratings as possible for a user on ingredients, flavors, and techniques using ratings given by similar users, and then uses an average of the ratings of the content of a recipe to predict a new user-recipe rating.\ To be more specific, this involved three steps:
1. I used Equation 1 to find the nearest neighbors of a new user (as in the CF approach).
2. Then, I used the following equation to predict a new user's ratings on ingredients, flavors, and techniques that they hadn't already rated:
$$rat(u_i, ingr_d) = \frac{\sum_k Sim(u_i, u_k)rat(u_k, ingr_d)}{\text{num ratings of }d}$$3. Then I used Equation 8 to predict a user's rating on a new recipe $r_{ij}$.
Content-augmented matrix factorization: takes the matrix factorization approach to CF and augments the item data with content. I used NMF for this approach, like the basic matrix factorization approach. However, I incorporated recipe content information into this factorization by further factoring the matrix $Q$ (shape: num features by num recipes) as $X\Phi$, where $X$ is a matrix (shape: num recipes by num ingredients) in which $X_{id}$ is a binary indicator if ingredient $d$ is in recipe $i$. This results in the following NMF factorization of matrix $R$: $$R = P\Phi^TX^T$$ My source for this approach is Forbes et al. (2011). This updated NMF results in the following minimization of MSE: $$min_{\phi,p} \sum_{(u,i) \in TR} (r_{ui} - p_u\Phi^Tx_i^T)^2$$
Rather than implementing the updates to each matrix on each iteration of Gradient Descent by hand like I did for the basic matrix factorization, I decided to use Pytorch to calculate and update the matrices according to MSE loss. Once training is complete, a prediction for a new user-recipe pair is simply the $\hat r_{ui}$ entry in the reconstructed $\hat R = P\Phi^TX^T$ matrix.
Summary of Findings¶
I trained each model above and then generated predictions using each user-recipe pair in the testing set. I used Root Mean Square Error (RMSE), Mean Absolute Error (MAE) and coverage (ability to generate predictions) [6] to evaluate the performance of each approach. The baseline model randomly predicts ratings for user-recipe pair by sampling from a unifrom distribution.
Model | RMSE | MAE | Coverage |
---|---|---|---|
Baseline | 2.5185 | 2.04112 | 1.0 |
Vanilla CF | 0.9805 | 0.5617 | 1.0 |
Matrix factorization CF | 1.7859 | 1.4959 | 0.6837 |
CB | 1.3987 | 1.0338 | 0.3343 |
Content-augmented CF | 1.0787 | 0.8480 | 0.9995 |
Content-augmented matrix factorization | 4.1741 | 3.8136 | 1.1 |
Full Project¶
Data Preprocessing: Recipe Ingredients¶
I started by processing the ingredient map that was part of the Kaggle data. I created a unified list of ingredients for all recipes in the following way: for each recipe, I applied a SpaCy model to a string representation of its ingredients and only the noun and proper nouns were preserved, and then all noun and proper nouns were given a unique id and added to a de-duplicated list. The reason for this pre-processing step is that each word in the list was queried in the FlavorDB API, and it wouldn't do to query a modifier (e.g. "fresh\" in the ingredient "fresh strawberries\").
import pandas as pd
pd.options.mode.chained_assignment = None
import spacy
nlp = spacy.load("en_core_web_sm")
# Import ingredients data
ingr_map = pd.read_pickle("food.com_recipes/ingr_map.pkl")
def nounify_ingredient(ingredient_str):
# Remove ingredients that aren't nouns or proper nouns
if len(ingredient_str.split(" ")) == 1:
return ingredient_str
else:
return " ".join([t.text for t in nlp(ingredient_str) if t.pos_ == 'NOUN' or t.pos_ == "PROPN"])
def build_ingredient_dict(ingr_map):
# Filter by if the count in the data is > 10
filt_ingr = ingr_map[ingr_map['count'] > 10][['id', 'replaced']].drop_duplicates()
# Remove ingredients that aren't nouns or proper nouns
filt_ingr['ingredients'] = filt_ingr['replaced'].apply(nounify_ingredient)
ingr_dict = dict(zip(filt_ingr['id'], filt_ingr['ingredients']))
return ingr_dict
ingr_dict = build_ingredient_dict(ingr_map)
list(ingr_dict.items())[:10]
Querying FlavorDB¶
For each ingredient in this list, I queried the FlavorDB search API to find the most similar search result that exists in their DB. For the search match, the search term was then used in the query for specific flavor molecules, which were subsequently matched with their individualflavors.
Below is the flavor profile for "lettuce"
import urllib.request
import json
import seaborn as sns
import matplotlib.pyplot as plt
from wordcloud import WordCloud
from ast import literal_eval
def find_flavors(ingredient):
# For the query, spaces must be & symbols instead
search_term = ingredient.replace(" ", "&")
# Initiate list of flavors
flavors = []
# Create search url for ingredient to get back it's id
bu = f"https://cosylab.iiitd.edu.in/flavordb/entities?entity={search_term}&category="
with urllib.request.urlopen(bu) as url:
b = json.loads(json.loads(url.read().decode()))
# find the result that has the ingredient
b = list(filter(lambda x: x['entity_alias_readable'].lower() in ingredient, b))
# if nothing is returned, end func
if len(b) == 0:
return []
print(f"found ingredient: {b[0]['entity_alias_readable']}")
flavdb_id = b[0]['entity_id']
# Create search url for ingredient to get back flavor molecules
bu = f"https://cosylab.iiitd.edu.in/flavordb/entities_json?id={flavdb_id}"
with urllib.request.urlopen(bu) as url:
a = json.loads(url.read().decode())
for mol in a['molecules']:
flavors.extend(mol['fooddb_flavor_profile'].split("@"))
# Return a list of all flavors
return list(set(flavors))[1:]
with open('data/ing_flav_dict.txt', 'r') as f:
ing_flav_dict = json.load(f)
wordcloud = WordCloud(width = 800, height = 800,
background_color ='white',
min_font_size = 10).generate(" ".join(ing_flav_dict['lettuce']))
plt.figure(figsize = (8, 4), facecolor = None)
plt.imshow(wordcloud)
plt.axis("off")
plt.tight_layout(pad = 0)
plt.show()
Merging Flavor and Ingredients¶
I merged each recipe id with its constituent ingredients and each ingredient's corresponding flavors.
# Import the full recipes dataset and isolate the ingredients in each recipe
pp_ingr = pd.read_csv("food.com_recipes/PP_recipes.csv")[['id', 'ingredient_ids']]
# Change the string of list in this column to a list of strings
pp_ingr['ingredient_ids'] = pp_ingr['ingredient_ids'].apply(json.loads)
def ingr_match_and_filter(ingr_ids):
# Given that the recipes data only has ingredient ids, we want to match those ids to the ingredient string
try:
ingr_list = [ingr_dict[i] for i in ingr_ids]
except KeyError:
ingr_list = None
return ingr_list
pp_ingr['ingredients'] = pp_ingr['ingredient_ids'].apply(ingr_match_and_filter)
def flav_match(ingredients):
# Use the ing flav dict constructed before to select the flavors for each ingredient in a recipe
flav_list = []
for i in ingredients:
try:
flav_list_i = ing_flav_dict[i]
except KeyError:
continue
if len(flav_list_i) >0:
flav_list.extend(flav_list_i)
# De-duplicate
flav_list = list(set(flav_list))
return flav_list
def literal_return(val):
try:
return literal_eval(val)
except (ValueError, SyntaxError) as e:
return val
pp_ingr = pp_ingr[pp_ingr['ingredient_ids'].isna() == False]
pp_ingr['ingredients'] = pp_ingr['ingredient_ids'].apply(ingr_match_and_filter)
pp_ingr = pp_ingr[pp_ingr['ingredients'].isna() == False]
pp_ingr['flavors'] = pp_ingr['ingredients'].apply(flav_match)
pp_ingr = pp_ingr.rename(columns={'id':'recipe_id'})
pp_ingr[:10]
Below, as an example of the flavors, I plot 50 randomly chosen flavors and their counts across the recipes
pp_ingr['flavors'] = pp_ingr['flavors'].apply(literal_return)
# Make each flavor it's own row and aggregate by the number of mentions
flav_explode = pp_ingr.explode('flavors')
flav_explode = flav_explode.groupby("flavors").agg(mentions = ('recipe_id', 'count')).reset_index()
# Plot 50 random flavors and their counts across the recipes
sns.set_theme(style="whitegrid")
f, ax = plt.subplots(figsize=(4, 4))
sns.set_color_codes("pastel")
sns.barplot(x="mentions", y="flavors", data=flav_explode.sort_values(by='mentions').sample(40),
label="Total mentions")
plt.yticks(fontsize=8)
sns.set_color_codes("muted")
Merging Cooking Techniques¶
Next, I associated with the technique ids included with each reciipe.
pp_recipes = pd.read_csv("food.com_recipes/PP_recipes.csv")
with open("data/techniques.txt", 'r') as f:
techniques = [line.strip() for line in f]
# Associate each yes/no indication in the techniques column with
# a technique in the techniques list, and add the technique to the recipe if it is associated with a 1
pp_recipes = pp_recipes.rename(columns={'techniques':'technique_onehot'})
pp_recipes['techniques'] = pp_recipes['technique_onehot'].apply(
lambda x: [ing for (ing, onehot) in zip(techniques, json.loads(x)) if onehot ==1])
pp_techniques = pp_recipes[['id', 'techniques']]
pp_techniques = pp_techniques.rename(columns={'id': 'recipe_id'})
pp_techniques[:10]
Preprocessing Interactions Data¶
Now it's time to start working with the user interactions data, the main data source for Collaborative Filtering approaches. First of all, I combined the training the testing set that were given on Kaggle because I wanted to do my own train-test split. Let's start by visualizing the spread of ratings and the number of ratings for each recipe.
interactions_train = pd.read_csv("food.com_recipes/interactions_train.csv")
interactions_test = pd.read_csv("food.com_recipes/interactions_test.csv")
interactions = pd.concat([interactions_train, interactions_test])
# Calculate mean rating and number of ratings for each recipe
agg = interactions.groupby(["recipe_id", 'i']).agg(mean_rating = ('rating', 'mean'), number_of_ratings = ('rating', 'count'))
def plot_ratings_dist(agg):
fig = plt.figure(figsize=(10, 5))
fig_rows = 1
fig_columns = 2
fig.add_subplot(fig_rows, fig_columns, 1)
plt.hist(agg['mean_rating'], bins=10, ec='white', fc='purple')
plt.title('Dist of mean rating for recipes')
fig.add_subplot(fig_rows, fig_columns, 2)
plt.hist(agg['number_of_ratings'], bins=1000, ec='white', fc='purple')
plt.xlim(0,50)
plt.title('Dist of number of ratings for recipes')
plt.show()
plot_ratings_dist(agg)
Train-test Split¶
I performed a train-test split by filtering down the data on whether the recipe had at least 10 ratings and then randomly selecting one rating for each of the remaining recipes to be part of the test set. The idea here was to make sure each recipe was represented in both the train and the test sets. See the sizes of the testing and training sets below:
interactions_train = pd.read_csv("data/interactions_train_mm.csv", index_col=0)
interactions_test = pd.read_csv("data/interactions_test_mm.csv", index_col=0)
print(f"Size of training data: {interactions_train.shape}, size of testing data: {interactions_test.shape}")
Final Merging of Recipe Details and Interactions¶
The interactions data were merged with the user interactions data such that the rating for each recipe could also be associated with each detail of that recipe. At this point, each recipe "detail\" was collated and given a detail id (ingredients, flavors, and techniques). For example, when averaging the rating across all the recipes that has a particular ingredient, you would get the average rating for that ingredient across the whole data set.
def add_deets_to_recipe(interactions, techniques, ingredients):
# Merge interactions with techniques and ingredients
interactions_w_deets = interactions.merge(techniques, how='left', on='recipe_id').merge(ingredients, how='left', on='recipe_id').drop(columns=['ingredient_ids'])
# Turn into lists and combine
interactions_w_deets['ingredients'] = interactions_w_deets['ingredients'].apply(literal_return)
interactions_w_deets['techniques'] = interactions_w_deets['techniques'].apply(literal_return)
interactions_w_deets['flavors'] = interactions_w_deets['flavors'].apply(literal_return)
interactions_w_deets['deets'] = interactions_w_deets['ingredients'] + interactions_w_deets['techniques'] + interactions_w_deets['flavors']
# Explode deets such that each row is a different detail
interactions_w_deets = interactions_w_deets.explode('deets')
return interactions_w_deets
interactions_test_w_deets = pd.read_csv("data/interactions_test_w_deets.csv", index_col=0)
interactions_test_w_deets[interactions_test_w_deets['deets'].isna() == False][:20]
Recommendation Algorithms¶
Now it's time to start the prediction of ratings of new user-recipe pairs. First, I defined some functions to calculate evaluation metrics and plot the predictions against the true ratings in the test set. While the implementation details of the algorithms vary slightly, the overall idea is that each was trained using the training set and then predictions were made on the user-recipe pairs in the testing set so that the performance of each model could be evaluated. I used Root Mean Square Error (RMSE), Mean Absolute Error (MAE) and coverage (ability to generate predictions) [6] to evaluate the performance of each approach.
from sklearn.metrics import mean_squared_error, mean_absolute_error, r2_score
import numpy as np
%matplotlib inline
def calculate_metrics(results, true_total, model_name):
RMSE = np.sqrt(mean_squared_error(results['rating'], results['rating_pred']))
MAE = mean_absolute_error(results['rating'], results['rating_pred'])
coverage = len(results) / true_total
print(f"{model_name} : RMSE: {RMSE} MAE: {MAE} Coverage: {coverage}")
def plot_results(results, model_name):
f, ax = plt.subplots(figsize=(5, 4))
sns.color_palette("crest")
g = sns.boxplot(data=results, x='rating', y='rating_pred', palette='crest')
g.set_title(f"{model_name} Results")
plt.show()
0. Baseline algorithm¶
The baseline model randomly predicts ratings for user-recipe pairs by sampling from a uniform distribution
# For future use of sparse matrices, add 1 to all ratings to that the range is 1-6 instead of 0-5
interactions_train['rating'] += 1
interactions_test['rating'] += 1
# Random choice rom a uniform distribution between 1 and 6
rating_pred = np.random.uniform(1, 7, len(interactions_test))
BL = interactions_test[['u', 'i', 'rating']]
BL.loc[:, 'rating_pred'] = rating_pred
# Calculate the true total for the coverage metric
true_total = len(interactions_test[['u', 'i']].drop_duplicates())
# Calculate metrics and plot results of baseline model
calculate_metrics(BL, true_total, "BL")
plot_results(BL, "BL")
1. Vanilla Collaborative Filtering¶
For this model, I compared results for the most similar 10 users, most similar 50 users, and most similar 100 users.
Below is the code for generting predictions for new user-recipe pairs by averaging the ratings of the most similar 50 users.
import scipy.sparse as scsp
from sklearn.metrics.pairwise import cosine_similarity
# create user-recipe matrix
M = scsp.csr_matrix((interactions_train['rating'], (interactions_train['u'], interactions_train['i'])))
# Calculate user level rating average
user_average = np.array(M.sum(axis=1) / np.diff(M.indptr).reshape(-1,1))
# Create rating average sparse matrix using the locations of the nonzero elements of interactions matrix
nz = M.nonzero()
user_average = scsp.csr_matrix((user_average[nz[0]].T[0], nz))
# Calculate normalized ratings matrix
M_norm = M - user_average
# Calculate similarity between users
sim = cosine_similarity(M_norm)
# Initiate dataframe for evaluation
CF = interactions_test[['u', 'i', 'rating']]
CF.loc[:, 'rating_pred'] = np.nan
# For each user in the test set, calculate a prediction for each possible recipe in the test set
for u in interactions_test['u'].unique():
# get similarities for that user and delete their similarity with themself
simt = sim[u,:]
simt = np.delete(simt, u)
# get top 50 most similar users
most_similar_users = np.argpartition(simt, -50)[-50:]
# get the ratings for those most similar users
similar_user_ratings = M_norm[most_similar_users]
# calculate weighted score for each recipe (score times similarity)
score = scsp.csc_matrix(similar_user_ratings.multiply(simt[most_similar_users].reshape(-1,1)) )
# calculate average score for each recipe
col_totals = score.sum(axis=0)
col_counts = np.diff(score.indptr).astype(float)
col_counts[np.where(col_counts == 0)] = np.NAN
score_average = np.array(col_totals / col_counts)[0]
score_average[np.where(np.isnan(score_average) == True)] = 0
# Add in user average
score_prediction = score_average + np.max(user_average[u])
CF.loc[CF['u']==u, 'rating_pred'] = score_prediction[CF[CF['u'] == u]['i']]
CF = CF[CF['rating_pred'].isna() == False].drop_duplicates()
calculate_metrics(CF, true_total, "CF")
plot_results(CF, "CF")
2. Content-based¶
Below is the content based approach of generating predictions for new user-recipe pairs by averaging the ratings of that user on constituent ingredients, flavors, and techniques of the new recipe.
# Average the rating of each user on each detail across all recipes
interactions_train_w_deets = pd.read_csv("data/interactions_train_w_deets.csv")[['u', 'deets', 'rating']].drop_duplicates().reset_index(drop=True)
interactions_train_deets = interactions_train_w_deets.groupby(['deets', 'u']).mean().reset_index()
interactions_train_deets = interactions_train_deets[interactions_train_deets['deets'].str.len() > 0]
# Merge training ratings of details with testing true ratings of recipes
interactions_test_w_deets = pd.read_csv("data/interactions_test_w_deets.csv")[['u', 'i', 'deets', 'rating']].drop_duplicates().reset_index(drop=True)
pred_df = interactions_test_w_deets[['u', 'i', 'deets']].merge(interactions_train_deets[['u', 'deets', 'rating']], on=['u', 'deets'])
# Group on recipes to get average of detail ratings as predicted recipe rating
pred_df = pred_df.groupby(['u', 'i'])['rating'].mean().reset_index()
pred_df.columns = ['u', 'i', 'rating_pred']
# Merge back with testing data for evaluation
CB = interactions_test_w_deets[['u', 'i', 'rating']].merge(pred_df, how='inner', on=['i', 'u'])
CB = CB.drop_duplicates()
calculate_metrics(CB, true_total, "CB")
plot_results(CB, "CB")
3. Matrix Factorization for Collaborative Filtering¶
Below is a Non-negative matrix factorization approach (NMF) for matrix factorization that if often used for factoring sparse matrices. I based this class on the NMF class in the Surprise Python package and uses Surprise-style datasets, which have certain properties like the number of users and ratings.
class NMF():
# Matrix Factorization of a sparse matrix using NMF.
# Fitt by perfoming SGD on the Q and P matrices defined above in the notes.
# Loss function is MSE between the actual matrix and the product of Q and P.
# Predict by multiplying the resulting the corresponding entries of the Q and P matrices together.
def __init__(self, n_factors=15, n_epochs=50, reg_pu=.06,
reg_qi=.06, random_state=None):
# params:
# n_factors: the number of latent factors to decompose matrix into (the number of rows in in P and the number of columns in Q)
# n_epochs: the number of iterations of SGD
# reg_pu and reg_qi: the regularization terms to apply to the SGD updates
# random_state: optional
self.n_factors = n_factors
self.n_epochs = n_epochs
self.reg_pu = reg_pu
self.reg_qi = reg_qi
self.random_state = random_state
def fit(self, trainset):
# trainset: n x m matrix of ratings
self.trainset = trainset
self.sgd(trainset)
return self
def sgd(self, trainset):
# user and item factors
pu = np.random.uniform(0, 0.5, size=(trainset.n_users, self.n_factors))
qi = np.random.uniform(0, 0.5, size=(trainset.n_items, self.n_factors))
n_factors = self.n_factors
reg_pu = self.reg_pu
reg_qi = self.reg_qi
# auxiliary matrices used in optimization process
user_num = np.zeros((trainset.n_users, n_factors))
user_denom = np.zeros((trainset.n_users, n_factors))
item_num = np.zeros((trainset.n_items, n_factors))
item_denom = np.zeros((trainset.n_items, n_factors))
self.losses = []
for current_epoch in range(self.n_epochs):
print("Processing epoch {}".format(current_epoch))
user_num[:, :] = 0
user_denom[:, :] = 0
item_num[:, :] = 0
item_denom[:, :] = 0
epoch_loss = 0
l = 0
# Compute numerators and denominators for users and items factors
for u, i, r in trainset.all_ratings():
# compute current estimation and error
dot = 0 # <q_i, p_u>
for f in range(n_factors):
dot += qi[i, f] * pu[u, f]
est = dot
err = r - est
epoch_loss += err**2
l += 1
# compute numerators and denominators
for f in range(n_factors):
user_num[u, f] += qi[i, f] * r
user_denom[u, f] += qi[i, f] * est
item_num[i, f] += pu[u, f] * r
item_denom[i, f] += pu[u, f] * est
print(epoch_loss, l)
epoch_loss = epoch_loss / l
print(epoch_loss)
self.losses.append(epoch_loss)
# Update user factors
for u in trainset.all_users():
n_ratings = len(trainset.ur[u])
for f in range(n_factors):
if pu[u, f] != 0: # Can happen if user only has 0 ratings
user_denom[u, f] += n_ratings * reg_pu * pu[u, f]
pu[u, f] *= user_num[u, f] / user_denom[u, f]
# Update item factors
for i in trainset.all_items():
n_ratings = len(trainset.ir[i])
for f in range(n_factors):
if qi[i, f] != 0:
item_denom[i, f] += n_ratings * reg_qi * qi[i, f]
qi[i, f] *= item_num[i, f] / item_denom[i, f]
self.pu = np.asarray(pu)
self.qi = np.asarray(qi)
def predict(self, u, i):
known_user = self.trainset.knows_user(u)
known_item = self.trainset.knows_item(i)
if known_user and known_item:
est = np.dot(self.qi[i], self.pu[u])
else:
est = None
return est
I chose to perform the factorization into 40 latent factors because this number is a common default for recommendation packages in python. I decided to go with this approach rather choosing this number through cross-validation because I wanted to focus my time for this project more on exploring different algorithms rather than tuning this particular approach. I initialized the $P$ and $Q$ matrices as uniform random matrices between 0 and 1. The shapes of these matrices are:
$P$: number of users by number of features (40).
$Q$: number of features (40) by number of recipes.
I trained the model by performing Gradient Descent for 100 iterations according to the matrix updates in the notes. Below is the training loss across iterations and the distribution of testing predictions versus true ratings.
MF = pd.read_csv("results/MF.csv", index_col=0)
calculate_metrics(MF, true_total, "Matrix Factorization")
plot_results(MF, "Matrix Factorization")
4. Content-augmented Collaborative Filtering¶
For this approach, I made rating predictions for each new user-recipe pair by first selecting the 50 most similar users as determined by cosine similarity of their overall recipe ratings(like the vanilla CF approach), then calculating a weighted average of each similar user's rating on each ingredient/flavor/technique to predict the new user's rating on each of those itmes, and then averaging the ratings of all ingredients/flavors/techniques in a recipe.
First, I needed to create a unique index for the ingredients/flavors/techniques to use in matrix calculations.
# Create master list of all detail
deets = pd.concat([interactions_train_w_deets['deets'], interactions_test_w_deets['deets']]).drop_duplicates().reset_index(drop=True)
# Create idx for each detail
deet_id_map = pd.DataFrame(deets).reset_index().rename(columns={'index':'d'})
# Merge idx for each detail back into train and test detail data sets
interactions_train_w_deets = interactions_train_w_deets.merge(deet_id_map, how='left', on='deets')
interactions_test_w_deets = interactions_test_w_deets.merge(deet_id_map, how='left', on='deets')
# Find average user rating for each detail
interactions_train_deets = interactions_train_w_deets.drop(columns=['deets']).groupby(['u', 'd']).mean().reset_index()
interactions_train_deets
Next, create the user-detail matrix. For each user in the interactions test data, find the 50 most similar users using the previously defined user-recipe rating matrix. For those 50 users, average their ratings on each detail to predict the ratings of this test user on each detail. Finally, average those detail ratings grouped by recipe.
# Create user-detail rating matrix
num_users = np.max(interactions_train['u']) + 1
num_deets = deet_id_map['d'].iloc[-1] + 1
A = scsp.csr_matrix((interactions_train_deets['rating'], (interactions_train_deets['u'], interactions_train_deets['d'])), shape=(num_users, num_deets))
# Calculate similarity matrix
sim = cosine_similarity(M_norm, dense_output=False)
# Initialize data frame
CA_CF = pd.DataFrame()
# For each user in the test set, calculate a prediction for each possible recipe in the test set
for u in interactions_test_w_deets['u'].unique():
# get similarities for that user and delete their similarity with themself
simt = sim[u,:].toarray()[0]
simt = np.delete(simt, u)
# get top 50 most similar users
most_similar_users = np.argpartition(simt, -50)[-50:]
# get the ratings for those most similar users
similar_user_deet_ratings = scsp.csc_matrix(A[most_similar_users])
# calculate average rating of all 50 similar users on each detail
col_counts = np.diff(similar_user_deet_ratings.indptr).astype(float)
col_counts[np.where(col_counts == 0)] = np.NAN
deet_pred = similar_user_deet_ratings.sum(axis=0) / col_counts
deet_pred = pd.DataFrame(deet_pred.T).reset_index().rename(columns={'index': 'd', 0: 'deet_rating_pred'})
# For this user, find average of all detail ratings grouped by recipe
pred_df = interactions_test_w_deets[interactions_test_w_deets['u'] == u]
pred_df = pred_df.merge(deet_pred, how='left', on='d')
pred_df = pred_df.groupby(['u', 'i'])['deet_rating_pred'].mean().reset_index().rename(columns={'deet_rating_pred':'rating_pred'})
CA_CF = pd.concat([CA_CF, pred_df])
# Merge back to test data to get true ratings
CA_CF = CA_CF.merge(interactions_test_w_deets[['u', 'i', 'rating']].drop_duplicates(), how='left', on=['u', 'i'])
CA_CF = CA_CF[CA_CF['rating_pred'].isna() == False].drop_duplicates()
calculate_metrics(CA_CF, true_total, "Content-Augmented CF")
plot_results(CA_CF, "Content-Augmented CF")
Here, we again face the issue of there being very little variation in the predictions across all the true ratings. Unfortunately it appears that the main effect of including recipe-content information in the CF approach had the effect of bringing down the average rating by about 0.5 without doing much else. My guess as to why this is happening is that in this sparse data set, most users may have only rated a couple of recipes and therefore very little in that user across recipes. Extending the rating of this recipe to it's content data would not actually give a more varied look at the user's preferences, and might have teh effect of diluting the variation in the data set more by increasing the number of observations with the same associated rating.
5. Content-augmented Matrix Factorization¶
For this algorithm, I used a smaller dataset that I created to be 1/10 the size of the original dataset. Additionally, I used only ingredients rather than ingredients, flavors, and techniques. I did this because the training process was taking a very long time and I wanted to be able to iterate quicker to understand the SGD process of this algorithm better.
interactions_train = pd.read_csv("data/interactions_train_small_mm.csv")
interactions_test = pd.read_csv("data/interactions_test_small_mm.csv")
pp_recipes = pd.read_csv("data/pp_recipes.csv")
ingr = pd.read_csv("data/pp_ingr.csv", index_col=0)
ingr = ingr.rename(columns={'id':'recipe_id'})
def add_ingr_to_recipe(interactions, ingredients):
interactions_w_deets = interactions.merge(ingredients, how='left', on='recipe_id').drop(columns=['ingredient_ids'])
interactions_w_deets['ingredients'] = interactions_w_deets['ingredients'].apply(literal_return)
interactions_w_deets = interactions_w_deets.explode('ingredients')
return interactions_w_deets
# Add details to training data
interactions_train_w_deets = add_ingr_to_recipe(interactions_train, ingr)
interactions_train_w_deets = interactions_train_w_deets[['user_id', 'recipe_id', 'ingredients', 'rating', 'u', 'i']].drop_duplicates().reset_index(drop=True)
# Add details to testing data
interactions_test_w_deets = add_ingr_to_recipe(interactions_test, ingr)
interactions_test_w_deets = interactions_test_w_deets[['user_id', 'recipe_id', 'ingredients', 'rating', 'u', 'i']].drop_duplicates().reset_index(drop=True)
# Combine test and training sets
interactions_w_deets = pd.concat([interactions_train_w_deets, interactions_test_w_deets])
Next, I started the MF process by creating the recipe-content matrix $X$, which is a factorization of the matrix $Q$ in the $R = PQ^T$ factorization. $X$ is a matrix in which $X_{id}$ is a binary indicator if ingredient $d$ is in recipe $i$.
# Create full list of ingredients and find unique id for each
ingredients = interactions_w_deets['ingredients'].drop_duplicates().reset_index(drop=True)
ingredients_id_map = pd.DataFrame(ingredients).reset_index().rename(columns={'index':'d'})
recipe_ingredients = interactions_w_deets.merge(ingredients_id_map, how='left', on=['ingredients'])[['d', 'i']].drop_duplicates()
# X indicates if recipes (rows) contain ingredients (columns)
X = scsp.csr_matrix((np.ones((len(recipe_ingredients))), (recipe_ingredients['i'], recipe_ingredients['d'])))
Next, I built a Pytorch model to perform SGD to find the factorization $\hat R = P\Phi^TX^T$. First the model initiates the matrices $P$ and $\Phi$ as random uniform distributions betwen 0 and 0.5. The three matrices had the following shapes:
- $P$: number of users by number of features.
- $\Phi$: number of features by number of recipe contents.
- $X$: number of recipe contents by number of recipes.
Next, the model calculates the MSE loss between $R$ and the factorization of $R$, and finally updates the $P$ and $\Phi$ along the gradients.
import torch
class PytorchLinearModel(torch.nn.Module):
def __init__(self, num_users, num_ingr, K, lr):
super().__init__()
# Set U and V as parameters of this model
self.P = torch.nn.Parameter(torch.zeros((num_users, K)))
self.PHI = torch.nn.Parameter(torch.zeros((num_ingr, K)))
# Xavier initialization is a great way to intialize parameters
# torch.nn.init.xavier_uniform_(self.U)
# torch.nn.init.xavier_uniform_(self.PHI)
torch.nn.init.uniform_(self.P, 0, 0.5)
torch.nn.init.uniform_(self.PHI, 0, 0.5)
# MSE using Pytorch
self.MSE = torch.nn.MSELoss()
# Optimizer for handling the gradient calculation and parameter updates
self.optimizer = torch.optim.SGD(self.parameters(), lr=lr)
def forward(self, X):
return self.UP @ self.PHI.T @ X.T
def calculate_loss(self, R, R_hat):
return self.MSE(R, R_hat)
def optimize(self, loss):
# Send in the loss tensor from `calculate_loss` to update parameters
self.optimizer.zero_grad() # Clear any previous epoch gradients
loss.backward() # Calc gradient
self.optimizer.step() # Update parameters
num_ingr = ingredients_id_map['d'].iloc[-1] + 1
MF_model = PytorchLinearModel(num_users, num_ingr, 40, 0.01)
As with previous factorization approach, I chose to use 40 latent factors and trained for 500 iterations. Below is the training loss across iterations and the distribution of testing predictions versus true ratings.
MF_hybrid = pd.read_csv("results/matrix_fact_hybrid.csv", index_col=0)
calculate_metrics(MF_hybrid, true_total, "Content-Augmented Matrix Factorization")
plot_results(MF_hybrid, "Content-Augmented Matrix Factorization")
Summary of Findings¶
Below is a comparison table of the performance of all the algorithms above. In summary, I did not find that including content-related recipe such as ingredients, flavors, and cooking techniques universally improved the quality of user-recipe pair rating predictions. Overall, the best performance was actually found from vanilla CF, despite the fact that matrix factorization is commonly considered to be the most performant CF approach. I imagine that if I had the computational resources to train a matrix factorization model for longer and perform a complete hyperparameter tuning, would get better results for that model. Compared to pure CB, CF approach did result in more accurate predictions, and combining content with vanilla CF improved the metrics in comparison with pure CB. | Model | RMSE | MAE | Coverage | | ---------------------------------------| -------| --------| ----------| | Baseline | 2.5185| 2.04112| 1.0| | Vanilla CF | 0.9805| 0.5617| 1.0| | Matrix factorization CF | 1.7859| 1.4959| 0.6837| | CB | 1.3987| 1.0338| 0.3343| | Content-augmented CF | 1.0787| 0.8480| 0.9995| | Content-augmented matrix factorization | 4.1741| 3.8136| 1.1|
For future experimentation, I would want to investigate the extent of overlap of recipe content (ingredients, etc.) across recipes. It's possible that if there is a large amount of overlap, the result of including more content data is actually diluting the personalized information in the data set. An additional avenue for experimentation is the distribution of ratings in the training data. While I'm sure that this skewed distribution had some effect on the performance of these algorithms, re-sampling the data so that the ratings were normally or uniformly distributed would result in a data set that is not representative of real-world rating systems where these algorithms would be implemented.
References¶
Trang Tran, T. N., Atas, M., Felfernig, A., & Stettinger, M. (2018). An overview of recommender systems in the healthy food domain. Journal of Intelligent Information Systems, 50 (pp. 501-526).
Pecune, F., Callebert, L., & Marsella, S. (2020, September). A Recommender System for Healthy and Personalized Recipes Recommendations. In HealthRecSys@ RecSys (pp. 15-20).
Van Pinxteren, Y., Geleijnse, G., & Kamsteeg, P. (2011, February). Deriving a recipe similarity measure for recommending healthful meals. In Proceedings of the 16th international conference on Intelligent user interfaces (pp. 105-114).
Masthoff, J. (2011). Group recommender systems: Combining individual models. Recommender systems handbook, Springer (pp. 677--702).
Ajitsaria, A. Build a Recommendation Enging with Collaborate Filtering. RealPython.com
Freyne, J., & Berkovsky, S. (2010, February). Intelligent food planning: personalized recipe recommendation. In Proceedings of the 15th international conference on Intelligent user interfaces (pp. 321-324).
Burke, R. (2002). Hybrid Recommender Systems: Survey and Experiments. User Model User-Adap Inter 12, (pp. 331--370).
Aberg, J. (2006, January). Dealing with Malnutrition: A Meal Planning System for Elderly. In AAAI spring symposium: argumentation for consumers of healthcare (pp. 1-7).
Luo, S. (2018, December). Introduction to Recommender System Approaches of Collaborative Filtering: Nearest Neighborhood and Matrix Factorization. towardsdatascience.com
Forbes, P., & Zhu, M. (2011, October). Content-boosted matrix factorization for recommender systems: experiments with recipe recommendation. In Proceedings of the fifth ACM conference on Recommender systems (pp. 261-264).
X. Luo, M. Zhou, Y. Xia and Q. Zhu. (May 2014). An Efficient Non-Negative Matrix-Factorization-Based Approach to Collaborative Filtering for Recommender Systems,\" in IEEE Transactions on Industrial Informatics, vol. 10, no. 2 (pp. 1273-1284)