Collaborative Filtering Applied to Educational Data Mining

Similar documents
CS246: Mining Massive Datasets Jure Leskovec, Stanford University

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

An Empirical Comparison of Collaborative Filtering Approaches on Netflix Data

Recommender Systems New Approaches with Netflix Dataset

CptS 570 Machine Learning Project: Netflix Competition. Parisa Rashidi Vikramaditya Jakkula. Team: MLSurvivors. Wednesday, December 12, 2007

Reddit Recommendation System Daniel Poon, Yu Wu, David (Qifan) Zhang CS229, Stanford University December 11 th, 2011

COMP 465: Data Mining Recommender Systems

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

Yelp Recommendation System

Data Mining Techniques

The exam is closed book, closed notes except your one-page (two-sided) cheat sheet.

Data Mining Techniques

Use of KNN for the Netflix Prize Ted Hong, Dimitris Tsamis Stanford University

Additive Regression Applied to a Large-Scale Collaborative Filtering Problem

Ensemble Learning and the Heritage Health Prize

Recommendation System Using Yelp Data CS 229 Machine Learning Jia Le Xu, Yingran Xu

Performance Comparison of Algorithms for Movie Rating Estimation

THE goal of a recommender system is to make predictions

A Dendrogram. Bioinformatics (Lec 17)

The exam is closed book, closed notes except your one-page (two-sided) cheat sheet.

Recommender System. What is it? How to build it? Challenges. R package: recommenderlab

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

On hybrid modular recommendation systems for video streaming

Machine Learning Classifiers and Boosting

The exam is closed book, closed notes except your one-page cheat sheet.

Collaborative Filtering for Netflix

A Linear Ensemble of Individual and Blended Models for Music Rating Prediction

Seminar Collaborative Filtering. KDD Cup. Ziawasch Abedjan, Arvid Heise, Felix Naumann

Lecture 7: Linear Regression (continued)

Using Social Networks to Improve Movie Rating Predictions

Recommendation Systems

CS294-1 Assignment 2 Report

Neural Network and Deep Learning. Donglin Zeng, Department of Biostatistics, University of North Carolina

Classification of Imbalanced Marketing Data with Balanced Random Sets

Advances in Collaborative Filtering

Advances in Collaborative Filtering

10-701/15-781, Fall 2006, Final

Recommender System Optimization through Collaborative Filtering

Machine Learning. Chao Lan

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

Ranking by Alternating SVM and Factorization Machine

CS 124/LINGUIST 180 From Languages to Information

Self-Organizing Map. presentation by Andreas Töscher. 19. May 2008

Neural Network Weight Selection Using Genetic Algorithms

CSE 158 Lecture 8. Web Mining and Recommender Systems. Extensions of latent-factor models, (and more on the Netflix prize)

Clustering-Based Personalization

Spatial Variation of Sea-Level Sea level reconstruction

Multi-label Classification. Jingzhou Liu Dec

CS570: Introduction to Data Mining

Akarsh Pokkunuru EECS Department Contractive Auto-Encoders: Explicit Invariance During Feature Extraction

Perceptron as a graph

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

Collaborative Filtering via Euclidean Embedding

ECS289: Scalable Machine Learning

Scaled Machine Learning at Matroid

Extension Study on Item-Based P-Tree Collaborative Filtering Algorithm for Netflix Prize

Predicting User Ratings Using Status Models on Amazon.com

Deep Learning and Its Applications

Lecture 26: Missing data

DS Machine Learning and Data Mining I. Alina Oprea Associate Professor, CCIS Northeastern University

BPR: Bayesian Personalized Ranking from Implicit Feedback

Machine Learning in Action

Variational Bayesian PCA versus k-nn on a Very Sparse Reddit Voting Dataset

Instance-based Learning

CS6220: DATA MINING TECHNIQUES

BINARY PRINCIPAL COMPONENT ANALYSIS IN THE NETFLIX COLLABORATIVE FILTERING TASK

Hybrid Recommendation Models for Binary User Preference Prediction Problem

BordaRank: A Ranking Aggregation Based Approach to Collaborative Filtering

CSE 258 Lecture 8. Web Mining and Recommender Systems. Extensions of latent-factor models, (and more on the Netflix prize)

CS 229 Final Project - Using machine learning to enhance a collaborative filtering recommendation system for Yelp

Hotel Recommendation Based on Hybrid Model

A Recommender System Based on Improvised K- Means Clustering Algorithm

Network Traffic Measurements and Analysis

General Instructions. Questions

Collaborative Filtering and Recommender Systems. Definitions. .. Spring 2009 CSC 466: Knowledge Discovery from Data Alexander Dekhtyar..

DS Machine Learning and Data Mining I. Alina Oprea Associate Professor, CCIS Northeastern University

Deep Generative Models Variational Autoencoders

Recommendation and Advertising. Shannon Quinn (with thanks to J. Leskovec, A. Rajaraman, and J. Ullman of Stanford University)

Preface to the Second Edition. Preface to the First Edition. 1 Introduction 1

Otto Group Product Classification Challenge

Personalize Movie Recommendation System CS 229 Project Final Writeup

CPSC 340: Machine Learning and Data Mining. Recommender Systems Fall 2017

291 Programming Assignment #3

Mining of Massive Datasets Jure Leskovec, Anand Rajaraman, Jeff Ullman Stanford University Infinite data. Filtering data streams

Nonparametric Importance Sampling for Big Data

Recommendation System for Netflix

HMC CS 158, Fall 2017 Problem Set 3 Programming: Regularized Polynomial Regression

CS246: Mining Massive Datasets Jure Leskovec, Stanford University.

Online Social Networks and Media

CS273 Midterm Exam Introduction to Machine Learning: Winter 2015 Tuesday February 10th, 2014

Opportunities and challenges in personalization of online hotel search

CS224W Project: Recommendation System Models in Product Rating Predictions

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

CS 224d: Assignment #1

Network embedding. Cheng Zheng

Regularized Committee of Extreme Learning Machine for Regression Problems

Lecture 13: Model selection and regularization

CS 124/LINGUIST 180 From Languages to Information

Web Personalization & Recommender Systems

Transcription:

Collaborative Filtering Applied to Educational Data Mining KDD Cup 200 July 25 th, 200

BigChaos @ KDD Team Dataset Solution Overview Michael Jahrer, Andreas Töscher from commendo research

Dataset Team Dataset Solution Overview Algebra 2008-2009 (9 million lines) Bridge to Algebra 2008-2009 (20 million lines) Predict correct first attempt Similarities to collaborative filtering (CF) KDD Cup Student CF User 5 5 Step 0 0 0 Item 4 2 3 2

Used Information Team Dataset Solution Overview Student and step Unit and section parsed from the provided hierarchy string Problem name Knowledge components (KC) Opportunity count Problem view

Solution Overview Team Dataset Solution Overview K Nearest Neighbor (KNN) Singular Value Decomposition (SVD) Factor Model (FM) Factor Model 2 (FM2) Factor Model 3 (FM3) Group Factor Model (GFM) Neural Network for blending 8-fold cross validation

KNN SVD FM

K Nearest Neighbor (KNN) KNN SVD FM Prediction is based on students with similar historic results Pearson correlation between students Use the K most similar students for prediction Problems Not all student correlations are defined with equal quality, so we use a correlation shrinkage Some steps are answered by no other students

KNN SVD FM ρ s s 2 = I s s 2 ρ s s 2 I s s 2 + α σ(x) = + e x ρ s s 2 = σ (δ ρ s s 2 + γ) c is = ĉ is = c is s S i (s;k) ρ s sc i s s S i (s;k) ρ s s s S i (s;k) ρ s s + µ s β s S i (s;k) ρ s s + β I s s 2... set of steps commonly answered by student s and student s 2 ĉ is... prediction for student s on step i ρ s s 2... Pearson correlation between student s and student s 2 S i (s; K)... to student s the K most similar students who answered step i α, β, γ, δ... meta parameters

KNN results KNN SVD FM Dataset RMSE Meta parameters K = 4, α = 2.9, β =.5, Algebra 2008-2009 0.3257 δ = 6.2, γ =.9 K = 4, α = 2.9, β =.5, B to A 2008-2009 0.3049 δ = 6.2, γ =.9 We have not used additional information = room for improvement Using step correlations does not work well

KNN SVD FM Singular Value Decomposition (SVD) Sparse student by step matrix C = [c is ], containing the correct first attempt values Represent a student s by a N dimensional feature vector b s Represent a step i by a N dimensional feature vector a i ĉ is = a T i b s

Results KNN SVD FM Parameters trained by stochastic gradient descent with L2 regularization λ Does not model biases and additional information Dataset RMSE Meta parameters Algebra 2008-2009 0.4462 N = 0, η = 0.002, λ = 0.02 B to A 2008-2009 0.368 N = 0, η = 0.002, λ = 0.02 B to A 2008-2009 0.359 N = 20, η = 0.002, λ = 0.0 B to A 2008-2009 0.378 N = 20, η = 0.002, λ = 0.03

Factor Model (FM) KNN SVD FM ĉ is = µ + ˆµ i + µ s + µ p(i) + µ x(i) + µ u(i) + (ˇµ k + µ ks ) K(i, s) + a i + K(i, s) k K(i,s) k K(i,s) á k T b s µ... global bias ˆµ i... bias for step i µ s... bias for student s µ p(i)... bias for problem p(i) µ x(i)... bias for section x(i) µ u(i)... bias for unit u(i) K(i, s)... set of knowledge components á k... N dimensional KC feature vector

FM results KNN SVD FM Significant improvement over SVD Models: unit, section, knowledge component FM2, FM3 and GFM contain further improvements Dataset RMSE Meta parameters Algebra 2008-2009 0.3078 N = 50, η = 0.0005, λ = 0.0 B to A 2008-2009 0.303 N = 50, η = 0.0005, λ = 0.0

Summary Thanks

Summary Thanks 2 layer neural network 80 sigmoid units on the first and second hidden layer 36 predictors for Algebra 2008-2009 37 predictors for Bridge to Algebra 2008-2009 Additional information

Feature Generation Summary Thanks

Results Summary Thanks Best individual results: Cross validation Leaderboard Algebra 2008-2009 0.2997 0.2849 Bridge to Algebra 2008-2009 0.2898 0.2837 Neural network blending: Cross validation Leaderboard Algebra 2008-2009 0.280925 0.277355 Bridge to Algebra 2008-2009 0.28847 0.28073

Summary Summary Thanks Ideas from collaborative filtering are suitable for the KDD Cup 200 An ensemble of different models is a good approach for a competition For sure there is a way to build simpler models with similar or better performance

Summary Thanks Thanks for your attention commendo research & consulting GmbH Austria