ALGORITHMS FOR DICTIONARY LEARNING

Size: px
Start display at page:

Download "ALGORITHMS FOR DICTIONARY LEARNING"

Transcription

1 ALGORITHMS FOR DICTIONARY LEARNING ANKUR MOITRA MASSACHUSETTS INSTITUTE OF TECHNOLOGY

2 SPARSE REPRESENTATIONS

3 SPARSE REPRESENTATIONS Many data- types are sparse in an appropriately chosen basis:

4 SPARSE REPRESENTATIONS Many data- types are sparse in an appropriately chosen basis: b i e.g. images, signals, data (n p)

5 SPARSE REPRESENTATIONS Many data- types are sparse in an appropriately chosen basis: dicoonary (n m) A x i b i e.g. images, signals, representaoons (m p) data (n p)

6 SPARSE REPRESENTATIONS Many data- types are sparse in an appropriately chosen basis: dicoonary (n m) at most k non- zeros A x i b i e.g. images, signals, representaoons (m p) data (n p)

7 SPARSE REPRESENTATIONS Many data- types are sparse in an appropriately chosen basis: dicoonary (n m) at most k non- zeros A x i b i Dic&onary Learning: Can we learn A from examples? e.g. images, signals, representaoons (m p) data (n p)

8 APPLICATIONS OF DICTIONARY LEARNING a.k.a. sparse coding

9 APPLICATIONS OF DICTIONARY LEARNING a.k.a. sparse coding Signal Processing/Sta&s&cs: De- noising, edge- detecoon, super- resoluoon Block compression for images/video

10 APPLICATIONS OF DICTIONARY LEARNING a.k.a. sparse coding Signal Processing/Sta&s&cs: De- noising, edge- detecoon, super- resoluoon Block compression for images/video Machine Learning: Sparsity as a regularizer to prevent over- fi[ng Learned sparse reps. play a key role in deep- learning

11 APPLICATIONS OF DICTIONARY LEARNING a.k.a. sparse coding Signal Processing/Sta&s&cs: De- noising, edge- detecoon, super- resoluoon Block compression for images/video Machine Learning: Sparsity as a regularizer to prevent over- fi[ng Learned sparse reps. play a key role in deep- learning Computa&onal Neuroscience (Olshausen- Field 1997): Applied to natural images yields filters with same qualitaove properoes as recepove field in V1

12 OUTLINE Are there efficient algorithms for dicoonary learning? Introduc&on Origins of Sparse Recovery A StochasOc Model; Our Results Provable Algorithms via Overlapping Clustering Uncertainty Principles ReformulaOon as Overlapping Clustering Analyzing Alterna&ng Minimiza&on Gradient Descent on Non- Convex Fctns

13 ORIGINS OF SPARSE RECOVERY Donoho- Stark, Donoho- Huo, Gribonval- Nielsen, Donoho- Elad: dicoonary (n m) A Incoherence: μ = max i j A i, A j / n x = b at most k non- zeros

14 ORIGINS OF SPARSE RECOVERY Donoho- Stark, Donoho- Huo, Gribonval- Nielsen, Donoho- Elad: dicoonary (n m) A Incoherence: μ = max i j A i, A j / n x = b at most k non- zeros for spikes- and- sines μ = 1

15 ORIGINS OF SPARSE RECOVERY Donoho- Stark, Donoho- Huo, Gribonval- Nielsen, Donoho- Elad: If k n / 2μ then x is the sparsest soluoon to the linear system, and can be found with l 1 - minimizaoon dicoonary (n m) A Incoherence: μ = max i j A i, A j / n x = b at most k non- zeros for spikes- and- sines μ = 1

16 THE FULL RANK CASE Are there efficient algorithms for dicoonary learning? Case #1: A has full column rank

17 THE FULL RANK CASE Are there efficient algorithms for dicoonary learning? Case #1: A has full column rank Theorem [Spielman, Wang, Wright 13]: There is a poly. Ome algorithm to exactly learn A when it has full column rank, for k n (hence m n)

18 THE FULL RANK CASE Are there efficient algorithms for dicoonary learning? Case #1: A has full column rank Theorem [Spielman, Wang, Wright 13]: There is a poly. Ome algorithm to exactly learn A when it has full column rank, for k n (hence m n) Approach: find the rows of A - 1, using L 1 - minimizaoon

19 THE FULL RANK CASE Are there efficient algorithms for dicoonary learning? Case #1: A has full column rank Theorem [Spielman, Wang, Wright 13]: There is a poly. Ome algorithm to exactly learn A when it has full column rank, for k n (hence m n) Approach: find the rows of A - 1, using L 1 - minimizaoon Stochas&c Model: unknown dicoonary A generate x with support of size k u.a.r., choose non- zero values independently, observe b = Ax

20 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely

21 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x)

22 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x) Claim: The sparsest vectors in row- span(x) (or B) are the X

23 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x) Claim: The sparsest vectors in row- span(x) (or B) are the X Can we find the sparsest vector in row- span(x)?

24 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x) Claim: The sparsest vectors in row- span(x) (or B) are the X Can we find the sparsest vector in row- span(x)? Approach #1: (P0): min w T B 0 s.t. w 0

25 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x) Claim: The sparsest vectors in row- span(x) (or B) are the X Can we find the sparsest vector in row- span(x)? Approach #1: NP- hard (P0): min w T B 1 s.t. w 0

26 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x) Claim: The sparsest vectors in row- span(x) (or B) are the X Can we find the sparsest vector in row- span(x)? Approach #2: L 1 - relaxa&on (P1): min w T B 1 s.t. w T r = 1 where we will set r later

27 (P1): min w T B 1 s.t. w T r = 1

28 (P1): min w T B 1 s.t. w T r = 1 Consider the bijecoon z = A T w, and set c = A - 1 r.

29 (P1): min w T B 1 s.t. w T r = 1 Consider the bijecoon z = A T w, and set r = Ac. We get: (P1): min w T AX 1 s.t. w T Ac = 1

30 (P1): min w T B 1 s.t. w T r = 1 Consider the bijecoon z = A T w, and set r = Ac. We get: (P1): min w T AX 1 s.t. w T Ac = 1 This is equivalent to: (Q1): min z T X 1 s.t. z T c = 1

31 (P1): min w T B 1 s.t. w T r = 1 Consider the bijecoon z = A T w, and set r = Ac. We get: (P1): min w T AX 1 s.t. w T Ac = 1 This is equivalent to: (Q1): min z T X 1 s.t. z T c = 1 Set r = column of B, then c = A - 1 r = column of X

32 (P1): min w T B 1 s.t. w T r = 1 Consider the bijecoon z = A T w, and set r = Ac. We get: (P1): min w T AX 1 s.t. w T Ac = 1 This is equivalent to: (Q1): min z T X 1 s.t. z T c = 1 Set r = column of B, then c = A - 1 r = column of X Claim: If c has a strictly largest coordinate ( c i > c j for j i) in absolute value, then whp the soln to (Q1) is e i

33 (P1): min w T B 1 s.t. w T r = 1 Consider the bijecoon z = A T w, and set r = Ac. We get: (P1): min w T AX 1 s.t. w T Ac = 1 Claim: Then the soln to (P1) is the i th row of X This is equivalent to: (Q1): min z T X 1 s.t. z T c = 1 Set r = column of B, then c = A - 1 r = column of X Claim: If c has a strictly largest coordinate ( c i > c j for j i) in absolute value, then whp the soln to (Q1) is e i

34 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x) Claim: The sparsest vectors in row- span(x) (or B) are the X Can we find the sparsest vector in row- span(x)? Approach #2: L 1 - relaxa&on (P1): min w T B 1 s.t. w T r = 1

35 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x) Claim: The sparsest vectors in row- span(x) (or B) are the X Can we find the sparsest vector in row- span(x)? Approach #2: L 1 - relaxa&on (P1): min w T B 1 s.t. w T r = 1 Hence we can find the rows of X, and solve for A

36 THE OVERCOMPLETE CASE What about overcomplete dicoonaries? (more expressive) Case #2: A is incoherent

37 THE OVERCOMPLETE CASE What about overcomplete dicoonaries? (more expressive) Case #2: A is incoherent Theorem [Arora, Ge, Moitra 13]: There is an algorithm to learn A within ε if it is n by m and μ- incoherent for k min( n/μ log n, m ½- η ) The running Ome and sample complexity are poly(n,m,log 1/ε)

38 THE OVERCOMPLETE CASE What about overcomplete dicoonaries? (more expressive) Case #2: A is incoherent Theorem [Arora, Ge, Moitra 13]: There is an algorithm to learn A within ε if it is n by m and μ- incoherent for k min( n/μ log n, m ½- η ) The running Ome and sample complexity are poly(n,m,log 1/ε) Approach: learn the support of the representaoons X = [ x ] first, by solving an overlapping clustering problem

39 THE OVERCOMPLETE CASE What about overcomplete dicoonaries? (more expressive) Case #2: A is incoherent Theorem [Arora, Ge, Moitra 13]: There is an algorithm to learn A within ε if it is n by m and μ- incoherent for k min( n/μ log n, m ½- η ) The running Ome and sample complexity are poly(n,m,log 1/ε) Approach: learn the support of the representaoons X = [ x ] first, by solving an overlapping clustering problem Theorem [Agarwal et al 13]: There is a poly. Ome algorithm to learn A if it is μ- incoherent for k n ¼ /μ

40 THE MODEL What about overcomplete dicoonaries? (more expressive) Case #2: A is incoherent

41 THE MODEL What about overcomplete dicoonaries? (more expressive) Case #2: A is incoherent Theorem [Barak, Kelner, Steurer 14]: There is a quasi- poly. Ome algorithm to learn A within any constant A if it is μ- incoherent for k n 1- η using the sum- of- squares hierarchy

42 THE MODEL What about overcomplete dicoonaries? (more expressive) Case #2: A is incoherent Theorem [Barak, Kelner, Steurer 14]: There is a quasi- poly. Ome algorithm to learn A within any constant A if it is μ- incoherent for k n 1- η using the sum- of- squares hierarchy Approach: find y that approximately maximizes E[ b T y 4 ] via a poly- logarithmic number of rounds; it is close to a coln of A

43 OUTLINE Are there efficient algorithms for dicoonary learning? Introduc&on Origins of Sparse Recovery A StochasOc Model; Our Results Provable Algorithms via Overlapping Clustering Uncertainty Principles ReformulaOon as Overlapping Clustering Analyzing Alterna&ng Minimiza&on Gradient Descent on Non- Convex Fctns

44 UNCERTAINTY PRINCIPLES

45 UNCERTAINTY PRINCIPLES Claim: Given A, b and k it is NP- hard to decide if there is a k- sparse x such that Ax = b

46 UNCERTAINTY PRINCIPLES Claim: Given A, b and k it is NP- hard to decide if there is a k- sparse x such that Ax = b Why is this easier for incoherent dicoonaries?

47 UNCERTAINTY PRINCIPLES Claim: Given A, b and k it is NP- hard to decide if there is a k- sparse x such that Ax = b Why is this easier for incoherent dicoonaries? Uncertainty Principle: If A is μ- incoherent then Ay, Ax y, x provided that x and y are k- sparse, for k n/2μ

48 UNCERTAINTY PRINCIPLES Claim: Given A, b and k it is NP- hard to decide if there is a k- sparse x such that Ax = b Why is this easier for incoherent dicoonaries? Uncertainty Principle: If A is μ- incoherent then Ay, Ax y, x provided that x and y are k- sparse, for k n/2μ Proof: A T A restricted to the support of x and y is k k and (A T A) i,j = 1 if i = j μ/ n if i j

49 UNCERTAINTY PRINCIPLES Claim: Given A, b and k it is NP- hard to decide if there is a k- sparse x such that Ax = b Why is this easier for incoherent dicoonaries? Uncertainty Principle: If A is μ- incoherent then Ay, Ax y, x provided that x and y are k- sparse, for k n/2μ Proof: A T A restricted to the support of x and y is k k and (A T A) i,j = 1 if i = j μ/ n Then use Gershgorin s Disk Thm if i j

50 UNCERTAINTY PRINCIPLES Claim: Given A, b and k it is NP- hard to decide if there is a k- sparse x such that Ax = b Why is this easier for incoherent dicoonaries? Uncertainty Principle: If A is μ- incoherent then Ay, Ax y, x provided that x and y are k- sparse, for k n/2μ

51 UNCERTAINTY PRINCIPLES Claim: Given A, b and k it is NP- hard to decide if there is a k- sparse x such that Ax = b Why is this easier for incoherent dicoonaries? Uncertainty Principle: If A is μ- incoherent then Ay, Ax y, x provided that x and y are k- sparse, for k n/2μ This principle can be used to establish uniqueness for sparse recovery, and things like b cannot be sparse in both standard and Fourier basis

52 A PAIR- WISE TEST

53 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support?

54 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support? supp(x) = supp(x ) =

55 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support? supp(x) = supp(x ) =

56 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support? supp(x) = supp(x ) = x', x zero non- zero maybe yes

57 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support? supp(x) = supp(x ) = x', x zero non- zero maybe yes x', x Ax, Ax Uncertainty Principle: for k- sparse x, incoherent A

58 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support? supp(x) = supp(x ) = x', x zero non- zero maybe yes x', x Ax, Ax zero non- zero maybe yes, whp Uncertainty Principle: for k- sparse x, incoherent A

59 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support? supp(x) = supp(x ) = Approach: Build a graph G on the p samples, with an edge btwn b and b if and only if b T b > 1/2

60 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support? supp(x) = supp(x ) = Approach: Build a graph G on the p samples, with an edge btwn b and b if and only if b T b > 1/2 For the purposes of this talk, probability of an edge between b, b is ½ iff supp(x) and supp(x ) intersect

61 OVERLAPPING CLUSTERING Let C i ={ b x i 0} (overlapping)

62 OVERLAPPING CLUSTERING Let C i ={ b x i 0} (overlapping) Can we find the clusters efficiently?

63 OVERLAPPING CLUSTERING Let C i ={ b x i 0} (overlapping) Can we find the clusters efficiently? Challenge: Given (x, x, x ) where all the pairs belong to a cluster together, do all three belong to a common cluster too? supp(x) = supp(x ) = supp(x ) =

64 OVERLAPPING CLUSTERING Let C i ={ b x i 0} (overlapping) Can we find the clusters efficiently? Challenge: Given (x, x, x ) where all the pairs belong to a cluster together, do all three belong to a common cluster too?

65 OVERLAPPING CLUSTERING Let C i ={ b x i 0} (overlapping) Can we find the clusters efficiently? Challenge: Given (x, x, x ) where all the pairs belong to a cluster together, do all three belong to a common cluster too? supp(x) = supp(x ) = supp(x ) =

66 A TRIPLE TEST Key Idea: Use new samples y

67 A TRIPLE TEST Key Idea: Use new samples y Case #1: all three intersect: supp(x) = supp(x ) = supp(x ) =

68 A TRIPLE TEST Key Idea: Use new samples y Case #1: all three intersect: Probability y intersects all three is at least k/m supp(x) = supp(x ) = supp(x ) = New sample y only needs to contain one element from their joint union

69 A TRIPLE TEST Key Idea: Use new samples y

70 A TRIPLE TEST Key Idea: Use new samples y Case #2: no common intersecoon supp(x) = supp(x ) = supp(x ) = New sample y needs to contain at least two elements from their joint union

71 A TRIPLE TEST Key Idea: Use new samples y Case #2: no common intersecoon, supp(x) supp(x ) C, etc Probability y intersects all three is at most O(Ck 3 /m 2 ) supp(x) = supp(x ) = supp(x ) = New sample y needs to contain at least two elements from their joint union

72 A TRIPLE TEST Key Idea: Use new samples y Case #1: all three intersect: Probability y intersects all three is at least k/m Case #2: no common intersecoon, supp(x) supp(x ) C, etc Probability y intersects all three is at most O(Ck 3 /m 2 )

73 A TRIPLE TEST Key Idea: Use new samples y Case #1: all three intersect: Probability y intersects all three is at least k/m Case #2: no common intersecoon, supp(x) supp(x ) C, etc Probability y intersects all three is at most O(Ck 3 /m 2 ) Triple Test: Given (x, x, x ) where all the pairs intersect If there are at least T samples y where (x, x, x, y) all pairwise intersect, ACCEPT else REJECT

74 FINDING ALL THE CLUSTERS We can build a clustering algorithm on this primiove: For each pair (x, x ), find all x that pass the triple test

75 FINDING ALL THE CLUSTERS We can build a clustering algorithm on this primiove: For each pair (x, x ), find all x that pass the triple test Claim: This set is the union of clusters corresponding to supp(x) supp(x )

76 FINDING ALL THE CLUSTERS We can build a clustering algorithm on this primiove: For each pair (x, x ), find all x that pass the triple test Claim: This set is the union of clusters corresponding to supp(x) supp(x ) Claim: For every cluster i, there is some x, x that uniquely idenofy it i.e. supp(x) supp(x ) = {i}

77 FINDING ALL THE CLUSTERS We can build a clustering algorithm on this primiove: For each pair (x, x ), find all x that pass the triple test Claim: This set is the union of clusters corresponding to supp(x) supp(x ) Claim: For every cluster i, there is some x, x that uniquely idenofy it i.e. supp(x) supp(x ) = {i} Output inclusion- wise minimal sets these are the clusters!

78 FINDING ALL THE CLUSTERS We can build a clustering algorithm on this primiove: For each pair (x, x ), find all x that pass the triple test Claim: This set is the union of clusters corresponding to supp(x) supp(x ) Claim: For every cluster i, there is some x, x that uniquely idenofy it i.e. supp(x) supp(x ) = {i} Output inclusion- wise minimal sets these are the clusters! Our full algorithm uses higher- order tests; analysis through connec&ons to piercing number

79 Many ways to get the dic&onary from the clustering

80 Many ways to get the dic&onary from the clustering Approach #1: RelaOve Signs Plan: Refine C i and find all the b s with x i > 0

81 Many ways to get the dic&onary from the clustering Approach #1: RelaOve Signs Plan: Refine C i and find all the b s with x i > 0 Intui&on: If supp(x) supp(x ) = {i}, the we can find relaove sign of x i and x i and there are many such pairs

82 Many ways to get the dic&onary from the clustering Approach #1: RelaOve Signs Plan: Refine C i and find all the b s with x i > 0 Intui&on: If supp(x) supp(x ) = {i}, the we can find relaove sign of x i and x i and there are many such pairs enough so that whp we can find all relaove signs by transi&vity

83 Many ways to get the dic&onary from the clustering Approach #1: RelaOve Signs Plan: Refine C i and find all the b s with x i > 0 Intui&on: If supp(x) supp(x ) = {i}, the we can find relaove sign of x i and x i and there are many such pairs enough so that whp we can find all relaove signs by transi&vity Claim: E[b Ax = b and x i > 0] = A i E[x i x i > 0] Hence their empirical average converges to A i

84 Many ways to get the dic&onary from the clustering

85 Many ways to get the dic&onary from the clustering Approach #2: SVD Suppose we restrict to samples b with x i 0.

86 Many ways to get the dic&onary from the clustering Approach #2: SVD Suppose we restrict to samples b with x i 0. Intui&on: E[bb T x i 0] has large variance in direcoon of A i

87 Many ways to get the dic&onary from the clustering Approach #2: SVD Suppose we restrict to samples b with x i 0. Intui&on: E[bb T x i 0] has large variance in direcoon of A i We also show that alternaong minimizaoon works when we re close enough (geometric convergence)

88 OUTLINE Are there efficient algorithms for dicoonary learning? Introduc&on Origins of Sparse Recovery A StochasOc Model; Our Results Provable Algorithms via Overlapping Clustering Uncertainty Principles ReformulaOon as Overlapping Clustering Analyzing Alterna&ng Minimiza&on (out of &me) Gradient Descent on Non- Convex Fctns

89 A CONCEPTUAL OVERVIEW smoothed analysis Overcomplete Models computaoonal geometry NonnegaOve Matrix FactorizaOon Tensor DecomposiOons Be{er EsOmators spectral methods New Models Is Learning ComputaOonally Easy? Topic Models experiments Method of Moments polynomial equaoons e.g. mixtures of Gaussians

90 A CONCEPTUAL OVERVIEW smoothed analysis Overcomplete Models computaoonal geometry NonnegaOve Matrix FactorizaOon Tensor DecomposiOons Be{er EsOmators spectral methods New Models Is Learning ComputaOonally Easy? Topic Models experiments Natural Algorithms Method of Moments polynomial equaoons convex opomizaoon DicOonary Learning e.g. mixtures of Gaussians

91 Summary: Provable algorithms for learning incoherent, overcomplete dicoonaries ConnecOons to overlapping clustering Analysis of alternaong minimizaoon gradient descent on non- convex objecove Why does it work even from a random ini&aliza&on?

92 Any QuesOons? Summary: Provable algorithms for learning incoherent, overcomplete dicoonaries ConnecOons to overlapping clustering Analysis of alternaong minimizaoon gradient descent on non- convex objecove Why does it work even from a random ini&aliza&on?

NEW ALGORITHMS FOR DICTIONARY LEARNING

NEW ALGORITHMS FOR DICTIONARY LEARNING NEW ALGORITHMS FOR DICTIONARY LEARNING ANKUR MOITRA MASSACHUSETTS INSTITUTE OF TECHNOLOGY joint work with Sanjeev Arora, Rong Ge and Tengyu Ma SPARSE REPRESENTATIONS SPARSE REPRESENTATIONS Many data- types

More information

Integral Geometry and the Polynomial Hirsch Conjecture

Integral Geometry and the Polynomial Hirsch Conjecture Integral Geometry and the Polynomial Hirsch Conjecture Jonathan Kelner, MIT Partially based on joint work with Daniel Spielman Introduction n A lot of recent work on Polynomial Hirsch Conjecture has focused

More information

BOOLEAN MATRIX FACTORIZATIONS. with applications in data mining Pauli Miettinen

BOOLEAN MATRIX FACTORIZATIONS. with applications in data mining Pauli Miettinen BOOLEAN MATRIX FACTORIZATIONS with applications in data mining Pauli Miettinen MATRIX FACTORIZATIONS BOOLEAN MATRIX FACTORIZATIONS o THE BOOLEAN MATRIX PRODUCT As normal matrix product, but with addition

More information

Signal Reconstruction from Sparse Representations: An Introdu. Sensing

Signal Reconstruction from Sparse Representations: An Introdu. Sensing Signal Reconstruction from Sparse Representations: An Introduction to Compressed Sensing December 18, 2009 Digital Data Acquisition Suppose we want to acquire some real world signal digitally. Applications

More information

Convex Optimization MLSS 2015

Convex Optimization MLSS 2015 Convex Optimization MLSS 2015 Constantine Caramanis The University of Texas at Austin The Optimization Problem minimize : f (x) subject to : x X. The Optimization Problem minimize : f (x) subject to :

More information

The convex geometry of inverse problems

The convex geometry of inverse problems The convex geometry of inverse problems Benjamin Recht Department of Computer Sciences University of Wisconsin-Madison Joint work with Venkat Chandrasekaran Pablo Parrilo Alan Willsky Linear Inverse Problems

More information

Sparse Reconstruction / Compressive Sensing

Sparse Reconstruction / Compressive Sensing Sparse Reconstruction / Compressive Sensing Namrata Vaswani Department of Electrical and Computer Engineering Iowa State University Namrata Vaswani Sparse Reconstruction / Compressive Sensing 1/ 20 The

More information

ELEG Compressive Sensing and Sparse Signal Representations

ELEG Compressive Sensing and Sparse Signal Representations ELEG 867 - Compressive Sensing and Sparse Signal Representations Gonzalo R. Arce Depart. of Electrical and Computer Engineering University of Delaware Fall 211 Compressive Sensing G. Arce Fall, 211 1 /

More information

Image Restoration and Background Separation Using Sparse Representation Framework

Image Restoration and Background Separation Using Sparse Representation Framework Image Restoration and Background Separation Using Sparse Representation Framework Liu, Shikun Abstract In this paper, we introduce patch-based PCA denoising and k-svd dictionary learning method for the

More information

Introduction to Topics in Machine Learning

Introduction to Topics in Machine Learning Introduction to Topics in Machine Learning Namrata Vaswani Department of Electrical and Computer Engineering Iowa State University Namrata Vaswani 1/ 27 Compressed Sensing / Sparse Recovery: Given y :=

More information

Section 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1

Section 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1 Section 5 Convex Optimisation 1 W. Dai (IC) EE4.66 Data Proc. Convex Optimisation 1 2018 page 5-1 Convex Combination Denition 5.1 A convex combination is a linear combination of points where all coecients

More information

CS675: Convex and Combinatorial Optimization Spring 2018 The Simplex Algorithm. Instructor: Shaddin Dughmi

CS675: Convex and Combinatorial Optimization Spring 2018 The Simplex Algorithm. Instructor: Shaddin Dughmi CS675: Convex and Combinatorial Optimization Spring 2018 The Simplex Algorithm Instructor: Shaddin Dughmi Algorithms for Convex Optimization We will look at 2 algorithms in detail: Simplex and Ellipsoid.

More information

SPARSE COMPONENT ANALYSIS FOR BLIND SOURCE SEPARATION WITH LESS SENSORS THAN SOURCES. Yuanqing Li, Andrzej Cichocki and Shun-ichi Amari

SPARSE COMPONENT ANALYSIS FOR BLIND SOURCE SEPARATION WITH LESS SENSORS THAN SOURCES. Yuanqing Li, Andrzej Cichocki and Shun-ichi Amari SPARSE COMPONENT ANALYSIS FOR BLIND SOURCE SEPARATION WITH LESS SENSORS THAN SOURCES Yuanqing Li, Andrzej Cichocki and Shun-ichi Amari Laboratory for Advanced Brain Signal Processing Laboratory for Mathematical

More information

Outlier Pursuit: Robust PCA and Collaborative Filtering

Outlier Pursuit: Robust PCA and Collaborative Filtering Outlier Pursuit: Robust PCA and Collaborative Filtering Huan Xu Dept. of Mechanical Engineering & Dept. of Mathematics National University of Singapore Joint w/ Constantine Caramanis, Yudong Chen, Sujay

More information

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2017

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2017 CPSC 340: Machine Learning and Data Mining Principal Component Analysis Fall 2017 Assignment 3: 2 late days to hand in tonight. Admin Assignment 4: Due Friday of next week. Last Time: MAP Estimation MAP

More information

Lecture 2 September 3

Lecture 2 September 3 EE 381V: Large Scale Optimization Fall 2012 Lecture 2 September 3 Lecturer: Caramanis & Sanghavi Scribe: Hongbo Si, Qiaoyang Ye 2.1 Overview of the last Lecture The focus of the last lecture was to give

More information

Blind Compressed Sensing Using Sparsifying Transforms

Blind Compressed Sensing Using Sparsifying Transforms Blind Compressed Sensing Using Sparsifying Transforms Saiprasad Ravishankar and Yoram Bresler Department of Electrical and Computer Engineering and Coordinated Science Laboratory University of Illinois

More information

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Minh Dao 1, Xiang Xiang 1, Bulent Ayhan 2, Chiman Kwan 2, Trac D. Tran 1 Johns Hopkins Univeristy, 3400

More information

Collaborative Sparsity and Compressive MRI

Collaborative Sparsity and Compressive MRI Modeling and Computation Seminar February 14, 2013 Table of Contents 1 T2 Estimation 2 Undersampling in MRI 3 Compressed Sensing 4 Model-Based Approach 5 From L1 to L0 6 Spatially Adaptive Sparsity MRI

More information

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)

More information

As a consequence of the operation, there are new incidences between edges and triangles that did not exist in K; see Figure II.9.

As a consequence of the operation, there are new incidences between edges and triangles that did not exist in K; see Figure II.9. II.4 Surface Simplification 37 II.4 Surface Simplification In applications it is often necessary to simplify the data or its representation. One reason is measurement noise, which we would like to eliminate,

More information

Note Set 4: Finite Mixture Models and the EM Algorithm

Note Set 4: Finite Mixture Models and the EM Algorithm Note Set 4: Finite Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine Finite Mixture Models A finite mixture model with K components, for

More information

AMS : Combinatorial Optimization Homework Problems - Week V

AMS : Combinatorial Optimization Homework Problems - Week V AMS 553.766: Combinatorial Optimization Homework Problems - Week V For the following problems, A R m n will be m n matrices, and b R m. An affine subspace is the set of solutions to a a system of linear

More information

A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009]

A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009] A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009] Yongjia Song University of Wisconsin-Madison April 22, 2010 Yongjia Song

More information

Combinatorial meets Convex Leveraging combinatorial structure to obtain faster algorithms, provably

Combinatorial meets Convex Leveraging combinatorial structure to obtain faster algorithms, provably Combinatorial meets Convex Leveraging combinatorial structure to obtain faster algorithms, provably Lorenzo Orecchia Department of Computer Science 6/30/15 Challenges in Optimization for Data Science 1

More information

Lecture 17 Sparse Convex Optimization

Lecture 17 Sparse Convex Optimization Lecture 17 Sparse Convex Optimization Compressed sensing A short introduction to Compressed Sensing An imaging perspective 10 Mega Pixels Scene Image compression Picture Why do we compress images? Introduction

More information

CSC 411 Lecture 18: Matrix Factorizations

CSC 411 Lecture 18: Matrix Factorizations CSC 411 Lecture 18: Matrix Factorizations Roger Grosse, Amir-massoud Farahmand, and Juan Carrasquilla University of Toronto UofT CSC 411: 18-Matrix Factorizations 1 / 27 Overview Recall PCA: project data

More information

Machine Learning and Data Mining. Clustering (1): Basics. Kalev Kask

Machine Learning and Data Mining. Clustering (1): Basics. Kalev Kask Machine Learning and Data Mining Clustering (1): Basics Kalev Kask Unsupervised learning Supervised learning Predict target value ( y ) given features ( x ) Unsupervised learning Understand patterns of

More information

STA 4273H: Sta-s-cal Machine Learning

STA 4273H: Sta-s-cal Machine Learning STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! h0p://www.cs.toronto.edu/~rsalakhu/ Lecture 3 Parametric Distribu>ons We want model the probability

More information

Optimal Denoising of Natural Images and their Multiscale Geometry and Density

Optimal Denoising of Natural Images and their Multiscale Geometry and Density Optimal Denoising of Natural Images and their Multiscale Geometry and Density Department of Computer Science and Applied Mathematics Weizmann Institute of Science, Israel. Joint work with Anat Levin (WIS),

More information

Combinatorial Selection and Least Absolute Shrinkage via The CLASH Operator

Combinatorial Selection and Least Absolute Shrinkage via The CLASH Operator Combinatorial Selection and Least Absolute Shrinkage via The CLASH Operator Volkan Cevher Laboratory for Information and Inference Systems LIONS / EPFL http://lions.epfl.ch & Idiap Research Institute joint

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Clustering and EM Barnabás Póczos & Aarti Singh Contents Clustering K-means Mixture of Gaussians Expectation Maximization Variational Methods 2 Clustering 3 K-

More information

Subspace Clustering with Global Dimension Minimization And Application to Motion Segmentation

Subspace Clustering with Global Dimension Minimization And Application to Motion Segmentation Subspace Clustering with Global Dimension Minimization And Application to Motion Segmentation Bryan Poling University of Minnesota Joint work with Gilad Lerman University of Minnesota The Problem of Subspace

More information

Graph Connectivity in Sparse Subspace Clustering

Graph Connectivity in Sparse Subspace Clustering www.nicta.com.au From imagination to impact Graph Connectivity in Sparse Subspace Clustering Behrooz Nasihatkon 24 March 2011 Outline 1 Subspace Clustering 2 Sparse Subspace Clustering 3 Graph Connectivity

More information

Compressed Sensing Algorithm for Real-Time Doppler Ultrasound Image Reconstruction

Compressed Sensing Algorithm for Real-Time Doppler Ultrasound Image Reconstruction Mathematical Modelling and Applications 2017; 2(6): 75-80 http://www.sciencepublishinggroup.com/j/mma doi: 10.11648/j.mma.20170206.14 ISSN: 2575-1786 (Print); ISSN: 2575-1794 (Online) Compressed Sensing

More information

COMP251: Disjoint sets

COMP251: Disjoint sets COMP251: Disjoint sets Jérôme Waldispühl School of Computer Science McGill University Based on slides from M. Langer (McGill) Announces Assignment 1 is due on Wednesday February 1 st. Submit your soluoon

More information

Shannon Switching Game

Shannon Switching Game EECS 495: Combinatorial Optimization Lecture 1 Shannon s Switching Game Shannon Switching Game In the Shannon switching game, two players, Join and Cut, alternate choosing edges on a graph G. Join s objective

More information

1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics

1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics 1. Introduction EE 546, Univ of Washington, Spring 2016 performance of numerical methods complexity bounds structural convex optimization course goals and topics 1 1 Some course info Welcome to EE 546!

More information

Outline. CS38 Introduction to Algorithms. Linear programming 5/21/2014. Linear programming. Lecture 15 May 20, 2014

Outline. CS38 Introduction to Algorithms. Linear programming 5/21/2014. Linear programming. Lecture 15 May 20, 2014 5/2/24 Outline CS38 Introduction to Algorithms Lecture 5 May 2, 24 Linear programming simplex algorithm LP duality ellipsoid algorithm * slides from Kevin Wayne May 2, 24 CS38 Lecture 5 May 2, 24 CS38

More information

Unsupervised Learning: Clustering

Unsupervised Learning: Clustering Unsupervised Learning: Clustering Vibhav Gogate The University of Texas at Dallas Slides adapted from Carlos Guestrin, Dan Klein & Luke Zettlemoyer Machine Learning Supervised Learning Unsupervised Learning

More information

Inverse Problems and Machine Learning

Inverse Problems and Machine Learning Inverse Problems and Machine Learning Julian Wörmann Research Group for Geometric Optimization and Machine Learning (GOL) 1 What are inverse problems? 2 Inverse Problems cause/ excitation 3 Inverse Problems

More information

K-Means and Gaussian Mixture Models

K-Means and Gaussian Mixture Models K-Means and Gaussian Mixture Models David Rosenberg New York University June 15, 2015 David Rosenberg (New York University) DS-GA 1003 June 15, 2015 1 / 43 K-Means Clustering Example: Old Faithful Geyser

More information

Kernels and representation

Kernels and representation Kernels and representation Corso di AA, anno 2017/18, Padova Fabio Aiolli 20 Dicembre 2017 Fabio Aiolli Kernels and representation 20 Dicembre 2017 1 / 19 (Hierarchical) Representation Learning Hierarchical

More information

IE598 Big Data Optimization Summary Nonconvex Optimization

IE598 Big Data Optimization Summary Nonconvex Optimization IE598 Big Data Optimization Summary Nonconvex Optimization Instructor: Niao He April 16, 2018 1 This Course Big Data Optimization Explore modern optimization theories, algorithms, and big data applications

More information

Robust Principal Component Analysis (RPCA)

Robust Principal Component Analysis (RPCA) Robust Principal Component Analysis (RPCA) & Matrix decomposition: into low-rank and sparse components Zhenfang Hu 2010.4.1 reference [1] Chandrasekharan, V., Sanghavi, S., Parillo, P., Wilsky, A.: Ranksparsity

More information

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Approximation algorithms Date: 11/27/18

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Approximation algorithms Date: 11/27/18 601.433/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Approximation algorithms Date: 11/27/18 22.1 Introduction We spent the last two lectures proving that for certain problems, we can

More information

Clustering Lecture 5: Mixture Model

Clustering Lecture 5: Mixture Model Clustering Lecture 5: Mixture Model Jing Gao SUNY Buffalo 1 Outline Basics Motivation, definition, evaluation Methods Partitional Hierarchical Density-based Mixture model Spectral methods Advanced topics

More information

Stanford University CS359G: Graph Partitioning and Expanders Handout 1 Luca Trevisan January 4, 2011

Stanford University CS359G: Graph Partitioning and Expanders Handout 1 Luca Trevisan January 4, 2011 Stanford University CS359G: Graph Partitioning and Expanders Handout 1 Luca Trevisan January 4, 2011 Lecture 1 In which we describe what this course is about. 1 Overview This class is about the following

More information

Approximation slides 1. An optimal polynomial algorithm for the Vertex Cover and matching in Bipartite graphs

Approximation slides 1. An optimal polynomial algorithm for the Vertex Cover and matching in Bipartite graphs Approximation slides 1 An optimal polynomial algorithm for the Vertex Cover and matching in Bipartite graphs Approximation slides 2 Linear independence A collection of row vectors {v T i } are independent

More information

Implicit Matrix Representations of Rational Bézier Curves and Surfaces

Implicit Matrix Representations of Rational Bézier Curves and Surfaces Implicit Matrix Representations of Rational Bézier Curves and Surfaces Laurent Busé INRIA Sophia Antipolis, France Laurent.Buse@inria.fr GD/SPM Conference, Denver, USA November 11, 2013 Overall motivation

More information

Efficient Iterative Semi-supervised Classification on Manifold

Efficient Iterative Semi-supervised Classification on Manifold . Efficient Iterative Semi-supervised Classification on Manifold... M. Farajtabar, H. R. Rabiee, A. Shaban, A. Soltani-Farani Sharif University of Technology, Tehran, Iran. Presented by Pooria Joulani

More information

Interpolation & Polynomial Approximation. Cubic Spline Interpolation II

Interpolation & Polynomial Approximation. Cubic Spline Interpolation II Interpolation & Polynomial Approximation Cubic Spline Interpolation II Numerical Analysis (9th Edition) R L Burden & J D Faires Beamer Presentation Slides prepared by John Carroll Dublin City University

More information

A Patch Prior for Dense 3D Reconstruction in Man-Made Environments

A Patch Prior for Dense 3D Reconstruction in Man-Made Environments A Patch Prior for Dense 3D Reconstruction in Man-Made Environments Christian Häne 1, Christopher Zach 2, Bernhard Zeisl 1, Marc Pollefeys 1 1 ETH Zürich 2 MSR Cambridge October 14, 2012 A Patch Prior for

More information

Sparse Models in Image Understanding And Computer Vision

Sparse Models in Image Understanding And Computer Vision Sparse Models in Image Understanding And Computer Vision Jayaraman J. Thiagarajan Arizona State University Collaborators Prof. Andreas Spanias Karthikeyan Natesan Ramamurthy Sparsity Sparsity of a vector

More information

CS Introduction to Data Mining Instructor: Abdullah Mueen

CS Introduction to Data Mining Instructor: Abdullah Mueen CS 591.03 Introduction to Data Mining Instructor: Abdullah Mueen LECTURE 8: ADVANCED CLUSTERING (FUZZY AND CO -CLUSTERING) Review: Basic Cluster Analysis Methods (Chap. 10) Cluster Analysis: Basic Concepts

More information

Image Restoration Using DNN

Image Restoration Using DNN Image Restoration Using DNN Hila Levi & Eran Amar Images were taken from: http://people.tuebingen.mpg.de/burger/neural_denoising/ Agenda Domain Expertise vs. End-to-End optimization Image Denoising and

More information

Introduc)on to Machine Learning: Clustering Algorithms How to Analyze Your Own Genome Fall 2013

Introduc)on to Machine Learning: Clustering Algorithms How to Analyze Your Own Genome Fall 2013 Introduc)on to Machine Learning: Clustering Algorithms 02-223 How to Analyze Your Own Genome Fall 203 Organizing data into clusters such that there is high intra- cluster similarity low inter- cluster

More information

Measurements and Bits: Compressed Sensing meets Information Theory. Dror Baron ECE Department Rice University dsp.rice.edu/cs

Measurements and Bits: Compressed Sensing meets Information Theory. Dror Baron ECE Department Rice University dsp.rice.edu/cs Measurements and Bits: Compressed Sensing meets Information Theory Dror Baron ECE Department Rice University dsp.rice.edu/cs Sensing by Sampling Sample data at Nyquist rate Compress data using model (e.g.,

More information

Interleaving Schemes on Circulant Graphs with Two Offsets

Interleaving Schemes on Circulant Graphs with Two Offsets Interleaving Schemes on Circulant raphs with Two Offsets Aleksandrs Slivkins Department of Computer Science Cornell University Ithaca, NY 14853 slivkins@cs.cornell.edu Jehoshua Bruck Department of Electrical

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-4: Constrained optimization Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428 June

More information

Face Recognition via Sparse Representation

Face Recognition via Sparse Representation Face Recognition via Sparse Representation John Wright, Allen Y. Yang, Arvind, S. Shankar Sastry and Yi Ma IEEE Trans. PAMI, March 2008 Research About Face Face Detection Face Alignment Face Recognition

More information

Lecture notes on the simplex method September We will present an algorithm to solve linear programs of the form. maximize.

Lecture notes on the simplex method September We will present an algorithm to solve linear programs of the form. maximize. Cornell University, Fall 2017 CS 6820: Algorithms Lecture notes on the simplex method September 2017 1 The Simplex Method We will present an algorithm to solve linear programs of the form maximize subject

More information

INCOHERENT DICTIONARY LEARNING FOR SPARSE REPRESENTATION BASED IMAGE DENOISING

INCOHERENT DICTIONARY LEARNING FOR SPARSE REPRESENTATION BASED IMAGE DENOISING INCOHERENT DICTIONARY LEARNING FOR SPARSE REPRESENTATION BASED IMAGE DENOISING Jin Wang 1, Jian-Feng Cai 2, Yunhui Shi 1 and Baocai Yin 1 1 Beijing Key Laboratory of Multimedia and Intelligent Software

More information

15-451/651: Design & Analysis of Algorithms October 11, 2018 Lecture #13: Linear Programming I last changed: October 9, 2018

15-451/651: Design & Analysis of Algorithms October 11, 2018 Lecture #13: Linear Programming I last changed: October 9, 2018 15-451/651: Design & Analysis of Algorithms October 11, 2018 Lecture #13: Linear Programming I last changed: October 9, 2018 In this lecture, we describe a very general problem called linear programming

More information

Clustering CS 550: Machine Learning

Clustering CS 550: Machine Learning Clustering CS 550: Machine Learning This slide set mainly uses the slides given in the following links: http://www-users.cs.umn.edu/~kumar/dmbook/ch8.pdf http://www-users.cs.umn.edu/~kumar/dmbook/dmslides/chap8_basic_cluster_analysis.pdf

More information

Mixture Models and the EM Algorithm

Mixture Models and the EM Algorithm Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine c 2017 1 Finite Mixture Models Say we have a data set D = {x 1,..., x N } where x i is

More information

Off-the-Grid Compressive Imaging: Recovery of Piecewise Constant Images from Few Fourier Samples

Off-the-Grid Compressive Imaging: Recovery of Piecewise Constant Images from Few Fourier Samples Off-the-Grid Compressive Imaging: Recovery of Piecewise Constant Images from Few Fourier Samples Greg Ongie PhD Candidate Department of Applied Math and Computational Sciences University of Iowa April

More information

Matching and Covering

Matching and Covering Matching and Covering Matchings A matching of size k in a graph G is a set of k pairwise disjoint edges The vertices belonging to the edges of a matching are saturated by the matching; the others are unsaturated

More information

15-451/651: Design & Analysis of Algorithms November 4, 2015 Lecture #18 last changed: November 22, 2015

15-451/651: Design & Analysis of Algorithms November 4, 2015 Lecture #18 last changed: November 22, 2015 15-451/651: Design & Analysis of Algorithms November 4, 2015 Lecture #18 last changed: November 22, 2015 While we have good algorithms for many optimization problems, the previous lecture showed that many

More information

Data Mining Learning from Large Data Sets

Data Mining Learning from Large Data Sets Data Mining Learning from Large Data Sets Lecture 8 Clustering large data sets 263-5200- 00L Andreas Krause Announcements! Homework 4 out tomorrow 2 Course organizapon! Retrieval! Given a query, find most

More information

Mathematical and Algorithmic Foundations Linear Programming and Matchings

Mathematical and Algorithmic Foundations Linear Programming and Matchings Adavnced Algorithms Lectures Mathematical and Algorithmic Foundations Linear Programming and Matchings Paul G. Spirakis Department of Computer Science University of Patras and Liverpool Paul G. Spirakis

More information

Lecture 11: Clustering and the Spectral Partitioning Algorithm A note on randomized algorithm, Unbiased estimates

Lecture 11: Clustering and the Spectral Partitioning Algorithm A note on randomized algorithm, Unbiased estimates CSE 51: Design and Analysis of Algorithms I Spring 016 Lecture 11: Clustering and the Spectral Partitioning Algorithm Lecturer: Shayan Oveis Gharan May nd Scribe: Yueqi Sheng Disclaimer: These notes have

More information

Linear Programming. Readings: Read text section 11.6, and sections 1 and 2 of Tom Ferguson s notes (see course homepage).

Linear Programming. Readings: Read text section 11.6, and sections 1 and 2 of Tom Ferguson s notes (see course homepage). Linear Programming Learning Goals. Introduce Linear Programming Problems. Widget Example, Graphical Solution. Basic Theory: Feasible Set, Vertices, Existence of Solutions. Equivalent formulations. Outline

More information

Spectral Clustering. Presented by Eldad Rubinstein Based on a Tutorial by Ulrike von Luxburg TAU Big Data Processing Seminar December 14, 2014

Spectral Clustering. Presented by Eldad Rubinstein Based on a Tutorial by Ulrike von Luxburg TAU Big Data Processing Seminar December 14, 2014 Spectral Clustering Presented by Eldad Rubinstein Based on a Tutorial by Ulrike von Luxburg TAU Big Data Processing Seminar December 14, 2014 What are we going to talk about? Introduction Clustering and

More information

Linear Programming. Widget Factory Example. Linear Programming: Standard Form. Widget Factory Example: Continued.

Linear Programming. Widget Factory Example. Linear Programming: Standard Form. Widget Factory Example: Continued. Linear Programming Widget Factory Example Learning Goals. Introduce Linear Programming Problems. Widget Example, Graphical Solution. Basic Theory:, Vertices, Existence of Solutions. Equivalent formulations.

More information

Robot Mapping. Least Squares Approach to SLAM. Cyrill Stachniss

Robot Mapping. Least Squares Approach to SLAM. Cyrill Stachniss Robot Mapping Least Squares Approach to SLAM Cyrill Stachniss 1 Three Main SLAM Paradigms Kalman filter Particle filter Graphbased least squares approach to SLAM 2 Least Squares in General Approach for

More information

Graphbased. Kalman filter. Particle filter. Three Main SLAM Paradigms. Robot Mapping. Least Squares Approach to SLAM. Least Squares in General

Graphbased. Kalman filter. Particle filter. Three Main SLAM Paradigms. Robot Mapping. Least Squares Approach to SLAM. Least Squares in General Robot Mapping Three Main SLAM Paradigms Least Squares Approach to SLAM Kalman filter Particle filter Graphbased Cyrill Stachniss least squares approach to SLAM 1 2 Least Squares in General! Approach for

More information

AMS /672: Graph Theory Homework Problems - Week V. Problems to be handed in on Wednesday, March 2: 6, 8, 9, 11, 12.

AMS /672: Graph Theory Homework Problems - Week V. Problems to be handed in on Wednesday, March 2: 6, 8, 9, 11, 12. AMS 550.47/67: Graph Theory Homework Problems - Week V Problems to be handed in on Wednesday, March : 6, 8, 9,,.. Assignment Problem. Suppose we have a set {J, J,..., J r } of r jobs to be filled by a

More information

Facial Expression Classification with Random Filters Feature Extraction

Facial Expression Classification with Random Filters Feature Extraction Facial Expression Classification with Random Filters Feature Extraction Mengye Ren Facial Monkey mren@cs.toronto.edu Zhi Hao Luo It s Me lzh@cs.toronto.edu I. ABSTRACT In our work, we attempted to tackle

More information

Efficient Imaging Algorithms on Many-Core Platforms

Efficient Imaging Algorithms on Many-Core Platforms Efficient Imaging Algorithms on Many-Core Platforms H. Köstler Dagstuhl, 22.11.2011 Contents Imaging Applications HDR Compression performance of PDE-based models Image Denoising performance of patch-based

More information

Methods for Intelligent Systems

Methods for Intelligent Systems Methods for Intelligent Systems Lecture Notes on Clustering (II) Davide Eynard eynard@elet.polimi.it Department of Electronics and Information Politecnico di Milano Davide Eynard - Lecture Notes on Clustering

More information

Large-Scale Lasso and Elastic-Net Regularized Generalized Linear Models

Large-Scale Lasso and Elastic-Net Regularized Generalized Linear Models Large-Scale Lasso and Elastic-Net Regularized Generalized Linear Models DB Tsai Steven Hillion Outline Introduction Linear / Nonlinear Classification Feature Engineering - Polynomial Expansion Big-data

More information

When Sparsity Meets Low-Rankness: Transform Learning With Non-Local Low-Rank Constraint For Image Restoration

When Sparsity Meets Low-Rankness: Transform Learning With Non-Local Low-Rank Constraint For Image Restoration When Sparsity Meets Low-Rankness: Transform Learning With Non-Local Low-Rank Constraint For Image Restoration Bihan Wen, Yanjun Li and Yoram Bresler Department of Electrical and Computer Engineering Coordinated

More information

Mathematical Programming Methods for Minimizing the Zero-norm over Polyhedral Sets. Francesco Rinaldi

Mathematical Programming Methods for Minimizing the Zero-norm over Polyhedral Sets. Francesco Rinaldi Sapienza, University of Rome Mathematical Programming Methods for Minimizing the Zero-norm over Polyhedral Sets Francesco Rinaldi Dissertation for the Degree of Philosophy Doctor in Operations Research

More information

Structurally Random Matrices

Structurally Random Matrices Fast Compressive Sampling Using Structurally Random Matrices Presented by: Thong Do (thongdo@jhu.edu) The Johns Hopkins University A joint work with Prof. Trac Tran, The Johns Hopkins University it Dr.

More information

Understanding Clustering Supervising the unsupervised

Understanding Clustering Supervising the unsupervised Understanding Clustering Supervising the unsupervised Janu Verma IBM T.J. Watson Research Center, New York http://jverma.github.io/ jverma@us.ibm.com @januverma Clustering Grouping together similar data

More information

arxiv: v1 [math.co] 15 Dec 2009

arxiv: v1 [math.co] 15 Dec 2009 ANOTHER PROOF OF THE FACT THAT POLYHEDRAL CONES ARE FINITELY GENERATED arxiv:092.2927v [math.co] 5 Dec 2009 VOLKER KAIBEL Abstract. In this note, we work out a simple inductive proof showing that every

More information

An incomplete survey of matrix completion. Benjamin Recht Department of Computer Sciences University of Wisconsin-Madison

An incomplete survey of matrix completion. Benjamin Recht Department of Computer Sciences University of Wisconsin-Madison An incomplete survey of matrix completion Benjamin Recht Department of Computer Sciences University of Wisconsin-Madison Abstract Setup: Matrix Completion M = M ij known for black cells M ij unknown for

More information

Algebraic Iterative Methods for Computed Tomography

Algebraic Iterative Methods for Computed Tomography Algebraic Iterative Methods for Computed Tomography Per Christian Hansen DTU Compute Department of Applied Mathematics and Computer Science Technical University of Denmark Per Christian Hansen Algebraic

More information

Parallelization in the Big Data Regime 5: Data Parallelization? Sham M. Kakade

Parallelization in the Big Data Regime 5: Data Parallelization? Sham M. Kakade Parallelization in the Big Data Regime 5: Data Parallelization? Sham M. Kakade Machine Learning for Big Data CSE547/STAT548 University of Washington S. M. Kakade (UW) Optimization for Big data 1 / 23 Announcements...

More information

Iteratively Re-weighted Least Squares for Sums of Convex Functions

Iteratively Re-weighted Least Squares for Sums of Convex Functions Iteratively Re-weighted Least Squares for Sums of Convex Functions James Burke University of Washington Jiashan Wang LinkedIn Frank Curtis Lehigh University Hao Wang Shanghai Tech University Daiwei He

More information

NP-Complete Reductions 2

NP-Complete Reductions 2 x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 12 22 32 CS 447 11 13 21 23 31 33 Algorithms NP-Complete Reductions 2 Prof. Gregory Provan Department of Computer Science University College Cork 1 Lecture Outline NP-Complete

More information

Sparse wavelet expansions for seismic tomography: Methods and algorithms

Sparse wavelet expansions for seismic tomography: Methods and algorithms Sparse wavelet expansions for seismic tomography: Methods and algorithms Ignace Loris Université Libre de Bruxelles International symposium on geophysical imaging with localized waves 24 28 July 2011 (Joint

More information

Optimization for Machine Learning

Optimization for Machine Learning with a focus on proximal gradient descent algorithm Department of Computer Science and Engineering Outline 1 History & Trends 2 Proximal Gradient Descent 3 Three Applications A Brief History A. Convex

More information

by conservation of flow, hence the cancelation. Similarly, we have

by conservation of flow, hence the cancelation. Similarly, we have Chapter 13: Network Flows and Applications Network: directed graph with source S and target T. Non-negative edge weights represent capacities. Assume no edges into S or out of T. (If necessary, we can

More information

Clustering: Centroid-Based Partitioning

Clustering: Centroid-Based Partitioning Clustering: Centroid-Based Partitioning Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong 1 / 29 Y Tao Clustering: Centroid-Based Partitioning In this lecture, we

More information

Recovery of Piecewise Smooth Images from Few Fourier Samples

Recovery of Piecewise Smooth Images from Few Fourier Samples Recovery of Piecewise Smooth Images from Few Fourier Samples Greg Ongie*, Mathews Jacob Computational Biomedical Imaging Group (CBIG) University of Iowa SampTA 2015 Washington, D.C. 1. Introduction 2.

More information

Image Deblurring Using Adaptive Sparse Domain Selection and Adaptive Regularization

Image Deblurring Using Adaptive Sparse Domain Selection and Adaptive Regularization Volume 3, No. 3, May-June 2012 International Journal of Advanced Research in Computer Science RESEARCH PAPER Available Online at www.ijarcs.info ISSN No. 0976-5697 Image Deblurring Using Adaptive Sparse

More information

Pareto Optimal Solutions for Smoothed Analysts

Pareto Optimal Solutions for Smoothed Analysts Pareto Optimal Solutions for Smoothed Analysts Ankur Moitra, IAS joint work with Ryan O Donnell, CMU September 26, 2011 Ankur Moitra (IAS) Pareto September 26, 2011 values weights v1 w 1...... v i w i

More information

Sparse Solutions to Linear Inverse Problems. Yuzhe Jin

Sparse Solutions to Linear Inverse Problems. Yuzhe Jin Sparse Solutions to Linear Inverse Problems Yuzhe Jin Outline Intro/Background Two types of algorithms Forward Sequential Selection Methods Diversity Minimization Methods Experimental results Potential

More information