ALGORITHMS FOR DICTIONARY LEARNING
|
|
- Patrick Hicks
- 5 years ago
- Views:
Transcription
1 ALGORITHMS FOR DICTIONARY LEARNING ANKUR MOITRA MASSACHUSETTS INSTITUTE OF TECHNOLOGY
2 SPARSE REPRESENTATIONS
3 SPARSE REPRESENTATIONS Many data- types are sparse in an appropriately chosen basis:
4 SPARSE REPRESENTATIONS Many data- types are sparse in an appropriately chosen basis: b i e.g. images, signals, data (n p)
5 SPARSE REPRESENTATIONS Many data- types are sparse in an appropriately chosen basis: dicoonary (n m) A x i b i e.g. images, signals, representaoons (m p) data (n p)
6 SPARSE REPRESENTATIONS Many data- types are sparse in an appropriately chosen basis: dicoonary (n m) at most k non- zeros A x i b i e.g. images, signals, representaoons (m p) data (n p)
7 SPARSE REPRESENTATIONS Many data- types are sparse in an appropriately chosen basis: dicoonary (n m) at most k non- zeros A x i b i Dic&onary Learning: Can we learn A from examples? e.g. images, signals, representaoons (m p) data (n p)
8 APPLICATIONS OF DICTIONARY LEARNING a.k.a. sparse coding
9 APPLICATIONS OF DICTIONARY LEARNING a.k.a. sparse coding Signal Processing/Sta&s&cs: De- noising, edge- detecoon, super- resoluoon Block compression for images/video
10 APPLICATIONS OF DICTIONARY LEARNING a.k.a. sparse coding Signal Processing/Sta&s&cs: De- noising, edge- detecoon, super- resoluoon Block compression for images/video Machine Learning: Sparsity as a regularizer to prevent over- fi[ng Learned sparse reps. play a key role in deep- learning
11 APPLICATIONS OF DICTIONARY LEARNING a.k.a. sparse coding Signal Processing/Sta&s&cs: De- noising, edge- detecoon, super- resoluoon Block compression for images/video Machine Learning: Sparsity as a regularizer to prevent over- fi[ng Learned sparse reps. play a key role in deep- learning Computa&onal Neuroscience (Olshausen- Field 1997): Applied to natural images yields filters with same qualitaove properoes as recepove field in V1
12 OUTLINE Are there efficient algorithms for dicoonary learning? Introduc&on Origins of Sparse Recovery A StochasOc Model; Our Results Provable Algorithms via Overlapping Clustering Uncertainty Principles ReformulaOon as Overlapping Clustering Analyzing Alterna&ng Minimiza&on Gradient Descent on Non- Convex Fctns
13 ORIGINS OF SPARSE RECOVERY Donoho- Stark, Donoho- Huo, Gribonval- Nielsen, Donoho- Elad: dicoonary (n m) A Incoherence: μ = max i j A i, A j / n x = b at most k non- zeros
14 ORIGINS OF SPARSE RECOVERY Donoho- Stark, Donoho- Huo, Gribonval- Nielsen, Donoho- Elad: dicoonary (n m) A Incoherence: μ = max i j A i, A j / n x = b at most k non- zeros for spikes- and- sines μ = 1
15 ORIGINS OF SPARSE RECOVERY Donoho- Stark, Donoho- Huo, Gribonval- Nielsen, Donoho- Elad: If k n / 2μ then x is the sparsest soluoon to the linear system, and can be found with l 1 - minimizaoon dicoonary (n m) A Incoherence: μ = max i j A i, A j / n x = b at most k non- zeros for spikes- and- sines μ = 1
16 THE FULL RANK CASE Are there efficient algorithms for dicoonary learning? Case #1: A has full column rank
17 THE FULL RANK CASE Are there efficient algorithms for dicoonary learning? Case #1: A has full column rank Theorem [Spielman, Wang, Wright 13]: There is a poly. Ome algorithm to exactly learn A when it has full column rank, for k n (hence m n)
18 THE FULL RANK CASE Are there efficient algorithms for dicoonary learning? Case #1: A has full column rank Theorem [Spielman, Wang, Wright 13]: There is a poly. Ome algorithm to exactly learn A when it has full column rank, for k n (hence m n) Approach: find the rows of A - 1, using L 1 - minimizaoon
19 THE FULL RANK CASE Are there efficient algorithms for dicoonary learning? Case #1: A has full column rank Theorem [Spielman, Wang, Wright 13]: There is a poly. Ome algorithm to exactly learn A when it has full column rank, for k n (hence m n) Approach: find the rows of A - 1, using L 1 - minimizaoon Stochas&c Model: unknown dicoonary A generate x with support of size k u.a.r., choose non- zero values independently, observe b = Ax
20 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely
21 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x)
22 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x) Claim: The sparsest vectors in row- span(x) (or B) are the X
23 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x) Claim: The sparsest vectors in row- span(x) (or B) are the X Can we find the sparsest vector in row- span(x)?
24 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x) Claim: The sparsest vectors in row- span(x) (or B) are the X Can we find the sparsest vector in row- span(x)? Approach #1: (P0): min w T B 0 s.t. w 0
25 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x) Claim: The sparsest vectors in row- span(x) (or B) are the X Can we find the sparsest vector in row- span(x)? Approach #1: NP- hard (P0): min w T B 1 s.t. w 0
26 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x) Claim: The sparsest vectors in row- span(x) (or B) are the X Can we find the sparsest vector in row- span(x)? Approach #2: L 1 - relaxa&on (P1): min w T B 1 s.t. w T r = 1 where we will set r later
27 (P1): min w T B 1 s.t. w T r = 1
28 (P1): min w T B 1 s.t. w T r = 1 Consider the bijecoon z = A T w, and set c = A - 1 r.
29 (P1): min w T B 1 s.t. w T r = 1 Consider the bijecoon z = A T w, and set r = Ac. We get: (P1): min w T AX 1 s.t. w T Ac = 1
30 (P1): min w T B 1 s.t. w T r = 1 Consider the bijecoon z = A T w, and set r = Ac. We get: (P1): min w T AX 1 s.t. w T Ac = 1 This is equivalent to: (Q1): min z T X 1 s.t. z T c = 1
31 (P1): min w T B 1 s.t. w T r = 1 Consider the bijecoon z = A T w, and set r = Ac. We get: (P1): min w T AX 1 s.t. w T Ac = 1 This is equivalent to: (Q1): min z T X 1 s.t. z T c = 1 Set r = column of B, then c = A - 1 r = column of X
32 (P1): min w T B 1 s.t. w T r = 1 Consider the bijecoon z = A T w, and set r = Ac. We get: (P1): min w T AX 1 s.t. w T Ac = 1 This is equivalent to: (Q1): min z T X 1 s.t. z T c = 1 Set r = column of B, then c = A - 1 r = column of X Claim: If c has a strictly largest coordinate ( c i > c j for j i) in absolute value, then whp the soln to (Q1) is e i
33 (P1): min w T B 1 s.t. w T r = 1 Consider the bijecoon z = A T w, and set r = Ac. We get: (P1): min w T AX 1 s.t. w T Ac = 1 Claim: Then the soln to (P1) is the i th row of X This is equivalent to: (Q1): min z T X 1 s.t. z T c = 1 Set r = column of B, then c = A - 1 r = column of X Claim: If c has a strictly largest coordinate ( c i > c j for j i) in absolute value, then whp the soln to (Q1) is e i
34 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x) Claim: The sparsest vectors in row- span(x) (or B) are the X Can we find the sparsest vector in row- span(x)? Approach #2: L 1 - relaxa&on (P1): min w T B 1 s.t. w T r = 1
35 Nota&on: AX = B, where the columns of B, X are samples and their representaoons respecovely Claim: row- span(b) = row- span(x) Claim: The sparsest vectors in row- span(x) (or B) are the X Can we find the sparsest vector in row- span(x)? Approach #2: L 1 - relaxa&on (P1): min w T B 1 s.t. w T r = 1 Hence we can find the rows of X, and solve for A
36 THE OVERCOMPLETE CASE What about overcomplete dicoonaries? (more expressive) Case #2: A is incoherent
37 THE OVERCOMPLETE CASE What about overcomplete dicoonaries? (more expressive) Case #2: A is incoherent Theorem [Arora, Ge, Moitra 13]: There is an algorithm to learn A within ε if it is n by m and μ- incoherent for k min( n/μ log n, m ½- η ) The running Ome and sample complexity are poly(n,m,log 1/ε)
38 THE OVERCOMPLETE CASE What about overcomplete dicoonaries? (more expressive) Case #2: A is incoherent Theorem [Arora, Ge, Moitra 13]: There is an algorithm to learn A within ε if it is n by m and μ- incoherent for k min( n/μ log n, m ½- η ) The running Ome and sample complexity are poly(n,m,log 1/ε) Approach: learn the support of the representaoons X = [ x ] first, by solving an overlapping clustering problem
39 THE OVERCOMPLETE CASE What about overcomplete dicoonaries? (more expressive) Case #2: A is incoherent Theorem [Arora, Ge, Moitra 13]: There is an algorithm to learn A within ε if it is n by m and μ- incoherent for k min( n/μ log n, m ½- η ) The running Ome and sample complexity are poly(n,m,log 1/ε) Approach: learn the support of the representaoons X = [ x ] first, by solving an overlapping clustering problem Theorem [Agarwal et al 13]: There is a poly. Ome algorithm to learn A if it is μ- incoherent for k n ¼ /μ
40 THE MODEL What about overcomplete dicoonaries? (more expressive) Case #2: A is incoherent
41 THE MODEL What about overcomplete dicoonaries? (more expressive) Case #2: A is incoherent Theorem [Barak, Kelner, Steurer 14]: There is a quasi- poly. Ome algorithm to learn A within any constant A if it is μ- incoherent for k n 1- η using the sum- of- squares hierarchy
42 THE MODEL What about overcomplete dicoonaries? (more expressive) Case #2: A is incoherent Theorem [Barak, Kelner, Steurer 14]: There is a quasi- poly. Ome algorithm to learn A within any constant A if it is μ- incoherent for k n 1- η using the sum- of- squares hierarchy Approach: find y that approximately maximizes E[ b T y 4 ] via a poly- logarithmic number of rounds; it is close to a coln of A
43 OUTLINE Are there efficient algorithms for dicoonary learning? Introduc&on Origins of Sparse Recovery A StochasOc Model; Our Results Provable Algorithms via Overlapping Clustering Uncertainty Principles ReformulaOon as Overlapping Clustering Analyzing Alterna&ng Minimiza&on Gradient Descent on Non- Convex Fctns
44 UNCERTAINTY PRINCIPLES
45 UNCERTAINTY PRINCIPLES Claim: Given A, b and k it is NP- hard to decide if there is a k- sparse x such that Ax = b
46 UNCERTAINTY PRINCIPLES Claim: Given A, b and k it is NP- hard to decide if there is a k- sparse x such that Ax = b Why is this easier for incoherent dicoonaries?
47 UNCERTAINTY PRINCIPLES Claim: Given A, b and k it is NP- hard to decide if there is a k- sparse x such that Ax = b Why is this easier for incoherent dicoonaries? Uncertainty Principle: If A is μ- incoherent then Ay, Ax y, x provided that x and y are k- sparse, for k n/2μ
48 UNCERTAINTY PRINCIPLES Claim: Given A, b and k it is NP- hard to decide if there is a k- sparse x such that Ax = b Why is this easier for incoherent dicoonaries? Uncertainty Principle: If A is μ- incoherent then Ay, Ax y, x provided that x and y are k- sparse, for k n/2μ Proof: A T A restricted to the support of x and y is k k and (A T A) i,j = 1 if i = j μ/ n if i j
49 UNCERTAINTY PRINCIPLES Claim: Given A, b and k it is NP- hard to decide if there is a k- sparse x such that Ax = b Why is this easier for incoherent dicoonaries? Uncertainty Principle: If A is μ- incoherent then Ay, Ax y, x provided that x and y are k- sparse, for k n/2μ Proof: A T A restricted to the support of x and y is k k and (A T A) i,j = 1 if i = j μ/ n Then use Gershgorin s Disk Thm if i j
50 UNCERTAINTY PRINCIPLES Claim: Given A, b and k it is NP- hard to decide if there is a k- sparse x such that Ax = b Why is this easier for incoherent dicoonaries? Uncertainty Principle: If A is μ- incoherent then Ay, Ax y, x provided that x and y are k- sparse, for k n/2μ
51 UNCERTAINTY PRINCIPLES Claim: Given A, b and k it is NP- hard to decide if there is a k- sparse x such that Ax = b Why is this easier for incoherent dicoonaries? Uncertainty Principle: If A is μ- incoherent then Ay, Ax y, x provided that x and y are k- sparse, for k n/2μ This principle can be used to establish uniqueness for sparse recovery, and things like b cannot be sparse in both standard and Fourier basis
52 A PAIR- WISE TEST
53 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support?
54 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support? supp(x) = supp(x ) =
55 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support? supp(x) = supp(x ) =
56 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support? supp(x) = supp(x ) = x', x zero non- zero maybe yes
57 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support? supp(x) = supp(x ) = x', x zero non- zero maybe yes x', x Ax, Ax Uncertainty Principle: for k- sparse x, incoherent A
58 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support? supp(x) = supp(x ) = x', x zero non- zero maybe yes x', x Ax, Ax zero non- zero maybe yes, whp Uncertainty Principle: for k- sparse x, incoherent A
59 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support? supp(x) = supp(x ) = Approach: Build a graph G on the p samples, with an edge btwn b and b if and only if b T b > 1/2
60 A PAIR- WISE TEST Given Ax = b and Ax = b, do x and x have intersecoon support? supp(x) = supp(x ) = Approach: Build a graph G on the p samples, with an edge btwn b and b if and only if b T b > 1/2 For the purposes of this talk, probability of an edge between b, b is ½ iff supp(x) and supp(x ) intersect
61 OVERLAPPING CLUSTERING Let C i ={ b x i 0} (overlapping)
62 OVERLAPPING CLUSTERING Let C i ={ b x i 0} (overlapping) Can we find the clusters efficiently?
63 OVERLAPPING CLUSTERING Let C i ={ b x i 0} (overlapping) Can we find the clusters efficiently? Challenge: Given (x, x, x ) where all the pairs belong to a cluster together, do all three belong to a common cluster too? supp(x) = supp(x ) = supp(x ) =
64 OVERLAPPING CLUSTERING Let C i ={ b x i 0} (overlapping) Can we find the clusters efficiently? Challenge: Given (x, x, x ) where all the pairs belong to a cluster together, do all three belong to a common cluster too?
65 OVERLAPPING CLUSTERING Let C i ={ b x i 0} (overlapping) Can we find the clusters efficiently? Challenge: Given (x, x, x ) where all the pairs belong to a cluster together, do all three belong to a common cluster too? supp(x) = supp(x ) = supp(x ) =
66 A TRIPLE TEST Key Idea: Use new samples y
67 A TRIPLE TEST Key Idea: Use new samples y Case #1: all three intersect: supp(x) = supp(x ) = supp(x ) =
68 A TRIPLE TEST Key Idea: Use new samples y Case #1: all three intersect: Probability y intersects all three is at least k/m supp(x) = supp(x ) = supp(x ) = New sample y only needs to contain one element from their joint union
69 A TRIPLE TEST Key Idea: Use new samples y
70 A TRIPLE TEST Key Idea: Use new samples y Case #2: no common intersecoon supp(x) = supp(x ) = supp(x ) = New sample y needs to contain at least two elements from their joint union
71 A TRIPLE TEST Key Idea: Use new samples y Case #2: no common intersecoon, supp(x) supp(x ) C, etc Probability y intersects all three is at most O(Ck 3 /m 2 ) supp(x) = supp(x ) = supp(x ) = New sample y needs to contain at least two elements from their joint union
72 A TRIPLE TEST Key Idea: Use new samples y Case #1: all three intersect: Probability y intersects all three is at least k/m Case #2: no common intersecoon, supp(x) supp(x ) C, etc Probability y intersects all three is at most O(Ck 3 /m 2 )
73 A TRIPLE TEST Key Idea: Use new samples y Case #1: all three intersect: Probability y intersects all three is at least k/m Case #2: no common intersecoon, supp(x) supp(x ) C, etc Probability y intersects all three is at most O(Ck 3 /m 2 ) Triple Test: Given (x, x, x ) where all the pairs intersect If there are at least T samples y where (x, x, x, y) all pairwise intersect, ACCEPT else REJECT
74 FINDING ALL THE CLUSTERS We can build a clustering algorithm on this primiove: For each pair (x, x ), find all x that pass the triple test
75 FINDING ALL THE CLUSTERS We can build a clustering algorithm on this primiove: For each pair (x, x ), find all x that pass the triple test Claim: This set is the union of clusters corresponding to supp(x) supp(x )
76 FINDING ALL THE CLUSTERS We can build a clustering algorithm on this primiove: For each pair (x, x ), find all x that pass the triple test Claim: This set is the union of clusters corresponding to supp(x) supp(x ) Claim: For every cluster i, there is some x, x that uniquely idenofy it i.e. supp(x) supp(x ) = {i}
77 FINDING ALL THE CLUSTERS We can build a clustering algorithm on this primiove: For each pair (x, x ), find all x that pass the triple test Claim: This set is the union of clusters corresponding to supp(x) supp(x ) Claim: For every cluster i, there is some x, x that uniquely idenofy it i.e. supp(x) supp(x ) = {i} Output inclusion- wise minimal sets these are the clusters!
78 FINDING ALL THE CLUSTERS We can build a clustering algorithm on this primiove: For each pair (x, x ), find all x that pass the triple test Claim: This set is the union of clusters corresponding to supp(x) supp(x ) Claim: For every cluster i, there is some x, x that uniquely idenofy it i.e. supp(x) supp(x ) = {i} Output inclusion- wise minimal sets these are the clusters! Our full algorithm uses higher- order tests; analysis through connec&ons to piercing number
79 Many ways to get the dic&onary from the clustering
80 Many ways to get the dic&onary from the clustering Approach #1: RelaOve Signs Plan: Refine C i and find all the b s with x i > 0
81 Many ways to get the dic&onary from the clustering Approach #1: RelaOve Signs Plan: Refine C i and find all the b s with x i > 0 Intui&on: If supp(x) supp(x ) = {i}, the we can find relaove sign of x i and x i and there are many such pairs
82 Many ways to get the dic&onary from the clustering Approach #1: RelaOve Signs Plan: Refine C i and find all the b s with x i > 0 Intui&on: If supp(x) supp(x ) = {i}, the we can find relaove sign of x i and x i and there are many such pairs enough so that whp we can find all relaove signs by transi&vity
83 Many ways to get the dic&onary from the clustering Approach #1: RelaOve Signs Plan: Refine C i and find all the b s with x i > 0 Intui&on: If supp(x) supp(x ) = {i}, the we can find relaove sign of x i and x i and there are many such pairs enough so that whp we can find all relaove signs by transi&vity Claim: E[b Ax = b and x i > 0] = A i E[x i x i > 0] Hence their empirical average converges to A i
84 Many ways to get the dic&onary from the clustering
85 Many ways to get the dic&onary from the clustering Approach #2: SVD Suppose we restrict to samples b with x i 0.
86 Many ways to get the dic&onary from the clustering Approach #2: SVD Suppose we restrict to samples b with x i 0. Intui&on: E[bb T x i 0] has large variance in direcoon of A i
87 Many ways to get the dic&onary from the clustering Approach #2: SVD Suppose we restrict to samples b with x i 0. Intui&on: E[bb T x i 0] has large variance in direcoon of A i We also show that alternaong minimizaoon works when we re close enough (geometric convergence)
88 OUTLINE Are there efficient algorithms for dicoonary learning? Introduc&on Origins of Sparse Recovery A StochasOc Model; Our Results Provable Algorithms via Overlapping Clustering Uncertainty Principles ReformulaOon as Overlapping Clustering Analyzing Alterna&ng Minimiza&on (out of &me) Gradient Descent on Non- Convex Fctns
89 A CONCEPTUAL OVERVIEW smoothed analysis Overcomplete Models computaoonal geometry NonnegaOve Matrix FactorizaOon Tensor DecomposiOons Be{er EsOmators spectral methods New Models Is Learning ComputaOonally Easy? Topic Models experiments Method of Moments polynomial equaoons e.g. mixtures of Gaussians
90 A CONCEPTUAL OVERVIEW smoothed analysis Overcomplete Models computaoonal geometry NonnegaOve Matrix FactorizaOon Tensor DecomposiOons Be{er EsOmators spectral methods New Models Is Learning ComputaOonally Easy? Topic Models experiments Natural Algorithms Method of Moments polynomial equaoons convex opomizaoon DicOonary Learning e.g. mixtures of Gaussians
91 Summary: Provable algorithms for learning incoherent, overcomplete dicoonaries ConnecOons to overlapping clustering Analysis of alternaong minimizaoon gradient descent on non- convex objecove Why does it work even from a random ini&aliza&on?
92 Any QuesOons? Summary: Provable algorithms for learning incoherent, overcomplete dicoonaries ConnecOons to overlapping clustering Analysis of alternaong minimizaoon gradient descent on non- convex objecove Why does it work even from a random ini&aliza&on?
NEW ALGORITHMS FOR DICTIONARY LEARNING
NEW ALGORITHMS FOR DICTIONARY LEARNING ANKUR MOITRA MASSACHUSETTS INSTITUTE OF TECHNOLOGY joint work with Sanjeev Arora, Rong Ge and Tengyu Ma SPARSE REPRESENTATIONS SPARSE REPRESENTATIONS Many data- types
More informationIntegral Geometry and the Polynomial Hirsch Conjecture
Integral Geometry and the Polynomial Hirsch Conjecture Jonathan Kelner, MIT Partially based on joint work with Daniel Spielman Introduction n A lot of recent work on Polynomial Hirsch Conjecture has focused
More informationBOOLEAN MATRIX FACTORIZATIONS. with applications in data mining Pauli Miettinen
BOOLEAN MATRIX FACTORIZATIONS with applications in data mining Pauli Miettinen MATRIX FACTORIZATIONS BOOLEAN MATRIX FACTORIZATIONS o THE BOOLEAN MATRIX PRODUCT As normal matrix product, but with addition
More informationSignal Reconstruction from Sparse Representations: An Introdu. Sensing
Signal Reconstruction from Sparse Representations: An Introduction to Compressed Sensing December 18, 2009 Digital Data Acquisition Suppose we want to acquire some real world signal digitally. Applications
More informationConvex Optimization MLSS 2015
Convex Optimization MLSS 2015 Constantine Caramanis The University of Texas at Austin The Optimization Problem minimize : f (x) subject to : x X. The Optimization Problem minimize : f (x) subject to :
More informationThe convex geometry of inverse problems
The convex geometry of inverse problems Benjamin Recht Department of Computer Sciences University of Wisconsin-Madison Joint work with Venkat Chandrasekaran Pablo Parrilo Alan Willsky Linear Inverse Problems
More informationSparse Reconstruction / Compressive Sensing
Sparse Reconstruction / Compressive Sensing Namrata Vaswani Department of Electrical and Computer Engineering Iowa State University Namrata Vaswani Sparse Reconstruction / Compressive Sensing 1/ 20 The
More informationELEG Compressive Sensing and Sparse Signal Representations
ELEG 867 - Compressive Sensing and Sparse Signal Representations Gonzalo R. Arce Depart. of Electrical and Computer Engineering University of Delaware Fall 211 Compressive Sensing G. Arce Fall, 211 1 /
More informationImage Restoration and Background Separation Using Sparse Representation Framework
Image Restoration and Background Separation Using Sparse Representation Framework Liu, Shikun Abstract In this paper, we introduce patch-based PCA denoising and k-svd dictionary learning method for the
More informationIntroduction to Topics in Machine Learning
Introduction to Topics in Machine Learning Namrata Vaswani Department of Electrical and Computer Engineering Iowa State University Namrata Vaswani 1/ 27 Compressed Sensing / Sparse Recovery: Given y :=
More informationSection 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1
Section 5 Convex Optimisation 1 W. Dai (IC) EE4.66 Data Proc. Convex Optimisation 1 2018 page 5-1 Convex Combination Denition 5.1 A convex combination is a linear combination of points where all coecients
More informationCS675: Convex and Combinatorial Optimization Spring 2018 The Simplex Algorithm. Instructor: Shaddin Dughmi
CS675: Convex and Combinatorial Optimization Spring 2018 The Simplex Algorithm Instructor: Shaddin Dughmi Algorithms for Convex Optimization We will look at 2 algorithms in detail: Simplex and Ellipsoid.
More informationSPARSE COMPONENT ANALYSIS FOR BLIND SOURCE SEPARATION WITH LESS SENSORS THAN SOURCES. Yuanqing Li, Andrzej Cichocki and Shun-ichi Amari
SPARSE COMPONENT ANALYSIS FOR BLIND SOURCE SEPARATION WITH LESS SENSORS THAN SOURCES Yuanqing Li, Andrzej Cichocki and Shun-ichi Amari Laboratory for Advanced Brain Signal Processing Laboratory for Mathematical
More informationOutlier Pursuit: Robust PCA and Collaborative Filtering
Outlier Pursuit: Robust PCA and Collaborative Filtering Huan Xu Dept. of Mechanical Engineering & Dept. of Mathematics National University of Singapore Joint w/ Constantine Caramanis, Yudong Chen, Sujay
More informationCPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2017
CPSC 340: Machine Learning and Data Mining Principal Component Analysis Fall 2017 Assignment 3: 2 late days to hand in tonight. Admin Assignment 4: Due Friday of next week. Last Time: MAP Estimation MAP
More informationLecture 2 September 3
EE 381V: Large Scale Optimization Fall 2012 Lecture 2 September 3 Lecturer: Caramanis & Sanghavi Scribe: Hongbo Si, Qiaoyang Ye 2.1 Overview of the last Lecture The focus of the last lecture was to give
More informationBlind Compressed Sensing Using Sparsifying Transforms
Blind Compressed Sensing Using Sparsifying Transforms Saiprasad Ravishankar and Yoram Bresler Department of Electrical and Computer Engineering and Coordinated Science Laboratory University of Illinois
More informationDetecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference
Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Minh Dao 1, Xiang Xiang 1, Bulent Ayhan 2, Chiman Kwan 2, Trac D. Tran 1 Johns Hopkins Univeristy, 3400
More informationCollaborative Sparsity and Compressive MRI
Modeling and Computation Seminar February 14, 2013 Table of Contents 1 T2 Estimation 2 Undersampling in MRI 3 Compressed Sensing 4 Model-Based Approach 5 From L1 to L0 6 Spatially Adaptive Sparsity MRI
More informationFMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu
FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)
More informationAs a consequence of the operation, there are new incidences between edges and triangles that did not exist in K; see Figure II.9.
II.4 Surface Simplification 37 II.4 Surface Simplification In applications it is often necessary to simplify the data or its representation. One reason is measurement noise, which we would like to eliminate,
More informationNote Set 4: Finite Mixture Models and the EM Algorithm
Note Set 4: Finite Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine Finite Mixture Models A finite mixture model with K components, for
More informationAMS : Combinatorial Optimization Homework Problems - Week V
AMS 553.766: Combinatorial Optimization Homework Problems - Week V For the following problems, A R m n will be m n matrices, and b R m. An affine subspace is the set of solutions to a a system of linear
More informationA fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009]
A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009] Yongjia Song University of Wisconsin-Madison April 22, 2010 Yongjia Song
More informationCombinatorial meets Convex Leveraging combinatorial structure to obtain faster algorithms, provably
Combinatorial meets Convex Leveraging combinatorial structure to obtain faster algorithms, provably Lorenzo Orecchia Department of Computer Science 6/30/15 Challenges in Optimization for Data Science 1
More informationLecture 17 Sparse Convex Optimization
Lecture 17 Sparse Convex Optimization Compressed sensing A short introduction to Compressed Sensing An imaging perspective 10 Mega Pixels Scene Image compression Picture Why do we compress images? Introduction
More informationCSC 411 Lecture 18: Matrix Factorizations
CSC 411 Lecture 18: Matrix Factorizations Roger Grosse, Amir-massoud Farahmand, and Juan Carrasquilla University of Toronto UofT CSC 411: 18-Matrix Factorizations 1 / 27 Overview Recall PCA: project data
More informationMachine Learning and Data Mining. Clustering (1): Basics. Kalev Kask
Machine Learning and Data Mining Clustering (1): Basics Kalev Kask Unsupervised learning Supervised learning Predict target value ( y ) given features ( x ) Unsupervised learning Understand patterns of
More informationSTA 4273H: Sta-s-cal Machine Learning
STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! h0p://www.cs.toronto.edu/~rsalakhu/ Lecture 3 Parametric Distribu>ons We want model the probability
More informationOptimal Denoising of Natural Images and their Multiscale Geometry and Density
Optimal Denoising of Natural Images and their Multiscale Geometry and Density Department of Computer Science and Applied Mathematics Weizmann Institute of Science, Israel. Joint work with Anat Levin (WIS),
More informationCombinatorial Selection and Least Absolute Shrinkage via The CLASH Operator
Combinatorial Selection and Least Absolute Shrinkage via The CLASH Operator Volkan Cevher Laboratory for Information and Inference Systems LIONS / EPFL http://lions.epfl.ch & Idiap Research Institute joint
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Clustering and EM Barnabás Póczos & Aarti Singh Contents Clustering K-means Mixture of Gaussians Expectation Maximization Variational Methods 2 Clustering 3 K-
More informationSubspace Clustering with Global Dimension Minimization And Application to Motion Segmentation
Subspace Clustering with Global Dimension Minimization And Application to Motion Segmentation Bryan Poling University of Minnesota Joint work with Gilad Lerman University of Minnesota The Problem of Subspace
More informationGraph Connectivity in Sparse Subspace Clustering
www.nicta.com.au From imagination to impact Graph Connectivity in Sparse Subspace Clustering Behrooz Nasihatkon 24 March 2011 Outline 1 Subspace Clustering 2 Sparse Subspace Clustering 3 Graph Connectivity
More informationCompressed Sensing Algorithm for Real-Time Doppler Ultrasound Image Reconstruction
Mathematical Modelling and Applications 2017; 2(6): 75-80 http://www.sciencepublishinggroup.com/j/mma doi: 10.11648/j.mma.20170206.14 ISSN: 2575-1786 (Print); ISSN: 2575-1794 (Online) Compressed Sensing
More informationCOMP251: Disjoint sets
COMP251: Disjoint sets Jérôme Waldispühl School of Computer Science McGill University Based on slides from M. Langer (McGill) Announces Assignment 1 is due on Wednesday February 1 st. Submit your soluoon
More informationShannon Switching Game
EECS 495: Combinatorial Optimization Lecture 1 Shannon s Switching Game Shannon Switching Game In the Shannon switching game, two players, Join and Cut, alternate choosing edges on a graph G. Join s objective
More information1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics
1. Introduction EE 546, Univ of Washington, Spring 2016 performance of numerical methods complexity bounds structural convex optimization course goals and topics 1 1 Some course info Welcome to EE 546!
More informationOutline. CS38 Introduction to Algorithms. Linear programming 5/21/2014. Linear programming. Lecture 15 May 20, 2014
5/2/24 Outline CS38 Introduction to Algorithms Lecture 5 May 2, 24 Linear programming simplex algorithm LP duality ellipsoid algorithm * slides from Kevin Wayne May 2, 24 CS38 Lecture 5 May 2, 24 CS38
More informationUnsupervised Learning: Clustering
Unsupervised Learning: Clustering Vibhav Gogate The University of Texas at Dallas Slides adapted from Carlos Guestrin, Dan Klein & Luke Zettlemoyer Machine Learning Supervised Learning Unsupervised Learning
More informationInverse Problems and Machine Learning
Inverse Problems and Machine Learning Julian Wörmann Research Group for Geometric Optimization and Machine Learning (GOL) 1 What are inverse problems? 2 Inverse Problems cause/ excitation 3 Inverse Problems
More informationK-Means and Gaussian Mixture Models
K-Means and Gaussian Mixture Models David Rosenberg New York University June 15, 2015 David Rosenberg (New York University) DS-GA 1003 June 15, 2015 1 / 43 K-Means Clustering Example: Old Faithful Geyser
More informationKernels and representation
Kernels and representation Corso di AA, anno 2017/18, Padova Fabio Aiolli 20 Dicembre 2017 Fabio Aiolli Kernels and representation 20 Dicembre 2017 1 / 19 (Hierarchical) Representation Learning Hierarchical
More informationIE598 Big Data Optimization Summary Nonconvex Optimization
IE598 Big Data Optimization Summary Nonconvex Optimization Instructor: Niao He April 16, 2018 1 This Course Big Data Optimization Explore modern optimization theories, algorithms, and big data applications
More informationRobust Principal Component Analysis (RPCA)
Robust Principal Component Analysis (RPCA) & Matrix decomposition: into low-rank and sparse components Zhenfang Hu 2010.4.1 reference [1] Chandrasekharan, V., Sanghavi, S., Parillo, P., Wilsky, A.: Ranksparsity
More information/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Approximation algorithms Date: 11/27/18
601.433/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Approximation algorithms Date: 11/27/18 22.1 Introduction We spent the last two lectures proving that for certain problems, we can
More informationClustering Lecture 5: Mixture Model
Clustering Lecture 5: Mixture Model Jing Gao SUNY Buffalo 1 Outline Basics Motivation, definition, evaluation Methods Partitional Hierarchical Density-based Mixture model Spectral methods Advanced topics
More informationStanford University CS359G: Graph Partitioning and Expanders Handout 1 Luca Trevisan January 4, 2011
Stanford University CS359G: Graph Partitioning and Expanders Handout 1 Luca Trevisan January 4, 2011 Lecture 1 In which we describe what this course is about. 1 Overview This class is about the following
More informationApproximation slides 1. An optimal polynomial algorithm for the Vertex Cover and matching in Bipartite graphs
Approximation slides 1 An optimal polynomial algorithm for the Vertex Cover and matching in Bipartite graphs Approximation slides 2 Linear independence A collection of row vectors {v T i } are independent
More informationImplicit Matrix Representations of Rational Bézier Curves and Surfaces
Implicit Matrix Representations of Rational Bézier Curves and Surfaces Laurent Busé INRIA Sophia Antipolis, France Laurent.Buse@inria.fr GD/SPM Conference, Denver, USA November 11, 2013 Overall motivation
More informationEfficient Iterative Semi-supervised Classification on Manifold
. Efficient Iterative Semi-supervised Classification on Manifold... M. Farajtabar, H. R. Rabiee, A. Shaban, A. Soltani-Farani Sharif University of Technology, Tehran, Iran. Presented by Pooria Joulani
More informationInterpolation & Polynomial Approximation. Cubic Spline Interpolation II
Interpolation & Polynomial Approximation Cubic Spline Interpolation II Numerical Analysis (9th Edition) R L Burden & J D Faires Beamer Presentation Slides prepared by John Carroll Dublin City University
More informationA Patch Prior for Dense 3D Reconstruction in Man-Made Environments
A Patch Prior for Dense 3D Reconstruction in Man-Made Environments Christian Häne 1, Christopher Zach 2, Bernhard Zeisl 1, Marc Pollefeys 1 1 ETH Zürich 2 MSR Cambridge October 14, 2012 A Patch Prior for
More informationSparse Models in Image Understanding And Computer Vision
Sparse Models in Image Understanding And Computer Vision Jayaraman J. Thiagarajan Arizona State University Collaborators Prof. Andreas Spanias Karthikeyan Natesan Ramamurthy Sparsity Sparsity of a vector
More informationCS Introduction to Data Mining Instructor: Abdullah Mueen
CS 591.03 Introduction to Data Mining Instructor: Abdullah Mueen LECTURE 8: ADVANCED CLUSTERING (FUZZY AND CO -CLUSTERING) Review: Basic Cluster Analysis Methods (Chap. 10) Cluster Analysis: Basic Concepts
More informationImage Restoration Using DNN
Image Restoration Using DNN Hila Levi & Eran Amar Images were taken from: http://people.tuebingen.mpg.de/burger/neural_denoising/ Agenda Domain Expertise vs. End-to-End optimization Image Denoising and
More informationIntroduc)on to Machine Learning: Clustering Algorithms How to Analyze Your Own Genome Fall 2013
Introduc)on to Machine Learning: Clustering Algorithms 02-223 How to Analyze Your Own Genome Fall 203 Organizing data into clusters such that there is high intra- cluster similarity low inter- cluster
More informationMeasurements and Bits: Compressed Sensing meets Information Theory. Dror Baron ECE Department Rice University dsp.rice.edu/cs
Measurements and Bits: Compressed Sensing meets Information Theory Dror Baron ECE Department Rice University dsp.rice.edu/cs Sensing by Sampling Sample data at Nyquist rate Compress data using model (e.g.,
More informationInterleaving Schemes on Circulant Graphs with Two Offsets
Interleaving Schemes on Circulant raphs with Two Offsets Aleksandrs Slivkins Department of Computer Science Cornell University Ithaca, NY 14853 slivkins@cs.cornell.edu Jehoshua Bruck Department of Electrical
More informationProgramming, numerics and optimization
Programming, numerics and optimization Lecture C-4: Constrained optimization Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428 June
More informationFace Recognition via Sparse Representation
Face Recognition via Sparse Representation John Wright, Allen Y. Yang, Arvind, S. Shankar Sastry and Yi Ma IEEE Trans. PAMI, March 2008 Research About Face Face Detection Face Alignment Face Recognition
More informationLecture notes on the simplex method September We will present an algorithm to solve linear programs of the form. maximize.
Cornell University, Fall 2017 CS 6820: Algorithms Lecture notes on the simplex method September 2017 1 The Simplex Method We will present an algorithm to solve linear programs of the form maximize subject
More informationINCOHERENT DICTIONARY LEARNING FOR SPARSE REPRESENTATION BASED IMAGE DENOISING
INCOHERENT DICTIONARY LEARNING FOR SPARSE REPRESENTATION BASED IMAGE DENOISING Jin Wang 1, Jian-Feng Cai 2, Yunhui Shi 1 and Baocai Yin 1 1 Beijing Key Laboratory of Multimedia and Intelligent Software
More information15-451/651: Design & Analysis of Algorithms October 11, 2018 Lecture #13: Linear Programming I last changed: October 9, 2018
15-451/651: Design & Analysis of Algorithms October 11, 2018 Lecture #13: Linear Programming I last changed: October 9, 2018 In this lecture, we describe a very general problem called linear programming
More informationClustering CS 550: Machine Learning
Clustering CS 550: Machine Learning This slide set mainly uses the slides given in the following links: http://www-users.cs.umn.edu/~kumar/dmbook/ch8.pdf http://www-users.cs.umn.edu/~kumar/dmbook/dmslides/chap8_basic_cluster_analysis.pdf
More informationMixture Models and the EM Algorithm
Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine c 2017 1 Finite Mixture Models Say we have a data set D = {x 1,..., x N } where x i is
More informationOff-the-Grid Compressive Imaging: Recovery of Piecewise Constant Images from Few Fourier Samples
Off-the-Grid Compressive Imaging: Recovery of Piecewise Constant Images from Few Fourier Samples Greg Ongie PhD Candidate Department of Applied Math and Computational Sciences University of Iowa April
More informationMatching and Covering
Matching and Covering Matchings A matching of size k in a graph G is a set of k pairwise disjoint edges The vertices belonging to the edges of a matching are saturated by the matching; the others are unsaturated
More information15-451/651: Design & Analysis of Algorithms November 4, 2015 Lecture #18 last changed: November 22, 2015
15-451/651: Design & Analysis of Algorithms November 4, 2015 Lecture #18 last changed: November 22, 2015 While we have good algorithms for many optimization problems, the previous lecture showed that many
More informationData Mining Learning from Large Data Sets
Data Mining Learning from Large Data Sets Lecture 8 Clustering large data sets 263-5200- 00L Andreas Krause Announcements! Homework 4 out tomorrow 2 Course organizapon! Retrieval! Given a query, find most
More informationMathematical and Algorithmic Foundations Linear Programming and Matchings
Adavnced Algorithms Lectures Mathematical and Algorithmic Foundations Linear Programming and Matchings Paul G. Spirakis Department of Computer Science University of Patras and Liverpool Paul G. Spirakis
More informationLecture 11: Clustering and the Spectral Partitioning Algorithm A note on randomized algorithm, Unbiased estimates
CSE 51: Design and Analysis of Algorithms I Spring 016 Lecture 11: Clustering and the Spectral Partitioning Algorithm Lecturer: Shayan Oveis Gharan May nd Scribe: Yueqi Sheng Disclaimer: These notes have
More informationLinear Programming. Readings: Read text section 11.6, and sections 1 and 2 of Tom Ferguson s notes (see course homepage).
Linear Programming Learning Goals. Introduce Linear Programming Problems. Widget Example, Graphical Solution. Basic Theory: Feasible Set, Vertices, Existence of Solutions. Equivalent formulations. Outline
More informationSpectral Clustering. Presented by Eldad Rubinstein Based on a Tutorial by Ulrike von Luxburg TAU Big Data Processing Seminar December 14, 2014
Spectral Clustering Presented by Eldad Rubinstein Based on a Tutorial by Ulrike von Luxburg TAU Big Data Processing Seminar December 14, 2014 What are we going to talk about? Introduction Clustering and
More informationLinear Programming. Widget Factory Example. Linear Programming: Standard Form. Widget Factory Example: Continued.
Linear Programming Widget Factory Example Learning Goals. Introduce Linear Programming Problems. Widget Example, Graphical Solution. Basic Theory:, Vertices, Existence of Solutions. Equivalent formulations.
More informationRobot Mapping. Least Squares Approach to SLAM. Cyrill Stachniss
Robot Mapping Least Squares Approach to SLAM Cyrill Stachniss 1 Three Main SLAM Paradigms Kalman filter Particle filter Graphbased least squares approach to SLAM 2 Least Squares in General Approach for
More informationGraphbased. Kalman filter. Particle filter. Three Main SLAM Paradigms. Robot Mapping. Least Squares Approach to SLAM. Least Squares in General
Robot Mapping Three Main SLAM Paradigms Least Squares Approach to SLAM Kalman filter Particle filter Graphbased Cyrill Stachniss least squares approach to SLAM 1 2 Least Squares in General! Approach for
More informationAMS /672: Graph Theory Homework Problems - Week V. Problems to be handed in on Wednesday, March 2: 6, 8, 9, 11, 12.
AMS 550.47/67: Graph Theory Homework Problems - Week V Problems to be handed in on Wednesday, March : 6, 8, 9,,.. Assignment Problem. Suppose we have a set {J, J,..., J r } of r jobs to be filled by a
More informationFacial Expression Classification with Random Filters Feature Extraction
Facial Expression Classification with Random Filters Feature Extraction Mengye Ren Facial Monkey mren@cs.toronto.edu Zhi Hao Luo It s Me lzh@cs.toronto.edu I. ABSTRACT In our work, we attempted to tackle
More informationEfficient Imaging Algorithms on Many-Core Platforms
Efficient Imaging Algorithms on Many-Core Platforms H. Köstler Dagstuhl, 22.11.2011 Contents Imaging Applications HDR Compression performance of PDE-based models Image Denoising performance of patch-based
More informationMethods for Intelligent Systems
Methods for Intelligent Systems Lecture Notes on Clustering (II) Davide Eynard eynard@elet.polimi.it Department of Electronics and Information Politecnico di Milano Davide Eynard - Lecture Notes on Clustering
More informationLarge-Scale Lasso and Elastic-Net Regularized Generalized Linear Models
Large-Scale Lasso and Elastic-Net Regularized Generalized Linear Models DB Tsai Steven Hillion Outline Introduction Linear / Nonlinear Classification Feature Engineering - Polynomial Expansion Big-data
More informationWhen Sparsity Meets Low-Rankness: Transform Learning With Non-Local Low-Rank Constraint For Image Restoration
When Sparsity Meets Low-Rankness: Transform Learning With Non-Local Low-Rank Constraint For Image Restoration Bihan Wen, Yanjun Li and Yoram Bresler Department of Electrical and Computer Engineering Coordinated
More informationMathematical Programming Methods for Minimizing the Zero-norm over Polyhedral Sets. Francesco Rinaldi
Sapienza, University of Rome Mathematical Programming Methods for Minimizing the Zero-norm over Polyhedral Sets Francesco Rinaldi Dissertation for the Degree of Philosophy Doctor in Operations Research
More informationStructurally Random Matrices
Fast Compressive Sampling Using Structurally Random Matrices Presented by: Thong Do (thongdo@jhu.edu) The Johns Hopkins University A joint work with Prof. Trac Tran, The Johns Hopkins University it Dr.
More informationUnderstanding Clustering Supervising the unsupervised
Understanding Clustering Supervising the unsupervised Janu Verma IBM T.J. Watson Research Center, New York http://jverma.github.io/ jverma@us.ibm.com @januverma Clustering Grouping together similar data
More informationarxiv: v1 [math.co] 15 Dec 2009
ANOTHER PROOF OF THE FACT THAT POLYHEDRAL CONES ARE FINITELY GENERATED arxiv:092.2927v [math.co] 5 Dec 2009 VOLKER KAIBEL Abstract. In this note, we work out a simple inductive proof showing that every
More informationAn incomplete survey of matrix completion. Benjamin Recht Department of Computer Sciences University of Wisconsin-Madison
An incomplete survey of matrix completion Benjamin Recht Department of Computer Sciences University of Wisconsin-Madison Abstract Setup: Matrix Completion M = M ij known for black cells M ij unknown for
More informationAlgebraic Iterative Methods for Computed Tomography
Algebraic Iterative Methods for Computed Tomography Per Christian Hansen DTU Compute Department of Applied Mathematics and Computer Science Technical University of Denmark Per Christian Hansen Algebraic
More informationParallelization in the Big Data Regime 5: Data Parallelization? Sham M. Kakade
Parallelization in the Big Data Regime 5: Data Parallelization? Sham M. Kakade Machine Learning for Big Data CSE547/STAT548 University of Washington S. M. Kakade (UW) Optimization for Big data 1 / 23 Announcements...
More informationIteratively Re-weighted Least Squares for Sums of Convex Functions
Iteratively Re-weighted Least Squares for Sums of Convex Functions James Burke University of Washington Jiashan Wang LinkedIn Frank Curtis Lehigh University Hao Wang Shanghai Tech University Daiwei He
More informationNP-Complete Reductions 2
x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 12 22 32 CS 447 11 13 21 23 31 33 Algorithms NP-Complete Reductions 2 Prof. Gregory Provan Department of Computer Science University College Cork 1 Lecture Outline NP-Complete
More informationSparse wavelet expansions for seismic tomography: Methods and algorithms
Sparse wavelet expansions for seismic tomography: Methods and algorithms Ignace Loris Université Libre de Bruxelles International symposium on geophysical imaging with localized waves 24 28 July 2011 (Joint
More informationOptimization for Machine Learning
with a focus on proximal gradient descent algorithm Department of Computer Science and Engineering Outline 1 History & Trends 2 Proximal Gradient Descent 3 Three Applications A Brief History A. Convex
More informationby conservation of flow, hence the cancelation. Similarly, we have
Chapter 13: Network Flows and Applications Network: directed graph with source S and target T. Non-negative edge weights represent capacities. Assume no edges into S or out of T. (If necessary, we can
More informationClustering: Centroid-Based Partitioning
Clustering: Centroid-Based Partitioning Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong 1 / 29 Y Tao Clustering: Centroid-Based Partitioning In this lecture, we
More informationRecovery of Piecewise Smooth Images from Few Fourier Samples
Recovery of Piecewise Smooth Images from Few Fourier Samples Greg Ongie*, Mathews Jacob Computational Biomedical Imaging Group (CBIG) University of Iowa SampTA 2015 Washington, D.C. 1. Introduction 2.
More informationImage Deblurring Using Adaptive Sparse Domain Selection and Adaptive Regularization
Volume 3, No. 3, May-June 2012 International Journal of Advanced Research in Computer Science RESEARCH PAPER Available Online at www.ijarcs.info ISSN No. 0976-5697 Image Deblurring Using Adaptive Sparse
More informationPareto Optimal Solutions for Smoothed Analysts
Pareto Optimal Solutions for Smoothed Analysts Ankur Moitra, IAS joint work with Ryan O Donnell, CMU September 26, 2011 Ankur Moitra (IAS) Pareto September 26, 2011 values weights v1 w 1...... v i w i
More informationSparse Solutions to Linear Inverse Problems. Yuzhe Jin
Sparse Solutions to Linear Inverse Problems Yuzhe Jin Outline Intro/Background Two types of algorithms Forward Sequential Selection Methods Diversity Minimization Methods Experimental results Potential
More information