Local Linear Embedding. Katelyn Stringer ASTR 689 December 1, 2015

Similar documents
Recognizing Handwritten Digits Using the LLE Algorithm with Back Propagation

Non-linear dimension reduction

Evgeny Maksakov Advantages and disadvantages: Advantages and disadvantages: Advantages and disadvantages: Advantages and disadvantages:

Large-Scale Face Manifold Learning

Dimension Reduction CS534

Dimension Reduction of Image Manifolds

Remote Sensing Data Classification Using Combined Spectral and Spatial Local Linear Embedding (CSSLE)

Robust Pose Estimation using the SwissRanger SR-3000 Camera

Face detection and recognition. Many slides adapted from K. Grauman and D. Lowe

Locality Preserving Projections (LPP) Abstract

The Curse of Dimensionality

Differential Structure in non-linear Image Embedding Functions

Head Frontal-View Identification Using Extended LLE

Data preprocessing Functional Programming and Intelligent Algorithms

Discriminate Analysis

SYDE Winter 2011 Introduction to Pattern Recognition. Clustering

Recognition: Face Recognition. Linda Shapiro EE/CSE 576

Locality Preserving Projections (LPP) Abstract

K-Local Hyperplane and Convex Distance Nearest Neighbor Algorithms

Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Generalized Principal Component Analysis CVPR 2007

Nonlinear Dimensionality Reduction Applied to the Classification of Images

Unsupervised Learning

COMPRESSED DETECTION VIA MANIFOLD LEARNING. Hyun Jeong Cho, Kuang-Hung Liu, Jae Young Park. { zzon, khliu, jaeypark

A Course in Machine Learning

Locally Linear Landmarks for large-scale manifold learning

Iterative Non-linear Dimensionality Reduction by Manifold Sculpting

Work 2. Case-based reasoning exercise

( ) =cov X Y = W PRINCIPAL COMPONENT ANALYSIS. Eigenvectors of the covariance matrix are the principal components

Image Similarities for Learning Video Manifolds. Selen Atasoy MICCAI 2011 Tutorial

Face detection and recognition. Detection Recognition Sally

Laplacian Faces: A Face Recognition Tool

SELECTION OF THE OPTIMAL PARAMETER VALUE FOR THE LOCALLY LINEAR EMBEDDING ALGORITHM. Olga Kouropteva, Oleg Okun and Matti Pietikäinen

Model Fitting. Introduction to Computer Vision CSE 152 Lecture 11

Non-linear CCA and PCA by Alignment of Local Models

Metric Learning Applied for Automatic Large Image Classification

CSE 258 Lecture 5. Web Mining and Recommender Systems. Dimensionality Reduction

Neural Networks for Machine Learning. Lecture 15a From Principal Components Analysis to Autoencoders

Anomaly Detection on Data Streams with High Dimensional Data Environment

Learning a Manifold as an Atlas Supplementary Material

2. Data Preprocessing

Globally and Locally Consistent Unsupervised Projection

Part I. Hierarchical clustering. Hierarchical Clustering. Hierarchical clustering. Produces a set of nested clusters organized as a

Robust PDF Table Locator

DIMENSION REDUCTION FOR HYPERSPECTRAL DATA USING RANDOMIZED PCA AND LAPLACIAN EIGENMAPS

Local Features: Detection, Description & Matching

CSC 411: Lecture 14: Principal Components Analysis & Autoencoders

Robotics Programming Laboratory

Network Traffic Measurements and Analysis

10-701/15-781, Fall 2006, Final

CSE 255 Lecture 5. Data Mining and Predictive Analytics. Dimensionality Reduction

Manifold Clustering. Abstract. 1. Introduction

CSC 411: Lecture 14: Principal Components Analysis & Autoencoders

Applying Supervised Learning

Robot Manifolds for Direct and Inverse Kinematics Solutions

Finding Structure in CyTOF Data

Hierarchical Clustering 4/5/17

FIFI-LS: Basic Cube Analysis using SOSPEX

Robust Kernel Methods in Clustering and Dimensionality Reduction Problems

Homework 4: Clustering, Recommenders, Dim. Reduction, ML and Graph Mining (due November 19 th, 2014, 2:30pm, in class hard-copy please)

Digital Image Processing Chapter 11: Image Description and Representation

Feature selection. Term 2011/2012 LSI - FIB. Javier Béjar cbea (LSI - FIB) Feature selection Term 2011/ / 22

A Taxonomy of Semi-Supervised Learning Algorithms

Unsupervised Learning

A Stochastic Optimization Approach for Unsupervised Kernel Regression

RDRToolbox A package for nonlinear dimension reduction with Isomap and LLE.

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

Lecture Topic Projects

Modelling and Visualization of High Dimensional Data. Sample Examination Paper

CS6716 Pattern Recognition

Cluster Analysis. Mu-Chun Su. Department of Computer Science and Information Engineering National Central University 2003/3/11 1

Last week. Multi-Frame Structure from Motion: Multi-View Stereo. Unknown camera viewpoints

Face Detection and Alignment. Prof. Xin Yang HUST

Announcements. Recognition I. Gradient Space (p,q) What is the reflectance map?

Face Recognition with Weighted Locally Linear Embedding

Robust Locally Linear Embedding

InterAxis: Steering Scatterplot Axes via Observation-Level Interaction

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points]

MUSI-6201 Computational Music Analysis

04 - Normal Estimation, Curves

Lecture 26: Missing data

3. Data Preprocessing. 3.1 Introduction

Panoramic Image Stitching

Surface Reconstruction with MLS

Nonlinear Dimensionality Reduction Applied to the Binary Classification of Images

Solution Sketches Midterm Exam COSC 6342 Machine Learning March 20, 2013

LOCAL LINEAR APPROXIMATION OF PRINCIPAL CURVE PROJECTIONS. Peng Zhang, Esra Ataer-Cansizoglu, Deniz Erdogmus

MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER

CSE 158. Web Mining and Recommender Systems. Midterm recap

Dimension reduction for hyperspectral imaging using laplacian eigenmaps and randomized principal component analysis

Digital Geometry Processing

CIE L*a*b* color model

Scale Invariant Feature Transform

Raman Images. Jeremy M. Shaver 1, Eunah Lee 2, Andrew Whitley 2, R. Scott Koch. 1. Eigenvector Research, Inc. 2. HORIBA Jobin Yvon, Inc.

Overview. Efficient Simplification of Point-sampled Surfaces. Introduction. Introduction. Neighborhood. Local Surface Analysis

Manifold Preserving Edit Propagation

Spectral Surface Reconstruction from Noisy Point Clouds

The Artifact Subspace Reconstruction method. Christian A Kothe SCCN / INC / UCSD January 2013

10/14/2017. Dejan Sarka. Anomaly Detection. Sponsors

Feature Detectors and Descriptors: Corners, Lines, etc.

Transcription:

Local Linear Embedding Katelyn Stringer ASTR 689 December 1, 2015

Idea Behind LLE Good at making nonlinear high-dimensional data easier for computers to analyze Example: A high-dimensional surface Think derivatives: local tangential hyperplane is a good approximation LLE records locations of points on this local tangential manifold based on locations of neighboring points It s like describing your address in terms of how far your house is from other buildings

Idea Behind LLE Good at making nonlinear high-dimensional data easier for computers to analyze Example: A high-dimensional surface Think derivatives: local tangential hyperplane is a good approximation WHY? LLE records locations of points on this local tangential manifold based on locations of neighboring points It s like describing your address in terms of how far your house is from other buildings

1

2

2 3

3 2 2

Reducing the Dimensionality of Data: Local Linear Embedding of Sloan Galaxy Spectra Jake Vanderplas, Andrew Connolly (University of Washington) 4 5 *All of the following images and information came from this paper unless otherwise noted!*

The LLE Algorithm You have a set of data vectors Want to map them to coordinate system For each data vector x i, the indices of the K nearest neighbors are represented by

LLE Algorithm, cont. Assume that each point x i lies near a locally linear low dimensional manifold Find K nearest neighbors (based on Euclidean distance) Find the local covariance matrix ( is the j th nearest neighbor to x i ) D in d pq = (X ps X qs ) 2 Regularize covariance matrix to produce a stable solution s Matrix element (authors used )

LLE Algorithm, cont. Determine the optimal weights for each nearest neighbor by minimizing the reconstruction error Impose Solve for eigenvalues Or solve for w (i) and scale to 1

LLE Algorithm, cont. Create overarching W matrix with all the weight vectors as columns, W ji = 0 if point j is not a nearest neighbor of point i Create matrix Solve for eigenvalues of Eigenvectors y corresponding to D out +1 lowest eigenvalues are the new basis vectors (except the first, λ=0 just gives a translation)

LLE Algorithm, cont. Errors are given by cost functions:

LLE Algorithm, Recap Find nearest neighbors Find weights to map each point in terms of its neighbors Find embedded coords Use weights to map points in new coords 1

Our Sloan Spectra Sample 8711 total spectra, z < 0.36 1000 logarithmic wavelength bins (3800-9800 Å) Corrected for sky absorption & normalized Equivalent widths, line positions in headers Prior Classifications: o QSOs = Quasi-Stellar Objects, include Quasars Broad line Narrow line o Emission Galaxies o Quiescent Galaxies o Absorption Galaxies

QSOs: What we look for Hydrogen emission lines 3x > noise Broad line: larger redshift o Line widths > 1200 km/s Narrow line: smaller redshift

Galaxies Emission: Star-forming galaxies o Hydrogen emission > 3 x noise Absorption: Balmer absorption > 3σ Quiescent: red and dead, Balmer emission < 3σ

Apply LLE

Progression of Emission Galaxy Spectra

Progression of Broad-Line QSO Spectra

Progression of Narrow- Line QSO Spectra

Progression from Broad to Narrow-Line QSOs

Progression of Quiescent Galaxy Spectra

Objects Misclassified by Sloan

Outliers Found with LLE How to keep outliers from skewing your results?

Must Account for Outliers!

Must Account for Outliers! 2

Must Account for Outliers! 2

Must Account for Outliers! 2

Robust LLE (RLLE) Assign a reliability score to each data vector Outliers will not: o be a part of many local neighborhoods o lie near the best-fit hyperplane Perform an iteratively reweighted least-squares reduction on each data point to determine optimal weights for PCA reconstruction of data o Like finding best-fit hyperplane and determining each point s distance from it o Result: assigns local weights to each point for its contribution to local tangent space Sum all the local weights for each neighborhood Low reliability scores = outliers

Robust LLE (RLLE)

Robust LLE (RLLE) 6

Choosing the Value of K Too small: undersampled, lose too much information o results easily skewed by outliers and noise Too large: manifold is oversampled, can no longer assume local linearity o cannot reduce the dimensions as much How do we find the Goldilocks K value?

Finding K: Trial and Error Tried values from K=10 to K=30. Goal: Maximize the angle between the QSOs and the Emission galaxies Optimal K=23

Better than Other Methods? PCA- good at showing variance, weaknesses: nonlinear effects, Behavior of individual lines Line Ratio Diagramsgood w/ emission lines, weaknesses: ignores continuum, works poorly w/ low S/N

High computational costs Bottlenecks: Drawbacks o Nearest neighbor search o Calculating optimal projection vectors Strategies: sample your data into smaller subsets o Vanderplas and Connolly did this! (~150,000 ~9,000 spectra)

Sources A. Roweis, S., & Saul, L. 2000, Science, 290, 2323 B. Vanderplas, Jake, and Andrew Connolly. 2009. REDUCING THE DIMENSIONALITY OF DATA : LOCALLY LINEAR EMBEDDING OF SLOAN GALAXY SPECTRA, 1365 79. doi:10.1088/0004-6256/138/5/1365. C. Saul, Lawrence K, and Sam T Roweis. n.d. An Introduction to Locally Linear Embedding, https://www.cs.nyu.edu/~roweis/lle/papers/lleintro.pdf.

Image Credits 1. http://web.stanford.edu/class/ee378b/papers/roweis.pdf 2. http://iopscience.iop.org/article/10.1088/0004-6256/138/5/1365/pdf 3. https://i.imgur.com/96t6i9mh.jpg 4. https://twitter.com/jakevdp 5. http://www.washington.edu/storycentral/story/scanningthe-sky/connolly-andrew-11/ 6. http://theactionelite.com/2012/07/the-return-of-chucknorris/