Nonparametric Risk Attribution for Factor Models of Portfolios. October 3, 2017 Kellie Ottoboni

Similar documents
Moving Beyond Linearity

Nonparametric Approaches to Regression

Moving Beyond Linearity

Lecture 27, April 24, Reading: See class website. Nonparametric regression and kernel smoothing. Structured sparse additive models (GroupSpAM)

Nonparametric Mixed-Effects Models for Longitudinal Data

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

Last time... Bias-Variance decomposition. This week

Lecture 9: Support Vector Machines

Mar. 20 Math 2335 sec 001 Spring 2014

STAT 705 Introduction to generalized additive models

CPSC 340: Machine Learning and Data Mining. Regularization Fall 2016

CPSC 340: Machine Learning and Data Mining. More Regularization Fall 2017

Generalized Additive Model

Generalized Additive Models

Nonparametric and Semiparametric Econometrics Lecture Notes for Econ 221. Yixiao Sun Department of Economics, University of California, San Diego

Machine Learning / Jan 27, 2010

Homework. Gaussian, Bishop 2.3 Non-parametric, Bishop 2.5 Linear regression Pod-cast lecture on-line. Next lectures:

Generalized additive models I

Section 4 Matching Estimator

Notes and Announcements

Model learning for robot control: a survey

Optimization Methods for Machine Learning (OMML)

Applying Supervised Learning

Spatial Interpolation & Geostatistics

Chapter 7: Dual Modeling in the Presence of Constant Variance

Goals of the Lecture. SOC6078 Advanced Statistics: 9. Generalized Additive Models. Limitations of the Multiple Nonparametric Models (2)

APPM/MATH Problem Set 4 Solutions

SUPERVISED LEARNING METHODS. Stanley Liang, PhD Candidate, Lassonde School of Engineering, York University Helix Science Engagement Programs 2018

Nonparametric Methods Recap

Spatial Interpolation - Geostatistics 4/3/2018

Adaptive Kernel Regression for Image Processing and Reconstruction

CS 229 Midterm Review

LOESS curve fitted to a population sampled from a sine wave with uniform noise added. The LOESS curve approximates the original sine wave.

Overview of various smoothers

Bootstrapping Method for 14 June 2016 R. Russell Rhinehart. Bootstrapping

Lecture 16: High-dimensional regression, non-linear regression

Lecture 24: Generalized Additive Models Stat 704: Data Analysis I, Fall 2010

NONPARAMETRIC REGRESSION TECHNIQUES

Lecture VIII. Global Approximation Methods: I

A popular method for moving beyond linearity. 2. Basis expansion and regularization 1. Examples of transformations. Piecewise-polynomials and splines

6. Linear Discriminant Functions

Topics in Machine Learning

Knowledge Discovery and Data Mining

CPSC 340: Machine Learning and Data Mining. Kernel Trick Fall 2017

Model Fitting, RANSAC. Jana Kosecka

Economics Nonparametric Econometrics

A Survey of Regression Methods for Proxy Functions

CPSC 340: Machine Learning and Data Mining. More Linear Classifiers Fall 2017

Robust Regression. Robust Data Mining Techniques By Boonyakorn Jantaranuson

Preface to the Second Edition. Preface to the First Edition. 1 Introduction 1

Spline Models. Introduction to CS and NCS. Regression splines. Smoothing splines

08 An Introduction to Dense Continuous Robotic Mapping

EE613 Machine Learning for Engineers LINEAR REGRESSION. Sylvain Calinon Robot Learning & Interaction Group Idiap Research Institute Nov.

Vision par ordinateur

Lecture #11: The Perceptron

Lecture 06 Decision Trees I

Introduction to ANSYS DesignXplorer

Spline fitting tool for scientometric applications: estimation of citation peaks and publication time-lags

Nonparametric regression using kernel and spline methods

Nonparametric Regression

Isosurface Visualization of Data with Nonparametric Models for Uncertainty

Lecture 17: Smoothing splines, Local Regression, and GAMs

Computational Physics PHYS 420

What is machine learning?

Computer Graphics Curves and Surfaces. Matthias Teschner

Applied Statistics : Practical 9

Predictive Analysis: Evaluation and Experimentation. Heejun Kim

Broad field that includes low-level operations as well as complex high-level algorithms

MODEL SELECTION AND REGULARIZATION PARAMETER CHOICE

Machine Learning. Topic 4: Linear Regression Models

22s:152 Applied Linear Regression. Chapter 18: Nonparametric Regression (Lowess smoothing)

Fitting. Fitting. Slides S. Lazebnik Harris Corners Pkwy, Charlotte, NC

Linear Regression and K-Nearest Neighbors 3/28/18

Apprenticeship Learning for Reinforcement Learning. with application to RC helicopter flight Ritwik Anand, Nick Haliday, Audrey Huang

Classification: Linear Discriminant Functions

Morningstar Direct SM Equity Attribution

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

CPSC 340: Machine Learning and Data Mining. Logistic Regression Fall 2016

Doubly Cyclic Smoothing Splines and Analysis of Seasonal Daily Pattern of CO2 Concentration in Antarctica

Separation of Surface Roughness Profile from Raw Contour based on Empirical Mode Decomposition Shoubin LIU 1, a*, Hui ZHANG 2, b

Bivariate Linear Regression James M. Murray, Ph.D. University of Wisconsin - La Crosse Updated: October 04, 2017

Inter-Data-Center Network Traffic Prediction with Elephant Flows

Using Multivariate Adaptive Regression Splines (MARS ) to enhance Generalised Linear Models. Inna Kolyshkina PriceWaterhouseCoopers

Withdrawn Equity Offerings: Event Study and Cross-Sectional Regression Analysis Using Eventus Software

A review of spline function selection procedures in R

Fitting to a set of data. Lecture on fitting

The K-modes and Laplacian K-modes algorithms for clustering

Locally Weighted Learning for Control. Alexander Skoglund Machine Learning Course AASS, June 2005

Splines and penalized regression

Basis Functions. Volker Tresp Summer 2017

1D Regression. i.i.d. with mean 0. Univariate Linear Regression: fit by least squares. Minimize: to get. The set of all possible functions is...

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2017

Generalized Additive Models

Module 4. Non-linear machine learning econometrics: Support Vector Machine

Cluster EMD and its Statistical Application

CSE 417T: Introduction to Machine Learning. Lecture 22: The Kernel Trick. Henry Chai 11/15/18

Analysis of Panel Data. Third Edition. Cheng Hsiao University of Southern California CAMBRIDGE UNIVERSITY PRESS

Non-Linear Regression. Business Analytics Practice Winter Term 2015/16 Stefan Feuerriegel

Spatial Interpolation & Geostatistics

Clustering. Discover groups such that samples within a group are more similar to each other than samples across groups.

Transcription:

Nonparametric Risk Attribution for Factor Models of Portfolios October 3, 2017 Kellie Ottoboni

Outline The problem Page 3 Additive model of returns Page 7 Euler s formula for risk decomposition Page 11 Nonparametric regression Page 13 Portfolio example Page 17 2

The problem Risk decomposition Factor models are used to forecast returns of a portfolio with known positions in underlying assets These models are used to predict risk Volatility, value-at-risk (VaR), and expected tail loss (ETL) Clients would often like to identify sources of risk in their portfolio Goal: divide up the overall risk among the factors 3

The problem Risk decomposition A function is homogeneous of order one if VaR and ETL have this property. VaR: the upper alpha quantile of the loss distribution ETL: the conditional expectation of the upper alpha tail of the loss distribution Euler s formula If a portfolio is comprised of K assets with known weights, and the risk measure is homogeneous of order one, then 4

The problem Risk decomposition What s wrong with using Euler s formula? If the portfolio is a nonlinear function of factors, then this additive decomposition does not hold. Nonlinear instruments (e.g. bonds, derivatives) Simple vs. log returns An asset s marginal contribution to risk is not its standalone risk Incremental contribution to risk is We have to estimate derivatives 5

Solution Simulate factor and portfolio returns Create synthetic factors Estimate risk contributions The portfolio is nonlinear in the factors. Previous results for partitioning portfolio risk do not apply. The portfolio is linear in the synthetic factors. We estimate them using the generalized additive model framework. The additive decomposition applies. The portfolio risk is partitioned amongst nonlinear transformations of the original factors. All estimation can be done nonparametrically: we don t need to make any distributional assumptions. 6

Additive models An algorithm for creating synthetic factors which are linearly related to portfolio returns.

Additive models Problem Factors influence the underlying assets which make up the portfolio. A linear relationship between factor and portfolio returns would allow us to use the Euler decomposition to attribute risk to factors. If the portfolio were made up of solely linear instruments (e.g. stocks), then we could estimate the factor contributions using linear regression But, factors influence portfolio returns in a nonlinear way when other instruments are present so this may be a bad approximation. 8

Additive models Solution If factors are independent, then there exists an additive decomposition called the Hajek projection: are measurable functions with finite second moments They can take any form and may vary for each factor. Idea: Create synthetic factors Then the portfolio return is a sum of synthetic factors 9

Additive models Estimation Iterate through factors and estimate in sequence, since they are independent Estimate relationship between the portfolio returns and the residuals of predicted returns using all other factors Smooth can be any regression method. We ll discuss specific methods later. 10

Euler s formula A theorem that allows us to decompose the risk of a portfolio into a sum of risk contributions

Euler s formula Solution Euler s formula: How do we calculate these derivatives? Hallerbach (1999), Tasche (1999), and Gourieroux et al. (2000), and others show that these partial derivatives can also be expressed as expected values: The problem of taking derivatives is just a regression problem! 12

Nonparametric regression Moving beyond least squares to find a best fit to the data.

Nonparametric regression Problem Linear regression is too rigid. Not all problems are linear! Idea: modify the least squares optimization problem to give more robust solutions We use these methods in both estimation steps: The smoother regression in creating synthetic factors The expected value in estimating the marginal risk contributions 14

Nonparametric regression Kernel methods Idea: do a separate regression at each input value, giving more weight to nearby observations Weighting is done using a kernel, a function made up of a distance between inputs and bandwidth h If x and y are far apart, the weight is small Small values of h lead to more smoothing near and far points have similar weights Kernel regression is a weighted linear regression problem at each x The solution is locally linear 15

Nonparametric regression Smoothing splines Idea: fit a cubic spline to the data, but penalize the roughness so it isn t too wiggly Cubic splines interpolate the data by putting cubic polynomials between every pair of points The second derivative of the function indicates roughness. Small = smooth The loss function trades off fit with roughness: Solution can be written in closed form, so it is fast to compute 16

Example Evaluating the method on simulated data from GX Labs.

Example Portfolio Standard December 2009 proxy 50-50 portfolio 30 factors First 20 are equity factors, with the 20 th being most important Last 10 are a mix of fixed-income and equity factors 10,000 simulation paths 18

Example Methods Step 1: create synthetic factors using smoothing splines OR use raw factors Step 2: estimate risk contributions using smoothing splines, OLS, or formulas derived from the Gaussian distribution Some comments: All splines were fit using default tuning parameters Local linear regression was order of magnitudes slower Estimated VaR and ETL at alpha=0.05 Measure the following Error: sum of risk contributions divided by estimated portfolio risk Time: seconds to run the two steps for VaR and ETL 19

Example Comparison of performance Method VaR Ratio ETL Ratio Time (s) Synthetic factors + smoothing splines 0.810 0.356 13.475 Synthetic factors + OLS 0.921 0.221 14.830 Synthetic factors + Gaussian formulas 0.219 0.219 14.119 Raw factors + OLS 0.002 0.168 0.055 Raw factors + Gaussian formulas 0.168 1.002 0.047 20

Example Synthetic Factors 21

Example Portfolio VaR Estimated risk contributions from smoothing splines + smoothing splines 22

Example Portfolio ETL Estimated risk contributions from smoothing splines + smoothing splines 23

Discussion Using smoothing splines to explain the relationship between factors and portfolio is computationally fast and more accurate than parametric methods that assume linearity or Gaussianity. Results make sense: what we believe a priori are the most important factors have the highest contribution to risk Interpretability is difficult Risk is attributed to synthetic factors, not raw factors The relationship between factors and synthetic factors may not be monotonic, invertible, etc. Methods assume factors are independent Instead of using the additive decomposition, one may use a Hoeffding decomposition which includes cross-terms. See Rosen and Saunders (2010). 24