Distributed Optimization via ADMM
|
|
- Vivian Evans
- 5 years ago
- Views:
Transcription
1 Distributed Optimization via ADMM Zhimin Peng Dept. Computational and Applied Mathematics Rice University Houston, TX Aug. 15, 2012 Main Reference: Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers. S. Boyd, N. Parikh, E. Chu, B. Peleato, and J. Eckstein Zhimin ADMM 1
2 Outline RICE 1 Alternating direction method of multipliers 2 Distributed Optimization 3 Lasso Zhimin ADMM 2
3 Alternating direction method of multipliers Alternating direction method of multipliers RICE ADMM problem form (with f, g convex) min s.t. f (x) + g(z) Ax + Bz = c - two sets of variables, with separable objective. Augmented Lagrangian Function L ρ (x, z, y) = f (x) + g(z) + y T (Ax + Bz c) + ρ 2 Ax + Bz c 2 2 ADMM: x k+1 z k+1 = arg min x L ρ (x, z k, y k ) = arg min z L ρ (x k+1, z, y k ) y k+1 = y k + ρ(ax k+1 + Bz k+1 c) Zhimin ADMM 3
4 Alternating direction method of multipliers Scaled Dual Form RICE Augmented Lagrangian Function: L ρ (x, z, y) = f (x) + g(z) + y T (Ax + Bz c) + ρ 2 Ax + Bz c 2 2 = f (x) + g(z) + ρ 2 Ax + Bz c + u const with u k = 1 ρ y k scaled form: x k+1 z k+1 = arg min x (f (x) + ρ 2 Ax + Bzk c + u k 2 2) = arg min z (g(z) + ρ 2 Ax k+1 + Bz c + u k 2 2) u k+1 = u k + (Ax k+1 + Bz k+1 c) Zhimin ADMM 4
5 Distributed Optimization Distributed Optimization RICE Consider global consensus problem with regularizer N min f i (x) + g(x) i=1 - x is the global variable ADMM form: min s.t. N f i (x i ) + g(z) i=1 x i z = 0 i = 1, 2,..., N - x i are local variables - z is the global variable, stored locally at each processor Zhimin ADMM 5
6 Distributed Optimization Augmented Lagrangian function L ρ (x, z, y) = g(z) + = g(z) + - with u i = 1 ρ y i - x i, u i are local variables ADMM: x k+1 i z k+1 N (f i (x i ) + yi T (x i z) + ρ 2 x i z 2 2) i=1 N (f i (x i ) + ρ 2 x i z + u i 2 2) + const i=1 = arg min x f i (x i ) + ρ 2 x i z k + u k i 2 2 = arg min g(z) + ρn z 2 z (uk + x k+1 ) 2 2 ui k+1 = ui k + xi k+1 z k+1 - with u k = 1 N N i=1 uk i, x k+1 = 1 N N i=1 x k+1 i Zhimin ADMM 6
7 Lasso RICE Lasso Problem: min 1 2 Ax b λ x 1 ADMM form: min 1 2 Ax b λ z 1 s.t. x z = 0 Scaled ADMM: x k+1 = (A T A + ρi ) 1 (A T b + ρ(z k u k )) z k+1 = S λ/ρ (x k+1 + u k ) u k+1 = u k + x k+1 z k+1 - updating x k+1 is very slow - S λ (x) is the soft-thresholding Q: How to parallelize it? Zhimin ADMM 7
8 ADMM form: min 1 2 Ax b λ z 1 s.t. x z = 0 Divide A and b into several blocks Distributed ADMM form: min s.t. 1 2 N A i x i b i λ z 1 i=1 x i z = 0, for i = 1, 2,..., N Zhimin ADMM 8
9 Distributed ADMM: ui k+1 = ui k + xi k z k, for i = 1, 2,..., N = (A T i A i + ρi ) 1 (A T i b + ρ(z k u k )), for i = 1, 2,..., N x k+1 i z k+1 = S λ/nρ (x k+1 + u k+1 ) - x i, u i are local variable, updated by local systems - z is a global variable In each iteration: - gather xi k+1, ui k+1 and average to get x k+1, u k+1 - scatter the average x k+1, u k+1 to processors - update z k+1 in each processor Zhimin ADMM 9
10 Zhimin ADMM 10
11 Zhimin ADMM 11
12 Zhimin ADMM 12
13 Zhimin ADMM 13
14 Zhimin ADMM 14
15 example Lasso RICE A = randn(1024, 3000) xs(randsample(1024, k)) = randn(k, 1), sparsity k = 200 b = Axs, no noise is introduced run on RICE STIC, with 1, 2, 4, 8, 16, 32 cores result: Termination Rule: xs z xs < 10 3 Zhimin ADMM 15
16 Progress with respect to xs z xs : Zhimin ADMM 16
17 Termination rules: ρ N z z prev < 10 5 nn y & x i z < 10 5 nn max( y, N z ) N is the number processors, n is the dimension of x Zhimin ADMM 17
18 Progress of prime and dual residual norm: Figure: iteration vs z z pre Figure: iteration vs z x Zhimin ADMM 18
19 Progress with respect to xs x xs 10 2 x xs # of iteration Figure: iteration vs xs x xs Zhimin ADMM 19
20 MPI & Cluster Lasso RICE MPI: Cluster: A library for message passing Designed for high performance massive parallel machine and workstation clusters MPI Init, MPI Comm rank, MPI Comm size, MPI Finalize Basic functions: MPI Send, MPI Recv, MPI Wait Collective MPI Communications: MPI Reduce, MPI Allreduce Compile: mpicc -o helloworld helloworld.c Run: mpiexec -n 4./helloworld Login: ssh username@stic.rice.edu -Y Transfer files: scp filename username@stic.rice.edu: Load applications: module load openmpi Zhimin ADMM 20
21 MPI Allreduce Lasso RICE Combines values from all processes and distributes the result back to all processes int MPI Allreduce( void sendbuf, void recvbuf, int count, MPI Datatype datatype, MPI Op op, MPI Comm comm) MPI Op: - MPI MAX : maximum value - MPI MAX : minimum value - MPI SUM: summation - MPI PROD: production - MPI LAND: logical and Other options: Zhimin ADMM 21
22 Amazon Elastic Compute RICE Features: Cloud computing Provide resizeable computing capacity Handle big data Connection: 1. Sign up for EC2. (aws.amazon.com) 2. Launch instance. (11 steps) 3. Connect to Linux Instance. (8 steps) 4. Terminate instance. Zhimin ADMM 22
The Alternating Direction Method of Multipliers
The Alternating Direction Method of Multipliers Customizable software solver package Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein April 27, 2016 1 / 28 Background The Dual Problem Consider
More informationThe Alternating Direction Method of Multipliers
The Alternating Direction Method of Multipliers With Adaptive Step Size Selection Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein October 8, 2015 1 / 30 Introduction Presentation Outline 1 Convex
More informationAlternating Direction Method of Multipliers
Alternating Direction Method of Multipliers CS 584: Big Data Analytics Material adapted from Stephen Boyd (https://web.stanford.edu/~boyd/papers/pdf/admm_slides.pdf) & Ryan Tibshirani (http://stat.cmu.edu/~ryantibs/convexopt/lectures/21-dual-meth.pdf)
More informationECE 574 Cluster Computing Lecture 13
ECE 574 Cluster Computing Lecture 13 Vince Weaver http://web.eece.maine.edu/~vweaver vincent.weaver@maine.edu 21 March 2017 Announcements HW#5 Finally Graded Had right idea, but often result not an *exact*
More informationDistributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers
Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers Stephen Boyd MIIS Xi An, 1/7/12 source: Distributed Optimization and Statistical Learning via the Alternating
More informationParallel and Distributed Sparse Optimization Algorithms
Parallel and Distributed Sparse Optimization Algorithms Part I Ruoyu Li 1 1 Department of Computer Science and Engineering University of Texas at Arlington March 19, 2015 Ruoyu Li (UTA) Parallel and Distributed
More informationParallel and Distributed Sparse Optimization
Parallel and Distributed Sparse Optimization Zhimin Peng Ming Yan Wotao Yin Computational and Applied Mathematics Rice University May 2013 outline 1. sparse optimization background 2. model and data distribution
More informationConvex Optimization: from Real-Time Embedded to Large-Scale Distributed
Convex Optimization: from Real-Time Embedded to Large-Scale Distributed Stephen Boyd Neal Parikh, Eric Chu, Yang Wang, Jacob Mattingley Electrical Engineering Department, Stanford University Springer Lectures,
More informationConvex Optimization and Machine Learning
Convex Optimization and Machine Learning Mengliu Zhao Machine Learning Reading Group School of Computing Science Simon Fraser University March 12, 2014 Mengliu Zhao SFU-MLRG March 12, 2014 1 / 25 Introduction
More informationA parallel implementation of the ADMM algorithm for power network control
Aalto University School of Science Tuomas Rintamäki A parallel implementation of the ADMM algorithm for power network control The document can be stored and made available to the public on the open internet
More informationParallel programming MPI
Parallel programming MPI Distributed memory Each unit has its own memory space If a unit needs data in some other memory space, explicit communication (often through network) is required Point-to-point
More informationIntroduction to the Message Passing Interface (MPI)
Introduction to the Message Passing Interface (MPI) CPS343 Parallel and High Performance Computing Spring 2018 CPS343 (Parallel and HPC) Introduction to the Message Passing Interface (MPI) Spring 2018
More informationApplication of Proximal Algorithms to Three Dimensional Deconvolution Microscopy
Application of Proximal Algorithms to Three Dimensional Deconvolution Microscopy Paroma Varma Stanford University paroma@stanford.edu Abstract In microscopy, shot noise dominates image formation, which
More informationMasterpraktikum - Scientific Computing, High Performance Computing
Masterpraktikum - Scientific Computing, High Performance Computing Message Passing Interface (MPI) and CG-method Michael Bader Alexander Heinecke Technische Universität München, Germany Outline MPI Hello
More informationDistributed Alternating Direction Method of Multipliers
Distributed Alternating Direction Method of Multipliers Ermin Wei and Asuman Ozdaglar Abstract We consider a network of agents that are cooperatively solving a global unconstrained optimization problem,
More informationParallel Programming. Using MPI (Message Passing Interface)
Parallel Programming Using MPI (Message Passing Interface) Message Passing Model Simple implementation of the task/channel model Task Process Channel Message Suitable for a multicomputer Number of processes
More informationAugmented Lagrangian Methods
Augmented Lagrangian Methods Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. IMA, August 2016 Stephen Wright (UW-Madison) Augmented Lagrangian IMA, August 2016 1 /
More informationMessage Passing Interface: Basic Course
Overview of DM- HPC2N, UmeåUniversity, 901 87, Sweden. April 23, 2015 Table of contents Overview of DM- 1 Overview of DM- Parallelism Importance Partitioning Data Distributed Memory Working on Abisko 2
More informationMasterpraktikum - Scientific Computing, High Performance Computing
Masterpraktikum - Scientific Computing, High Performance Computing Message Passing Interface (MPI) Thomas Auckenthaler Wolfgang Eckhardt Technische Universität München, Germany Outline Hello World P2P
More information15-440: Recitation 8
15-440: Recitation 8 School of Computer Science Carnegie Mellon University, Qatar Fall 2013 Date: Oct 31, 2013 I- Intended Learning Outcome (ILO): The ILO of this recitation is: Apply parallel programs
More informationConvex Optimization. Stephen Boyd
Convex Optimization Stephen Boyd Electrical Engineering Computer Science Management Science and Engineering Institute for Computational Mathematics & Engineering Stanford University Institute for Advanced
More informationTensor Sparse PCA and Face Recognition: A Novel Approach
Tensor Sparse PCA and Face Recognition: A Novel Approach Loc Tran Laboratoire CHArt EA4004 EPHE-PSL University, France tran0398@umn.edu Linh Tran Ho Chi Minh University of Technology, Vietnam linhtran.ut@gmail.com
More informationHPC Parallel Programing Multi-node Computation with MPI - I
HPC Parallel Programing Multi-node Computation with MPI - I Parallelization and Optimization Group TATA Consultancy Services, Sahyadri Park Pune, India TCS all rights reserved April 29, 2013 Copyright
More informationTopic Notes: Message Passing Interface (MPI)
Computer Science 400 Parallel Processing Siena College Fall 2008 Topic Notes: Message Passing Interface (MPI) The Message Passing Interface (MPI) was created by a standards committee in the early 1990
More informationSparse Optimization Lecture: Parallel and Distributed Sparse Optimization
Sparse Optimization Lecture: Parallel and Distributed Sparse Optimization Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know basics of parallel computing
More informationConvex optimization algorithms for sparse and low-rank representations
Convex optimization algorithms for sparse and low-rank representations Lieven Vandenberghe, Hsiao-Han Chao (UCLA) ECC 2013 Tutorial Session Sparse and low-rank representation methods in control, estimation,
More informationPractical Introduction to Message-Passing Interface (MPI)
1 Outline of the workshop 2 Practical Introduction to Message-Passing Interface (MPI) Bart Oldeman, Calcul Québec McGill HPC Bart.Oldeman@mcgill.ca Theoretical / practical introduction Parallelizing your
More informationPerformance Estimation of the Alternating Directions Method of Multipliers in a Distributed Environment
Performance Estimation of the Alternating Directions Method of Multipliers in a Distributed Environment Johan Mathe - johmathe@stanford.edu June 3, 20 Goal We will implement the ADMM global consensus algorithm
More informationOptimization for Machine Learning
Optimization for Machine Learning (Problems; Algorithms - C) SUVRIT SRA Massachusetts Institute of Technology PKU Summer School on Data Science (July 2017) Course materials http://suvrit.de/teaching.html
More informationThe Alternating Direction Method of Multipliers
Mid Year Review: AMSC/CMSC 663 and 664 The Alternating Direction Method of Multipliers An Adaptive Step-size Software Library Peter Sutor, Jr. psutor@umd.edu Project Advisor Dr. Tom Goldstein tomg@cs.umd.edu
More informationMPI Collective communication
MPI Collective communication CPS343 Parallel and High Performance Computing Spring 2018 CPS343 (Parallel and HPC) MPI Collective communication Spring 2018 1 / 43 Outline 1 MPI Collective communication
More informationAugmented Lagrangian Methods
Augmented Lagrangian Methods Mário A. T. Figueiredo 1 and Stephen J. Wright 2 1 Instituto de Telecomunicações, Instituto Superior Técnico, Lisboa, Portugal 2 Computer Sciences Department, University of
More informationLecture 4 Duality and Decomposition Techniques
Lecture 4 Duality and Decomposition Techniques Jie Lu (jielu@kth.se) Richard Combes Alexandre Proutiere Automatic Control, KTH September 19, 2013 Consider the primal problem Lagrange Duality Lagrangian
More informationBasic MPI Communications. Basic MPI Communications (cont d)
Basic MPI Communications MPI provides two non-blocking routines: MPI_Isend(buf,cnt,type,dst,tag,comm,reqHandle) buf: source of data to be sent cnt: number of data elements to be sent type: type of each
More informationIntroduction to MPI, the Message Passing Library
Chapter 3, p. 1/57 Basics of Basic Messages -To-? Introduction to, the Message Passing Library School of Engineering Sciences Computations for Large-Scale Problems I Chapter 3, p. 2/57 Outline Basics of
More informationMessage Passing Interface
MPSoC Architectures MPI Alberto Bosio, Associate Professor UM Microelectronic Departement bosio@lirmm.fr Message Passing Interface API for distributed-memory programming parallel code that runs across
More informationDISTRIBUTED NETWORK RESOURCE ALLOCATION WITH INTEGER CONSTRAINTS. Yujiao Cheng, Houfeng Huang, Gang Wu, Qing Ling
DISTRIBUTED NETWORK RESOURCE ALLOCATION WITH INTEGER CONSTRAINTS Yuao Cheng, Houfeng Huang, Gang Wu, Qing Ling Department of Automation, University of Science and Technology of China, Hefei, China ABSTRACT
More informationAgenda. MPI Application Example. Praktikum: Verteiltes Rechnen und Parallelprogrammierung Introduction to MPI. 1) Recap: MPI. 2) 2.
Praktikum: Verteiltes Rechnen und Parallelprogrammierung Introduction to MPI Agenda 1) Recap: MPI 2) 2. Übungszettel 3) Projektpräferenzen? 4) Nächste Woche: 3. Übungszettel, Projektauswahl, Konzepte 5)
More informationOptimization for Machine Learning
with a focus on proximal gradient descent algorithm Department of Computer Science and Engineering Outline 1 History & Trends 2 Proximal Gradient Descent 3 Three Applications A Brief History A. Convex
More informationMessage-Passing and MPI Programming
Message-Passing and MPI Programming 2.1 Transfer Procedures Datatypes and Collectives N.M. Maclaren Computing Service nmm1@cam.ac.uk ext. 34761 July 2010 These are the procedures that actually transfer
More informationPackage ADMM. May 29, 2018
Type Package Package ADMM May 29, 2018 Title Algorithms using Alternating Direction Method of Multipliers Version 0.3.0 Provides algorithms to solve popular optimization problems in statistics such as
More informationInvestigating Image Inpainting via the Alternating Direction Method of Multipliers
Investigating Image Inpainting via the Alternating Direction Method of Multipliers Jonathan Tuck Stanford University jonathantuck@stanford.edu Abstract In many imaging applications, there exists potential
More informationHIGH-dimensional data are commonly observed in various
1 Simplex Representation for Subspace Clustering Jun Xu 1, Student Member, IEEE, Deyu Meng 2, Member, IEEE, Lei Zhang 1, Fellow, IEEE 1 Department of Computing, The Hong Kong Polytechnic University, Hong
More informationProject Proposals. Xiang Zhang. Department of Computer Science Courant Institute of Mathematical Sciences New York University.
Project Proposals Xiang Zhang Department of Computer Science Courant Institute of Mathematical Sciences New York University March 26, 2013 Xiang Zhang (NYU) Project Proposals March 26, 2013 1 / 9 Contents
More informationConic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding
Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding B. O Donoghue E. Chu N. Parikh S. Boyd Convex Optimization and Beyond, Edinburgh, 11/6/2104 1 Outline Cone programming Homogeneous
More informationAdvanced Message-Passing Interface (MPI)
Outline of the workshop 2 Advanced Message-Passing Interface (MPI) Bart Oldeman, Calcul Québec McGill HPC Bart.Oldeman@mcgill.ca Morning: Advanced MPI Revision More on Collectives More on Point-to-Point
More informationSnapVX: A Network-Based Convex Optimization Solver
Journal of Machine Learning Research 18 (2017) 1-5 Submitted 9/15; Revised 10/16; Published 2/17 SnapVX: A Network-Based Convex Optimization Solver David Hallac Christopher Wong Steven Diamond Abhijit
More informationCollective Communications II
Collective Communications II Ned Nedialkov McMaster University Canada SE/CS 4F03 January 2014 Outline Scatter Example: parallel A b Distributing a matrix Gather Serial A b Parallel A b Allocating memory
More informationCS 179: GPU Programming. Lecture 14: Inter-process Communication
CS 179: GPU Programming Lecture 14: Inter-process Communication The Problem What if we want to use GPUs across a distributed system? GPU cluster, CSIRO Distributed System A collection of computers Each
More informationL15: Putting it together: N-body (Ch. 6)!
Outline L15: Putting it together: N-body (Ch. 6)! October 30, 2012! Review MPI Communication - Blocking - Non-Blocking - One-Sided - Point-to-Point vs. Collective Chapter 6 shows two algorithms (N-body
More informationIntroduction to MPI. May 20, Daniel J. Bodony Department of Aerospace Engineering University of Illinois at Urbana-Champaign
Introduction to MPI May 20, 2013 Daniel J. Bodony Department of Aerospace Engineering University of Illinois at Urbana-Champaign Top500.org PERFORMANCE DEVELOPMENT 1 Eflop/s 162 Pflop/s PROJECTED 100 Pflop/s
More informationIntroduction to MPI. Ricardo Fonseca. https://sites.google.com/view/rafonseca2017/
Introduction to MPI Ricardo Fonseca https://sites.google.com/view/rafonseca2017/ Outline Distributed Memory Programming (MPI) Message Passing Model Initializing and terminating programs Point to point
More informationMPI Tutorial. Shao-Ching Huang. High Performance Computing Group UCLA Institute for Digital Research and Education
MPI Tutorial Shao-Ching Huang High Performance Computing Group UCLA Institute for Digital Research and Education Center for Vision, Cognition, Learning and Art, UCLA July 15 22, 2013 A few words before
More informationA More Efficient Approach to Large Scale Matrix Completion Problems
A More Efficient Approach to Large Scale Matrix Completion Problems Matthew Olson August 25, 2014 Abstract This paper investigates a scalable optimization procedure to the low-rank matrix completion problem
More informationDistributed Memory Systems: Part IV
Chapter 5 Distributed Memory Systems: Part IV Max Planck Institute Magdeburg Jens Saak, Scientific Computing II 293/342 The Message Passing Interface is a standard for creation of parallel programs using
More informationEpigraph proximal algorithms for general convex programming
Epigraph proimal algorithms for general conve programming Matt Wytock, Po-Wei Wang and J. Zico Kolter Machine Learning Department Carnegie Mellon University mwytock@cs.cmu.edu Abstract This work aims at
More informationPOGS Proximal Operator Graph Solver
POGS Proximal Operator Graph Solver Chris Fougner June 4, 2014 1 Introduction As data sets continue to increase in size, it becomes increasingly important to design algorithms that scale with the size
More informationThe MPI Message-passing Standard Practical use and implementation (V) SPD Course 6/03/2017 Massimo Coppola
The MPI Message-passing Standard Practical use and implementation (V) SPD Course 6/03/2017 Massimo Coppola Intracommunicators COLLECTIVE COMMUNICATIONS SPD - MPI Standard Use and Implementation (5) 2 Collectives
More informationA primal-dual framework for mixtures of regularizers
A primal-dual framework for mixtures of regularizers Baran Gözcü baran.goezcue@epfl.ch Laboratory for Information and Inference Systems (LIONS) École Polytechnique Fédérale de Lausanne (EPFL) Switzerland
More informationMessage Passing Interface
Message Passing Interface DPHPC15 TA: Salvatore Di Girolamo DSM (Distributed Shared Memory) Message Passing MPI (Message Passing Interface) A message passing specification implemented
More informationDepartment of Electrical and Computer Engineering
LAGRANGIAN RELAXATION FOR GATE IMPLEMENTATION SELECTION Yi-Le Huang, Jiang Hu and Weiping Shi Department of Electrical and Computer Engineering Texas A&M University OUTLINE Introduction and motivation
More informationDistributed Machine Learning: An Intro. Chen Huang
: An Intro. Chen Huang Feature Engineering Group, Data Mining Lab, Big Data Research Center, UESTC Contents Background Some Examples Model Parallelism & Data Parallelism Parallelization Mechanisms Synchronous
More informationIntroduction to parallel computing with MPI
Introduction to parallel computing with MPI Sergiy Bubin Department of Physics Nazarbayev University Distributed Memory Environment image credit: LLNL Hybrid Memory Environment Most modern clusters and
More informationIntroduction to Parallel Programming
Introduction to Parallel Programming Ste phane Zuckerman Laboratoire ETIS Universite Paris-Seine, Universite de Cergy-Pontoise, ENSEA, CNRS F95000, Cergy, France October 19, 2018 S.Zuckerman (ETIS) Parallel
More informationCS 470 Spring Mike Lam, Professor. Distributed Programming & MPI
CS 470 Spring 2017 Mike Lam, Professor Distributed Programming & MPI MPI paradigm Single program, multiple data (SPMD) One program, multiple processes (ranks) Processes communicate via messages An MPI
More informationNetwork Lasso: Clustering and Optimization in Large Graphs
Network Lasso: Clustering and Optimization in Large Graphs David Hallac, Jure Leskovec, Stephen Boyd Stanford University September 28, 2015 Convex optimization Convex optimization is everywhere Introduction
More informationPraktikum: Verteiltes Rechnen und Parallelprogrammierung Introduction to MPI
Praktikum: Verteiltes Rechnen und Parallelprogrammierung Introduction to MPI Agenda 1) MPI für Java Installation OK? 2) 2. Übungszettel Grundidee klar? 3) Projektpräferenzen? 4) Nächste Woche: 3. Übungszettel,
More informationOur new HPC-Cluster An overview
Our new HPC-Cluster An overview Christian Hagen Universität Regensburg Regensburg, 15.05.2009 Outline 1 Layout 2 Hardware 3 Software 4 Getting an account 5 Compiling 6 Queueing system 7 Parallelization
More informationHigh Performance Computing
High Performance Computing Course Notes 2009-2010 2010 Message Passing Programming II 1 Communications Point-to-point communications: involving exact two processes, one sender and one receiver For example,
More informationDetecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference
Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Minh Dao 1, Xiang Xiang 1, Bulent Ayhan 2, Chiman Kwan 2, Trac D. Tran 1 Johns Hopkins Univeristy, 3400
More informationL19: Putting it together: N-body (Ch. 6)!
Administrative L19: Putting it together: N-body (Ch. 6)! November 22, 2011! Project sign off due today, about a third of you are done (will accept it tomorrow, otherwise 5% loss on project grade) Next
More informationSCREEN CONTENT IMAGE SEGMENTATION USING LEAST ABSOLUTE DEVIATION FITTING. Shervin Minaee and Yao Wang
SCREEN CONTENT IMAGE SEGMENTATION USING LEAST ABSOLUTE DEVIATION FITTING Shervin Minaee and Yao Wang Department of Electrical and Computer Engineering, Polytechnic School of Engineering, New York University,
More informationCollective Communications I
Collective Communications I Ned Nedialkov McMaster University Canada CS/SE 4F03 January 2016 Outline Introduction Broadcast Reduce c 2013 16 Ned Nedialkov 2/14 Introduction A collective communication involves
More informationMA471. Lecture 5. Collective MPI Communication
MA471 Lecture 5 Collective MPI Communication Today: When all the processes want to send, receive or both Excellent website for MPI command syntax available at: http://www-unix.mcs.anl.gov/mpi/www/ 9/10/2003
More informationThe flare Package for High Dimensional Linear Regression and Precision Matrix Estimation in R
Journal of Machine Learning Research 6 (205) 553-557 Submitted /2; Revised 3/4; Published 3/5 The flare Package for High Dimensional Linear Regression and Precision Matrix Estimation in R Xingguo Li Department
More informationCOLLABORATIVE RESOURCE ALLOCATION OVER A HYBRID CLOUD CENTER AND EDGE SERVER NETWORK *
Journal of Computational Mathematics Vol.35, No.4, 2017, 421 436. http://www.global-sci.org/jcm doi:10.4208/jcm.1608-m2016-0561 COLLABORATIVE RESOURCE ALLOCATION OVER A HYBRID CLOUD CENTER AND EDGE SERVER
More informationCS 224W Final Report: Community Detection for Distributed Optimization
CS 224W Final Report: Community Detection for Distributed Optimization Tri Dao trid@stanford.edu Rolland He rhe@stanford.edu Zhivko Zhechev zzhechev@stanford.edu 1 Introduction Distributed optimization
More informationDistributed Memory Programming with MPI
Distributed Memory Programming with MPI Moreno Marzolla Dip. di Informatica Scienza e Ingegneria (DISI) Università di Bologna moreno.marzolla@unibo.it Algoritmi Avanzati--modulo 2 2 Credits Peter Pacheco,
More informationCS 470 Spring Mike Lam, Professor. Distributed Programming & MPI
CS 470 Spring 2018 Mike Lam, Professor Distributed Programming & MPI MPI paradigm Single program, multiple data (SPMD) One program, multiple processes (ranks) Processes communicate via messages An MPI
More informationMessage Passing Interface. most of the slides taken from Hanjun Kim
Message Passing Interface most of the slides taken from Hanjun Kim Message Passing Pros Scalable, Flexible Cons Someone says it s more difficult than DSM MPI (Message Passing Interface) A standard message
More informationA Decentralized Energy Management System
A Decentralized Energy Management System Ceyhun Eksin, Ali Hooshmand and Ratnesh Sharma Energy Management Department at NEC Laboratories America, Inc., Cupertino, CA, 9514 USA. Email:{ceksin, ahooshmand,
More informationIteratively Re-weighted Least Squares for Sums of Convex Functions
Iteratively Re-weighted Least Squares for Sums of Convex Functions James Burke University of Washington Jiashan Wang LinkedIn Frank Curtis Lehigh University Hao Wang Shanghai Tech University Daiwei He
More informationMessage Passing Interface - MPI
Message Passing Interface - MPI Parallel and Distributed Computing Department of Computer Science and Engineering (DEI) Instituto Superior Técnico October 24, 2011 Many slides adapted from lectures by
More informationProgramming with MPI Collectives
Programming with MPI Collectives Jan Thorbecke Type to enter text Delft University of Technology Challenge the future Collectives Classes Communication types exercise: BroadcastBarrier Gather Scatter exercise:
More informationMinimum-Polytope-Based Linear Programming Decoder for LDPC Codes via ADMM Approach
Minimum-Polytope-Based Linear Programg Decoder for LDPC Codes via ADMM Approach Jing Bai, Yongchao Wang, Member, IEEE, Francis C. M. Lau, Senior Member, IEEE arxiv:90.07806v [cs.it] 23 Jan 209 Abstract
More informationRobust Principal Component Analysis (RPCA)
Robust Principal Component Analysis (RPCA) & Matrix decomposition: into low-rank and sparse components Zhenfang Hu 2010.4.1 reference [1] Chandrasekharan, V., Sanghavi, S., Parillo, P., Wilsky, A.: Ranksparsity
More informationMPI Performance Snapshot
User's Guide 2014-2015 Intel Corporation Legal Information No license (express or implied, by estoppel or otherwise) to any intellectual property rights is granted by this document. Intel disclaims all
More informationCollective Communication in MPI and Advanced Features
Collective Communication in MPI and Advanced Features Pacheco s book. Chapter 3 T. Yang, CS240A. Part of slides from the text book, CS267 K. Yelick from UC Berkeley and B. Gropp, ANL Outline Collective
More informationCollective Communication
Lab 14 Collective Communication Lab Objective: Learn how to use collective communication to increase the efficiency of parallel programs In the lab on the Trapezoidal Rule [Lab??], we worked to increase
More informationMPI. (message passing, MIMD)
MPI (message passing, MIMD) What is MPI? a message-passing library specification extension of C/C++ (and Fortran) message passing for distributed memory parallel programming Features of MPI Point-to-point
More informationLINE DETECTION IN SPECKLE IMAGES USING RADON TRANSFORM AND l 1 REGULARIZATION
LINE DETECTION IN SPECKLE IMAGES USING RADON TRANSFORM AND l 1 REGULARIZATION Nantheera Anantrasirichai 1, Marco Allinovi 2, Wesley Hayes 3, David Bull 1, Alin Achim 1 1 Visual Information Lab, University
More informationAn R Package flare for High Dimensional Linear Regression and Precision Matrix Estimation
An R Package flare for High Dimensional Linear Regression and Precision Matrix Estimation Xingguo Li Tuo Zhao Xiaoming Yuan Han Liu Abstract This paper describes an R package named flare, which implements
More informationLecture 21 May 21, 2018
Stats 300C: Theory of Statistics Spring 208 Lecture 2 May 2, 208 Prof. Emmanuel Candes Scrie: Sergio Camelo, Govinda Kamath, edited y Elena Tuzhilina Agenda. The Lasso 2. The Lasso dual 3. SURE for Lasso
More informationEfficient Iterative LP Decoding of LDPC Codes with Alternating Direction Method of Multipliers
Efficient Iterative LP Decoding of LDPC Codes with Alternating Direction Method of Multipliers Xiaojie Zhang Samsung R&D America, Dallas, Texas 758 Email: eric.zhang@samsung.com Paul H. Siegel University
More informationFirst day. Basics of parallel programming. RIKEN CCS HPC Summer School Hiroya Matsuba, RIKEN CCS
First day Basics of parallel programming RIKEN CCS HPC Summer School Hiroya Matsuba, RIKEN CCS Today s schedule: Basics of parallel programming 7/22 AM: Lecture Goals Understand the design of typical parallel
More informationUnion of Learned Sparsifying Transforms Based Low-Dose 3D CT Image Reconstruction
Union of Learned Sparsifying Transforms Based Low-Dose 3D CT Image Reconstruction Xuehang Zheng 1, Saiprasad Ravishankar 2, Yong Long 1, Jeff Fessler 2 1 University of Michigan - Shanghai Jiao Tong University
More informationAn introduction to MPI
An introduction to MPI C MPI is a Library for Message-Passing Not built in to compiler Function calls that can be made from any compiler, many languages Just link to it Wrappers: mpicc, mpif77 Fortran
More informationWireless Networks Research Seminar April 22nd 2013
Wireless Networks Research Seminar April 22nd 2013 Distributed Transmit Power Minimization in Wireless Sensor Networks via Cross-Layer Optimization NETS2020 Markus Leinonen, Juha Karjalainen, Marian Codreanu,
More informationWelcome to the introductory workshop in MPI programming at UNICC
Welcome...... to the introductory workshop in MPI programming at UNICC Schedule: 08.00-12.00 Hard work and a short coffee break Scope of the workshop: We will go through the basics of MPI-programming and
More informationLecture 19: November 5
0-725/36-725: Convex Optimization Fall 205 Lecturer: Ryan Tibshirani Lecture 9: November 5 Scribes: Hyun Ah Song Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have not
More information