Part II (c) Desktop Installation. Net Serpents LLC, USA
|
|
- Miles Parrish
- 6 years ago
- Views:
Transcription
1 Part II (c) Desktop ation
2 Desktop ation ation Supported Platforms Required Software Releases &Mirror Sites Configure Format Start/ Stop Verify
3 Supported Platforms ation GNU Linux supported for Development Production Demonstrated on 2000 node cluster Win32 Development only Not supported as a production platform
4 Required Software Required Software Following to be installed first Java 1.6.x or higher ssh: Ubuntu: ssh and rsync Windows: openssh
5 Releases and Mirror Sites Releases and Mirror Sites Releases Stable Release (July 2015) Stable release: (released Nov 2014) Earlier good releases: (April 2014) (GA Release Oct 2013) (Dec 2011) Visit:
6 Mirror Sites Mirror Sites Downloads available at several mirror sites Suggested by Apache: common Other mirror sites at: common
7 - Overview Hadoop on Ubuntu Step 1- Java Step 2 Create a dedicated hadoop user Step 3 ssh Step 4 Create ssh certificates Ubuntu is the most popular Linux distribution Step 5 Hadoop Step 6 Setup configuration Files Step 7 Format Step 8 Start/ Stop Step 9 - Verify
8 Step 1 - Java Login as an admin user: $ cd ~ # Update the source list $ sudo apt-get update $ sudo apt-get install default-jdk # Verify version of Java is or higher $ java -version java version "1.7.0_65" OpenJDK Runtime Environment (IcedTea 2.5.3) (7u ubuntu ) OpenJDK 64-Bit Server VM (build b04, mixed mode)
9 - Overview Hadoop on Ubuntu Step 1- Java Step 2 Create a dedicated hadoop user Step 3 ssh Step 4 Create ssh certificates Step 5 Hadoop Step 6 Setup configuration Files Step 7 Format Step 8 Start/ Stop Step 9 - Verify
10 Step 2 Create a dedicated Hadoop user # Create a hadoop group $ sudo addgroup hadoop Adding group `hadoop' (GID 1009) Done. # Create hadoop user $ sudo adduser --ingroup hadoop huser Adding user `huser'... Adding new user `huser' (1001) with group `hadoop'... Creating home directory `/home/huser'... Copying files from `/etc/skel'... Enter new UNIX password: Retype new UNIX password: passwd: password updated successfully Changing the user information for huser
11 Step 2 Create a dedicated Hadoop user Enter the new value, or press ENTER for the default Full Name []: Room Number []: Work Phone []: Home Phone []: Other []: Is the information correct? [Y/n] Y # Add new user to sudoers $ sudo adduser huser sudo [sudo] password for admin: Adding user `huser' to group `sudo'... Adding user huser to group sudo Done.
12 - Overview Hadoop on Ubuntu Step 1- Java Step 2 Create a dedicated hadoop user Step 3 ssh Step 4 Create ssh certificates Step 5 Hadoop Step 6 Setup configuration Files Step 7 Format Step 8 Start/ Stop Step 9 - Verify
13 Step 3 SSH $ sudo apt-get install ssh # Verify SSH is installed $ which ssh /usr/bin/ssh $ which sshd /usr/sbin/sshd
14 Step 4 Create SSH Certificates $ sudo su huser #Generate a key pair $ ssh-keygen -f ~/.ssh/id_rsa -t rsa -P "" Generating public/private rsa key pair. Enter file in which to save the key (/home/huser/.ssh/id_rsa): Created directory '/home/huser/.ssh'. Your identification has been saved in /home/huser/.ssh/id_rsa. Your public key has been saved in /home/huser/.ssh/id_rsa.pub. The key fingerprint is: 20:6c:f3:ff:0f:33:bf:30:72:c3:22:70:24:cc:2d:d3 huser@laptop The key's randomart image is: +--[ RSA 2048]----+.oo.o
15 Step 4 Create SSH Certificates # Create list of authorized keys to avoid being prompted for password $ cat $HOME/.ssh/id_rsa.pub >> $HOME/.ssh/authorized_keys
16 - Overview Hadoop on Ubuntu Step 1- Java Step 2 Create a dedicated hadoop user Step 3 ssh Step 4 Create ssh certificates Step 5 Hadoop Step 6 Setup configuration Files Step 7 Format Step 8 Start/ Stop Step 9 - Verify
17 Step 5 Hadoop # Download distribution from a mirror site $ sudo wget # Extract the files $ tar xvzf hadoop tar.gz $ cd hadoop # Move files to /usr/local/hadoop $ sudo mkdir /usr/local/hadoop $ sudo mv * /usr/local/hadoop # Change ownership to hadoop user $ sudo chown -R huser:hadoop /usr/local/hadoop
18 - Overview Hadoop on Ubuntu Step 1- Java Step 2 Create a dedicated hadoop user Step 3 ssh Step 4 Create ssh certificates Step 5 Hadoop Step 6 Setup configuration Files Step 7 Format Step 8 - Verify
19 Step 6 Configure # Update links to point to Java $update-alternatives --config java There is only one alternative in link group java (providing /usr/bin/java): /usr/lib/jvm/ java-7-openjdk-amd64/jre/bin/java Nothing to configure. # Note down JAVA_HOME variable value $ which javac /usr/bin/javac $ readlink -f /usr/bin/javac /usr/lib/jvm/java-7-openjdk-amd64/bin/javac (Note: JAVA_HOME would be everything before /bin/javac)
20 Step 6 Configure # Add variables to the end of.bashrc $ vi ~/.bashrc export JAVA_HOME=/usr/lib/jvm/java-7-openjdk-amd64 export HADOOP_INSTALL=/usr/local/hadoop export PATH=$PATH:$HADOOP_INSTALL/bin export PATH=$PATH:$HADOOP_INSTALL/sbin export HADOOP_MAPRED_HOME=$HADOOP_INSTALL export HADOOP_COMMON_HOME=$HADOOP_INSTALL export HADOOP_HDFS_HOME=$HADOOP_INSTALL
21 Step 6 Configure export YARN_HOME=$HADOOP_INSTALL export HADOOP_COMMON_LIB_NATIVE_DIR= $HADOOP_INSTALL/lib/native export HADOOP_OPTS="-Djava.library.path= $HADOOP_INSTALL/lib # Execute the commands in.bashrc $ source ~/.bashrc
22 Step 6 Configure # Configure hadoop-env.sh Change variable JAVA_HOME in hadoop-env.sh $ vi /usr/local/hadoop/etc/hadoop/hadoop-env.sh export JAVA_HOME=/usr/lib/jvm/java-7-openjdk-amd64
23 Step 6 Configure # Configure core-site.xml # First create a tmp folder for hadoop $ sudo mkdir -p /app/hadoop/tmp $ sudo chown huser:hadoop /app/hadoop/tmp # Modify core-site.xml $ vi /usr/local/hadoop/etc/hadoop/core-site.xml
24 Step 6 Configure Modify as follows: <configuration> <property> <name>hadoop.tmp.dir</name> <value>/app/hadoop/tmp</value> <description>a base for other temporary directories.</description> </property> <property> <name>fs.default.name</name> <value>hdfs://localhost:54310</value> <description>the name of the default file system. A URI whose scheme and authority determine the FileSystem implementation. The uri's scheme determines the config property (fs.scheme.impl) naming the FileSystem implementation class. The uri's authority is used to determine the host, port, etc. for a filesystem.</description> </property> </configuration>
25 Step 6 Configure # Configure mapred-site.xml # First copy the file from the template provided $ cp /usr/local/hadoop/etc/hadoop/mapredsite.xml.template /usr/local/hadoop/etc/hadoop/ mapred-site.xml # Modify mapred-site.xml $ vi /usr/local/hadoop/etc/hadoop/mapred-site.xml
26 Step 6 Configure Modify as follows: <configuration> <property> <name>mapred.job.tracker</name> <value>localhost:54311</value> <description>the host and port that the MapReduce job tracker runs at. </description> </property> </configuration>
27 Step 6 Configure # Configure hdfs-site.xml # First create the directories for data node and name node $ sudo mkdir -p /usr/local/hadoop_store/hdfs/namenode $ sudo mkdir -p /usr/local/hadoop_store/hdfs/datanode $ sudo chown -R huser:hadoop /usr/local/hadoop_store # Modify hdfs-site.xml $ vi /usr/local/hadoop/etc/hadoop/mapred-site.xml
28 Step 6 Configure Modify as follows: <configuration> <property> <name>dfs.replication</name> <value>1</value> <description>this is a default value for block replication. Tis could be different from the value specified when the file is created. This value is just a default if none is specified at file creation. </description> </property> <property> continued see next page
29 Step 6 Configure <property> <name>dfs.namenode.name.dir</name> <value>file:/usr/local/hadoop_store/hdfs/namenode</value> </property> <property> <name>dfs.datanode.data.dir</name> <value>file:/usr/local/hadoop_store/hdfs/datanode</value> </property> </configuration>
30 - Overview Hadoop on Ubuntu Step 1- Java Step 2 Create a dedicated hadoop user Step 3 ssh Step 4 Create ssh certificates Step 5 Hadoop Step 6 Setup configuration Files Step 7 Format Step 8 Start/ Stop Step 9 - Verify
31 Step 7 Format Format $ hadoop namenode -format DEPRECATED: Use of this script to execute hdfs command is deprecated. Instead use the hdfs command for it. 15/04/18 14:43:03 INFO namenode.namenode: STARTUP_MSG: /************************************************************ STARTUP_MSG: Starting NameNode STARTUP_MSG: host = laptop/ STARTUP_MSG: args = [-format] STARTUP_MSG: version = STARTUP_MSG: classpath = /usr/local/hadoop/etc/hadoop... STARTUP_MSG: java = 1.7.0_65 ************************************************************/ 15/04/18 14:43:03 INFO namenode.namenode: registered UNIX signal handlers for [TERM, HUP, INT] 15/04/18 14:43:03 INFO namenode.namenode: createnamenode [-format] 15/04/18 14:43:07 WARN util.nativecode
32 - Overview Hadoop on Ubuntu Step 1- Java Step 2 Create a dedicated hadoop user Step 3 ssh Step 4 Create ssh certificates Ubuntu is the most popular Linux distribution Step 5 Hadoop Step 6 Setup configuration Files Step 7 Format Step 8 Start/ Stop Step 9 - Verify
33 Step 8 Start/ Stop Start / Stop To view the available commands: $ ls /usr/local/hadoop/sbin Start hadoop (Login as hadoop user) Available commands: $start-all.sh (deprecated) this starts all daemons OR $start-dfs.sh $start-yarn.sh $mr-jobhistory-daemon.sh start historyserver -
34 Step 8 Start/ Stop Start / Stop To stop hadoop (Login as hadoop user): $stop-all.sh Or $ Stop-dfs.sh $ stop-yarn.sh $ mr-jobhistory-daemon.sh stop historyserver
35 - Overview Hadoop on Ubuntu Step 1- Java Step 2 Create a dedicated hadoop user Step 3 ssh Step 4 Create ssh certificates Ubuntu is the most popular Linux distribution Step 5 Hadoop Step 6 Setup configuration Files Step 7 Format Step 8 Start/ Stop Step 9 - Verify
36 Step 9 Verify Verify To verify: $ jps (to view running daemons)
37 Step 9 Verify Verify WebUI: Using port Eg.,
38 Quiz Quiz 1 - Which two of the following must be installed for hadoop installation to succeed a- ssh b- java c- ftp d- ruby 2 Which of the following Linux commands may be used to install a software like ssh or java a- get b- put c- get-apt d- startall.sh 3- Name the commands you would use to start the hadoop daemons and to stop them
39 Quiz Quiz 4 - Which of the following is NOT a configuration file for hadoop a- hadoop-env.sh b- core-site.xml c- mapred-site.xml d- hadoop-configuration.xml 5 True or false a- The user owning hadoop files must belong to a group called hadoop b- A hadoop cluster running on a single node is called pseudo-distributed mode c- A fully distributed hadoop configuration should have a minimum of three nodes
3 Hadoop Installation: Pseudo-distributed mode
Laboratory 3 Hadoop Installation: Pseudo-distributed mode Obecjective Hadoop can be run in 3 different modes. Different modes of Hadoop are 1. Standalone Mode Default mode of Hadoop HDFS is not utilized
More informationInstallation of Hadoop on Ubuntu
Installation of Hadoop on Ubuntu Various software and settings are required for Hadoop. This section is mainly developed based on rsqrl.com tutorial. 1- Install Java Software Java Version* Openjdk version
More informationApache Hadoop Installation and Single Node Cluster Configuration on Ubuntu A guide to install and setup Single-Node Apache Hadoop 2.
SDJ INFOSOFT PVT. LTD Apache Hadoop 2.6.0 Installation and Single Node Cluster Configuration on Ubuntu A guide to install and setup Single-Node Apache Hadoop 2.x Table of Contents Topic Software Requirements
More informationInstallation and Configuration Documentation
Installation and Configuration Documentation Release 1.0.1 Oshin Prem Sep 27, 2017 Contents 1 HADOOP INSTALLATION 3 1.1 SINGLE-NODE INSTALLATION................................... 3 1.2 MULTI-NODE INSTALLATION....................................
More informationHadoop Quickstart. Table of contents
Table of contents 1 Purpose...2 2 Pre-requisites...2 2.1 Supported Platforms... 2 2.2 Required Software... 2 2.3 Installing Software...2 3 Download...2 4 Prepare to Start the Hadoop Cluster...3 5 Standalone
More informationHadoop Setup on OpenStack Windows Azure Guide
CSCI4180 Tutorial- 2 Hadoop Setup on OpenStack Windows Azure Guide ZHANG, Mi mzhang@cse.cuhk.edu.hk Sep. 24, 2015 Outline Hadoop setup on OpenStack Ø Set up Hadoop cluster Ø Manage Hadoop cluster Ø WordCount
More informationInstalling Hadoop. You need a *nix system (Linux, Mac OS X, ) with a working installation of Java 1.7, either OpenJDK or the Oracle JDK. See, e.g.
Big Data Computing Instructor: Prof. Irene Finocchi Master's Degree in Computer Science Academic Year 2013-2014, spring semester Installing Hadoop Emanuele Fusco (fusco@di.uniroma1.it) Prerequisites You
More informationUNIT II HADOOP FRAMEWORK
UNIT II HADOOP FRAMEWORK Hadoop Hadoop is an Apache open source framework written in java that allows distributed processing of large datasets across clusters of computers using simple programming models.
More informationInria, Rennes Bretagne Atlantique Research Center
Hadoop TP 1 Shadi Ibrahim Inria, Rennes Bretagne Atlantique Research Center Getting started with Hadoop Prerequisites Basic Configuration Starting Hadoop Verifying cluster operation Hadoop INRIA S.IBRAHIM
More informationAbout the Tutorial. Audience. Prerequisites. Copyright & Disclaimer. HCatalog
About the Tutorial HCatalog is a table storage management tool for Hadoop that exposes the tabular data of Hive metastore to other Hadoop applications. It enables users with different data processing tools
More informationMulti-Node Cluster Setup on Hadoop. Tushar B. Kute,
Multi-Node Cluster Setup on Hadoop Tushar B. Kute, http://tusharkute.com What is Multi-node? Multi-node cluster Multinode Hadoop cluster as composed of Master- Slave Architecture to accomplishment of BigData
More informationInstalling Hadoop / Yarn, Hive 2.1.0, Scala , and Spark 2.0 on Raspberry Pi Cluster of 3 Nodes. By: Nicholas Propes 2016
Installing Hadoop 2.7.3 / Yarn, Hive 2.1.0, Scala 2.11.8, and Spark 2.0 on Raspberry Pi Cluster of 3 Nodes By: Nicholas Propes 2016 1 NOTES Please follow instructions PARTS in order because the results
More informationDeveloper s Manual. Version May, Computer Science Department, Texas Christian University
Developer s Manual Version 4.0 2 May, 2016 2015-2016 Computer Science Department, Texas Christian University Revision Signatures By signing the following document, the team member is acknowledging that
More informationBig Data Retrieving Required Information From Text Files Desmond Hill Yenumula B Reddy (Advisor)
Big Data Retrieving Required Information From Text Files Desmond Hill Yenumula B Reddy (Advisor) 1 OUTLINE Objective What is Big data Characteristics of Big Data Setup Requirements Hadoop Setup Word Count
More informationImplementation of Randomized Hydrodynamic Load Balancing Algorithm using Map Reduce framework on Open Source Platform of Hadoop
ISSN: 2393-8528 Contents lists available at www.ijicse.in International Journal of Innovative Computer Science & Engineering Volume 2 Issue 3; July-August-2015; Page No. 44-51 Implementation of Randomized
More informationBIG DATA TRAINING PRESENTATION
BIG DATA TRAINING PRESENTATION TOPICS TO BE COVERED HADOOP YARN MAP REDUCE SPARK FLUME SQOOP OOZIE AMBARI TOPICS TO BE COVERED FALCON RANGER KNOX SENTRY MASTER IMAGE INSTALLATION 1 JAVA INSTALLATION: 1.
More informationOutline Introduction Big Data Sources of Big Data Tools HDFS Installation Configuration Starting & Stopping Map Reduc.
D. Praveen Kumar Junior Research Fellow Department of Computer Science & Engineering Indian Institute of Technology (Indian School of Mines) Dhanbad, Jharkhand, India Head of IT & ITES, Skill Subsist Impels
More informationHadoop is essentially an operating system for distributed processing. Its primary subsystems are HDFS and MapReduce (and Yarn).
1 Hadoop Primer Hadoop is essentially an operating system for distributed processing. Its primary subsystems are HDFS and MapReduce (and Yarn). 2 Passwordless SSH Before setting up Hadoop, setup passwordless
More informationHadoop Setup Walkthrough
Hadoop 2.7.3 Setup Walkthrough This document provides information about working with Hadoop 2.7.3. 1 Setting Up Configuration Files... 2 2 Setting Up The Environment... 2 3 Additional Notes... 3 4 Selecting
More informationCloud Computing II. Exercises
Cloud Computing II Exercises Exercise 1 Creating a Private Cloud Overview In this exercise, you will install and configure a private cloud using OpenStack. This will be accomplished using a singlenode
More informationGetting Started with Hadoop/YARN
Getting Started with Hadoop/YARN Michael Völske 1 April 28, 2016 1 michael.voelske@uni-weimar.de Michael Völske Getting Started with Hadoop/YARN April 28, 2016 1 / 66 Outline Part One: Hadoop, HDFS, and
More informationBig Data Analytics by Using Hadoop
Governors State University OPUS Open Portal to University Scholarship All Capstone Projects Student Capstone Projects Spring 2015 Big Data Analytics by Using Hadoop Chaitanya Arava Governors State University
More informationGuidelines - Configuring PDI, MapReduce, and MapR
Guidelines - Configuring PDI, MapReduce, and MapR This page intentionally left blank. Contents Overview... 1 Set Up Your Environment... 2 Get MapR Server Information... 2 Set Up Your Host Environment...
More informationHadoop Lab 2 Exploring the Hadoop Environment
Programming for Big Data Hadoop Lab 2 Exploring the Hadoop Environment Video A short video guide for some of what is covered in this lab. Link for this video is on my module webpage 1 Open a Terminal window
More informationInstallation Guide. Community release
Installation Guide Community 151 release This document details step-by-step deployment procedures, system and environment requirements to assist Jumbune deployment 1 P a g e Table of Contents Introduction
More informationAbout 1. Chapter 1: Getting started with oozie 2. Remarks 2. Versions 2. Examples 2. Installation or Setup 2. Chapter 2: Oozie
oozie #oozie Table of Contents About 1 Chapter 1: Getting started with oozie 2 Remarks 2 Versions 2 Examples 2 Installation or Setup 2 Chapter 2: Oozie 101 7 Examples 7 Oozie Architecture 7 Oozie Application
More informationWelcome to getting started with Ubuntu Server. This System Administrator Manual. guide to be simple to follow, with step by step instructions
Welcome to getting started with Ubuntu 12.04 Server. This System Administrator Manual guide to be simple to follow, with step by step instructions with screenshots INDEX 1.Installation of Ubuntu 12.04
More informationThis brief tutorial provides a quick introduction to Big Data, MapReduce algorithm, and Hadoop Distributed File System.
About this tutorial Hadoop is an open-source framework that allows to store and process big data in a distributed environment across clusters of computers using simple programming models. It is designed
More informationLinux Essentials Objectives Topics:
Linux Essentials Linux Essentials is a professional development certificate program that covers basic knowledge for those working and studying Open Source and various distributions of Linux. Exam Objectives
More informationHadoop Cluster Implementation
Hadoop Cluster Implementation By Aysha Binta Sayed ID:2013-1-60-068 Supervised By Dr. Md. Shamim Akhter Assistant Professor Department of Computer Science and Engineering East West University A project
More informationRDMA for Apache Hadoop 2.x User Guide
1.3.0 User Guide HIGH-PERFORMANCE BIG DATA TEAM http://hibd.cse.ohio-state.edu NETWORK-BASED COMPUTING LABORATORY DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING THE OHIO STATE UNIVERSITY Copyright (c)
More informationHortonworks Technical Preview for Apache Falcon
Architecting the Future of Big Data Hortonworks Technical Preview for Apache Falcon Released: 11/20/2013 Architecting the Future of Big Data 2013 Hortonworks Inc. All Rights Reserved. Welcome to Hortonworks
More informationLinux Kung-Fu. James Droste UBNetDef Fall 2016
Linux Kung-Fu James Droste UBNetDef Fall 2016 $ init 1 GO TO https://apps.ubnetdef.org GO TO https://apps.ubnetdef.org GO TO https://apps.ubnetdef.org GO TO https://apps.ubnetdef.org GO TO https://apps.ubnetdef.org
More informationRunning various Bigtop components
Running various Bigtop components Running Hadoop Components One of the advantages of Bigtop is the ease of installation of the different Hadoop Components without having to hunt for a specific Hadoop Component
More informationRunning Kmeans Spark on EC2 Documentation
Running Kmeans Spark on EC2 Documentation Pseudo code Input: Dataset D, Number of clusters k Output: Data points with cluster memberships Step1: Read D from HDFS as RDD Step 2: Initialize first k data
More informationNetwork Monitoring & Management. A few Linux basics
Network Monitoring & Management A few Linux basics Our chosen platform Ubuntu Linux 14.04.3 LTS 64-bit LTS = Long Term Support no GUI, we administer using ssh Ubuntu is Debian underneath There are other
More informationHow to Install and Configure Big Data Edition for Hortonworks
How to Install and Configure Big Data Edition for Hortonworks 1993-2015 Informatica Corporation. No part of this document may be reproduced or transmitted in any form, by any means (electronic, photocopying,
More informationInstalling Lemur on Mac OS X and CSE Systems
Installing Lemur on Mac OS X 10.6.4 and CSE Systems Everything all at once For those of you who just want to copy and paste, here is the quick and dirty. # Config for black.cse.msu.edu # Note that you
More informationDATA MIGRATION METHODOLOGY FROM SQL TO COLUMN ORIENTED DATABASES (HBase)
1 DATA MIGRATION METHODOLOGY FROM SQL TO COLUMN ORIENTED DATABASES (HBase) Nosql No predefined schema/ less rigid schema Unstructured and constantly evolving data No Declarative Query Language Weaker Transactional
More informationThe Analysis and Implementation of the K - Means Algorithm Based on Hadoop Platform
Computer and Information Science; Vol. 11, No. 1; 2018 ISSN 1913-8989 E-ISSN 1913-8997 Published by Canadian Center of Science and Education The Analysis and Implementation of the K - Means Algorithm Based
More informationInitial setting up of VPN Java version.
VPN Linux 64 bit To access work documents from home you will need to create a VPN connection. This will allow you to securely connect to Exeter University network from other location such as your home.
More informationBPPM Patrol Agent Installation Steps on Linux and Automation Integration
BPPM Patrol Agent Installation Steps on Linux and Automation Integration Author: Karlis Peterson, Software Consultant, BMC Software Version: 1.0 Date: May 12, 2013 DISCLAIMER NOTICE This is Field Developed
More informationFEPS. SSH Access with Two-Factor Authentication. RSA Key-pairs
FEPS SSH Access with Two-Factor Authentication RSA Key-pairs access.eps.surrey.ac.uk Contents: Introduction - 3 RSA Key-pairs - 3 Where can I use my RSA Key-Pair? - 3 Step 1 Prepare to generate your RSA
More informationa) Define Hadoop Ecosystem Ans: Hadoop eco system consist of two components Hadoop Distributed File System(HDFS)
ADVANCED DATA ANALYTICS 14CS/IT 703. November, 2017 Seventh Semester Time: Three Hours IV/IV B.Tech (Supplementary) DEGREE EXAMINATION Computer Science Engineering Maximum : 60 Marks Answer Question No.1
More informationSQOOP - QUICK GUIDE SQOOP - INTRODUCTION
SQOOP - QUICK GUIDE http://www.tutorialspoint.com/sqoop/sqoop_quick_guide.htm Copyright tutorialspoint.com SQOOP - INTRODUCTION The traditional application management system, that is, the interaction of
More informationBooting a Galaxy Instance
Booting a Galaxy Instance Create Security Groups First time Only Create Security Group for Galaxy Name the group galaxy Click Manage Rules for galaxy Click Add Rule Choose HTTPS and Click Add Repeat Security
More informationSetting up a Chaincoin Masternode
Setting up a Chaincoin Masternode Introduction So you want to set up your own Chaincoin Masternode? You ve come to the right place! These instructions are correct as of April, 2017, and relate to version
More informationConfigure HOSTNAME by adding the hostname to the file /etc/sysconfig/network. Do the same to all the other 3(4) nodes.
Network setup As the root user execute the command "ifconfig" on each host. Take a note of ipaddress's of all machines I have a home LAN so my addresses are class C which might be in the format 192.168.192.x.
More informationCOUCHDB - INSTALLATION
COUCHDB - INSTALLATION http://www.tutorialspoint.com/couchdb/couchdb_installation.htm Copyright tutorialspoint.com This chapter teaches you how to install CouchDB in windows as well as Linux systems. Installing
More informationhalvade Documentation
halvade Documentation Release 1.1.0 Dries Decap Mar 12, 2018 Contents 1 Introduction 3 1.1 Recipes.................................................. 3 2 Installation 5 2.1 Build from source............................................
More informationPentaho MapReduce with MapR Client
Pentaho MapReduce with MapR Client Change log (if you want to use it): Date Version Author Changes Contents Overview... 1 Before You Begin... 1 Use Case: Run MapReduce Jobs on Cluster... 1 Set Up Your
More informationCMU MSP Intro to Hadoop
CMU MSP 36602 Intro to Hadoop H. Seltman, April 3 and 5 2017 1) Carl had created an MSP virtual machine that you can download as an appliance for VirtualBox (also used for SAS University Edition). See
More informationIn this exercise you will practice working with HDFS, the Hadoop. You will use the HDFS command line tool and the Hue File Browser
Access HDFS with Command Line and Hue Data Files (local): ~/labs/data/kb/* ~/labs/data/base_stations.tsv In this exercise you will practice working with HDFS, the Hadoop Distributed File System. You will
More informationPerl and R Scripting for Biologists
Perl and R Scripting for Biologists Lukas Mueller PLBR 4092 Course overview Linux basics (today) Linux advanced (Aure, next week) Why Linux? Free open source operating system based on UNIX specifications
More informationHadoop On Demand User Guide
Table of contents 1 Introduction...3 2 Getting Started Using HOD... 3 2.1 A typical HOD session... 3 2.2 Running hadoop scripts using HOD...5 3 HOD Features... 6 3.1 Provisioning and Managing Hadoop Clusters...6
More informationTutorial for Assignment 2.0
Tutorial for Assignment 2.0 Web Science and Web Technology Summer 2011 Slides based on last years tutorial by Florian Klien and Chris Körner 1 IMPORTANT The presented information has been tested on the
More informationSession 1 Big Data and Hadoop - Overview. - Dr. M. R. Sanghavi
Session 1 Big Data and Hadoop - Overview - Dr. M. R. Sanghavi Acknowledgement Prof. Kainjan M. Sanghavi For preparing this prsentation This presentation is available on my blog https://maheshsanghavi.wordpress.com/expert-talk-fdp-workshop/
More informationEverything about Linux User- and Filemanagement
Everything about Linux User- and Filemanagement Lukas Prokop 20. April 2009 Inhaltsverzeichnis 1 Who I am 2 1.1 whoami..................................... 3 1.2 passwd......................................
More informationIntroduction into Big Data analytics Lecture 3 Hadoop ecosystem. Janusz Szwabiński
Introduction into Big Data analytics Lecture 3 Hadoop ecosystem Janusz Szwabiński Outlook of today s talk Apache Hadoop Project Common use cases Getting started with Hadoop Single node cluster Further
More information11/8/17 GETTING STARTED
11/8/17 GETTING STARTED ENTANDO EXECUTIVE SUMMARY This document describes the basics of how to get up and run a web application based on Entando v4.3.1. TARGET AUDIENCE This document is targeted to users
More informationDownloading and installing Db2 Developer Community Edition on Ubuntu Linux Roger E. Sanders Yujing Ke Published on October 24, 2018
Downloading and installing Db2 Developer Community Edition on Ubuntu Linux Roger E. Sanders Yujing Ke Published on October 24, 2018 This guide will help you download and install IBM Db2 software, Data
More informationInstalling SmartSense on HDP
1 Installing SmartSense on HDP Date of Publish: 2018-07-12 http://docs.hortonworks.com Contents SmartSense installation... 3 SmartSense system requirements... 3 Operating system, JDK, and browser requirements...3
More informationAbout Backup and Restore, on page 1 Supported Backup and Restore Procedures, on page 3
About, on page 1 Supported Procedures, on page 3 Back Up Automation Data Using the GUI, on page 4 Restore Automation Data Using the GUI, on page 6 Schedule a Backup of Automation Data Using the GUI, on
More informationTo configure the patching repository so that it can copy patches to alternate locations, use SFTP, SCP, FTP, NFS, or a premounted file system.
Configuring Protocols to Stage and 1 Deploy Linux and UNIX Patches VCM supports patching of managed machines in distributed environments, either geographically or separated by firewalls. VCM uses a single
More informationOpenEMR Insights Configuration Instructions
OpenEMR Insights provides ETL and Dashboards to enhance your OpenEMR intelligence backed by Pentaho CE and Saiku Analytics CE. To see OpenEMR Insights in action, head over to www.oemr.org. There you will
More informationdocs.hortonworks.com
docs.hortonworks.com Hortonworks Data Platform : Security Administration Tools Guide Copyright 2012-2014 Hortonworks, Inc. Some rights reserved. The Hortonworks Data Platform, powered by Apache Hadoop,
More informationGetting Started with Hadoop
Getting Started with Hadoop May 28, 2018 Michael Völske, Shahbaz Syed Web Technology & Information Systems Bauhaus-Universität Weimar 1 webis 2018 What is Hadoop Started in 2004 by Yahoo Open-Source implementation
More informationJohn the Ripper on a Ubuntu MPI Cluster
John the Ripper on a Ubuntu 10.04 MPI Cluster Pétur Ingi Egilsson petur [at] petur [.] eu 1 Table of Contents Foreword...3 History...3 Requirements...3 Configuring the Server...3 Requirements...3 Required
More informationCompile and Run WordCount via Command Line
Aims This exercise aims to get you to: Compile, run, and debug MapReduce tasks via Command Line Compile, run, and debug MapReduce tasks via Eclipse One Tip on Hadoop File System Shell Following are the
More informationCryptography Application : SSH. Cyber Security & Network Security March, 2017 Dhaka, Bangladesh
Cryptography Application : SSH Cyber Security & Network Security 20-22 March, 2017 Dhaka, Bangladesh Issue Date: [31-12-2015] Revision: [v.1] What is Safely Authentication I am Assured of Which Host I
More informationMarketC - Masternode Setup Guide
MarketC - Masternode Setup Guide Preface In this guide we will be focusing on setting up a masternode for Marketc (CMK). This guide will focus on a typical "hot node" / "cold wallet" scenario. The "hot
More informationGIT. A free and open source distributed version control system. User Guide. January, Department of Computer Science and Engineering
GIT A free and open source distributed version control system User Guide January, 2018 Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur Table of Contents What is
More informationInstalling Datameer with MapR on an Edge Node
Installing Datameer with MapR on an Edge Node If Datameer is installed on an edge node and has to be connected with MapR, you also need to install the MapR client software on the edge node, so the node
More informationHOD User Guide. Table of contents
Table of contents 1 Introduction...3 2 Getting Started Using HOD... 3 2.1 A typical HOD session... 3 2.2 Running hadoop scripts using HOD...5 3 HOD Features... 6 3.1 Provisioning and Managing Hadoop Clusters...6
More informationMIS Week 10. Operating System Security. Unix/Linux basics
MIS 5170 Operating System Security Week 10 Unix/Linux basics Tonight s Plan 2 Questions from Last Week Review on-line posts In The News Download Kali Install Kali Unix/Linux Basics Scripting Appropriate
More informationPost Ubuntu Install Exercises
Post Ubuntu Install Exercises PacNOG 3 June 18 Rarotonga, Cook Islands 1. Get used to using sudo 2. Create an ainst account 3. Learn how to install software 4. Install gcc and make 5. Learn how to control
More informationGreenplum Data Loader Installation and User Guide
Greenplum DataLoader 1.2 Installation and User Guide Rev: A01 1 Copyright 2012 EMC Corporation. All rights reserved. EMC believes the information in this publication is accurate as of its publication date.
More informationProject 1 Setup. Some relevant details are the output of: 1. uname -a 2. cat /etc/*release 3. whereis java 4. java -version 5.
Project 1 Setup The purpose of this document is to help you to prepare your development machine for the project by: 1. Installing any missing tools 2. Setting up required environment variables and paths
More informationGetting Started. Table of contents. 1 Pig Setup Running Pig Pig Latin Statements Pig Properties Pig Tutorial...
Table of contents 1 Pig Setup... 2 2 Running Pig... 3 3 Pig Latin Statements... 6 4 Pig Properties... 8 5 Pig Tutorial... 9 1. Pig Setup 1.1. Requirements Mandatory Unix and Windows users need the following:
More informationLAB #7 Linux Tutorial
Gathering information: LAB #7 Linux Tutorial Find the password file on a Linux box Scenario You have access to a Linux computer. You must find the password file on the computer. Objective Get a listing
More informationExpedition. Hardening Guide Version Palo Alto Networks, Inc.
Expedition Hardening Guide Version 1.0 1 Palo Alto Networks, Inc. www.paloaltonetworks.com 2018 Palo Alto Networks, Inc. Palo Alto Networks is a registered trademark of Palo Alto Networks. You can find
More informationAims. Background. This exercise aims to get you to:
Aims This exercise aims to get you to: Import data into HBase using bulk load Read MapReduce input from HBase and write MapReduce output to HBase Manage data using Hive Manage data using Pig Background
More informationBUILD LINUX LEARNING LAB FOR FREE
HOW TO BUILD LINUX LEARNING LAB FOR FREE Proudly Presented by: LearnLinux.ca Learn Linux - kickstart your IT career Table of Contents 1. What is Linux? 2. VMware Workstation Player Download 3. VMware Workstation
More informationUbuntu Practice and Configuration Post Installation Exercises interlab at AIT Bangkok, Thailand
Ubuntu Practice and Configuration Post Installation Exercises interlab at AIT Bangkok, Thailand 1. Get used to using sudo 2. Create an inst account 3. Learn how to install software 4. Update /etc/apt/sources.list
More informationGetting Started with Pentaho and Cloudera QuickStart VM
Getting Started with Pentaho and Cloudera QuickStart VM This page intentionally left blank. Contents Overview... 1 Before You Begin... 1 Prerequisites... 1 Use Case: Development Sandbox for Pentaho and
More informationApache Manually Install Ubuntu From Usb
Apache Manually Install Ubuntu 12.10 From Usb On Windows 8 Using WinUSB, we can create startup disks for Windows Vista, 7, 8, and Windows PE. You can install WinUSB on Ubuntu 13.10/13.04/12.10/12.04 using
More informationHBase Installation and Configuration
Aims This exercise aims to get you to: Install and configure HBase Manage data using HBase Shell Install and configure Hive Manage data using Hive HBase Installation and Configuration 1. Download HBase
More informationLENS Server Maintenance Guide JZ 2017/07/28
LENS Server Maintenance Guide JZ 2017/07/28 Duty Maintain LENS server with minimum downtime Patch critical vulnerabilities Assist LAB member for using the LENS services Evaluate for custom requirements
More informationIntroduction to remote command line Linux. Research Computing Team University of Birmingham
Introduction to remote command line Linux Research Computing Team University of Birmingham Linux/UNIX/BSD/OSX/what? v All different v UNIX is the oldest, mostly now commercial only in large environments
More informationHANDS UP IF YOU DON T HAVE A VM OR IF YOU DON T REMEMBER YOUR PASSWORDS. Or something broke
HANDS UP IF YOU DON T HAVE A VM OR IF YOU DON T REMEMBER YOUR PASSWORDS Or something broke Securing your VM 101 Getting Comfy in Linux -> Comfort ++ You After This Talk Quick recap from last week. Setup
More informationSAS Event Stream Processing for Edge Computing 4.3: Deployment Guide
SAS Event Stream Processing for Edge Computing 4.3: Deployment Guide SAS Documentation June 2017 The correct bibliographic citation for this manual is as follows: SAS Institute Inc. 2017. SAS Event Stream
More informationHow to Install and Configure EBF16193 for Hortonworks HDP 2.3 and HotFix 3 Update 2
How to Install and Configure EBF16193 for Hortonworks HDP 2.3 and 9.6.1 HotFix 3 Update 2 1993-2015 Informatica Corporation. No part of this document may be reproduced or transmitted in any form, by any
More informationCluster Setup. Table of contents
Table of contents 1 Purpose...2 2 Pre-requisites...2 3 Installation...2 4 Configuration... 2 4.1 Configuration Files...2 4.2 Site Configuration... 3 5 Cluster Restartability... 10 5.1 Map/Reduce...10 6
More informationCreate Test Environment
Create Test Environment Describes how to set up the Trafodion test environment used by developers and testers Prerequisites Python Passwordless ssh If you already have an existing set of ssh keys If you
More informationSetting up VPS on Ovh public cloud and installing lamp server on Ubuntu instance
Setting up VPS on Ovh public cloud and installing lamp server on Ubuntu instance What is OVH Public Cloud Public Cloud Instances provides a choice of two types of virtual machines: the RAM instances are
More informationInstall and Configure Ubuntu on a VirtualBox Virtual Machine
Install and Configure Ubuntu on a VirtualBox Virtual Machine Ronald Mak Department of Computer Engineering Department of Computer Science January 11, 2019 Introduction Because the class will use Linux
More informationAn Overview of SSH. Presentation to Linux Users of Victoria. Melbourne, August 26, 2017
An Overview of SSH Presentation to Linux Users of Victoria Melbourne, August 26, 2017 http://levlafayette.com Utilisation and Rationale The most common use of SSH (secure shell) is remote login access
More informationLinux Kung Fu. Ross Ventresca UBNetDef, Fall 2017
Linux Kung Fu Ross Ventresca UBNetDef, Fall 2017 GOTO: https://apps.ubnetdef.org/ What is Linux? Linux generally refers to a group of Unix-like free and open source operating system distributions built
More informationManually Java 7 Update Bit Windows 8
Manually Java 7 Update 21 64 Bit Windows 8 Oracle strongly recommends that all Java SE 8 users upgrade to this release. Learn more arrow more arrow, Download JDK 7 Update 40 with NetBeans 7.3.1. Platform(s):
More informationInstalling Connector on Linux
CHAPTER 3 Revised: July 15, 2010 Overview This chapter provides a step-by-step guide to installing the Linux Connector on x86 and x86-64 servers running either Red Hat Enterprise Linux version 5 or Cent
More information